ion7-llm
API Reference
Chat pipeline + multi-session inference orchestration on top of ion7-core. Per-seq KV snapshots, prefix cache, three-channel streaming, schema-constrained sampling, interleaved-thinking tool loop, embeddings.
23
modules
129
documented functions
Sampler
Utilities