ion7-llm

ion7-llm

API Reference

Chat pipeline + multi-session inference orchestration on top of ion7-core. Per-seq KV snapshots, prefix cache, three-channel streaming, schema-constrained sampling, interleaved-thinking tool loop, embeddings.

23 modules
129 documented functions

Sampler

Utilities