TokenFactory · 4-page brief

TokenFactory — Production inference for the open-model frontier

Run gpt-oss-120B, Kimi-K2.5, Qwen3-Coder, GLM-5, DeepSeek V3.2 and the full open frontier behind one OpenAI-compatible API. Pay-as-you-go, batch, post-training and dedicated.

What's inside
Section 1

The catalog

40+ open and frontier models, evaluated and added within days of release. Same API, no migration.

Section 2

Four service modes

Inference, batch, post-training, dedicated deployments — pick by workload, mix freely.

Section 3

Performance

Heterogeneous serving, optimized runtimes per silicon class, and the operational SLAs you can sign for.

Section 4

Pricing

Per-token, transparent. Batch up to 50% lower than real-time. Dedicated reservations for predictable load.

More briefs

Ready to operationalize your agents?

Talk to our team about a pilot on Synaptix Cloud or on-prem.