Today both theDocumentation Index
Fetch the complete documentation index at: https://docs.arkor.ai/llms.txt
Use this file to discover all available pages before exploring further.
model field of createTrainer and Studio’s Playground (base-model mode) accept exactly one value: gemma-4-E4B-it. Below: what that value gets you, why the list is short on purpose, and what’s next.
What you can pick today
gemma-4-E4B-it is the Gemma 4 instruction-tuned build, packaged by Unsloth for fast LoRA / QLoRA fine-tuning. Every starter template (triage, translate, redaction) targets it by default.
unsloth/gemma-4-E4B-it; the unsloth/ prefix stays in code as a deliberate credit to the Unsloth project that produced the build.
Why one model, right now
The priority in alpha is making both training and inference feel fast on one base model, not building a wide catalog. Tuning the whole path fromarkor dev through arkor start to a chat-ready checkpoint against one model means optimizations to that model compound across runs, instead of getting diluted across many.
Gemma 4 fits this stage because it is small enough to iterate on quickly, capable enough to be useful inside a product, and openly licensed.
What’s coming
Gemma 4 family
Open the
model field to the full Gemma 4 family so you can pick the variant that matches your use case (size, capability, latency, quality).Other open-weight families
Expanding to additional open-weight families is on the roadmap Backlog. Today the focus stays on Gemma 4 to keep delivery on it fast and stable.
See also
- Trainer concept for what the
modelfield is and where it sits among the other fields. createTrainerreference for the fullTrainerInputtype.- Roadmap for the broader picture of what’s next.