Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.arkor.ai/llms.txt

Use this file to discover all available pages before exploring further.

Today both the model field of createTrainer and Studio’s Playground (base-model mode) accept exactly one value: gemma-4-E4B-it. Below: what that value gets you, why the list is short on purpose, and what’s next.

What you can pick today

gemma-4-E4B-it is the Gemma 4 instruction-tuned build, packaged by Unsloth for fast LoRA / QLoRA fine-tuning. Every starter template (triage, translate, redaction) targets it by default.
import { createTrainer } from "arkor";

export const trainer = createTrainer({
  name: "support-bot-v1",
  model: "unsloth/gemma-4-E4B-it",
  dataset: { type: "huggingface", name: "arkorlab/triage-demo" },
});
The canonical value the cloud API accepts is unsloth/gemma-4-E4B-it; the unsloth/ prefix stays in code as a deliberate credit to the Unsloth project that produced the build.

Why one model, right now

The priority in alpha is making both training and inference feel fast on one base model, not building a wide catalog. Tuning the whole path from arkor dev through arkor start to a chat-ready checkpoint against one model means optimizations to that model compound across runs, instead of getting diluted across many. Gemma 4 fits this stage because it is small enough to iterate on quickly, capable enough to be useful inside a product, and openly licensed.

What’s coming

Gemma 4 family

Open the model field to the full Gemma 4 family so you can pick the variant that matches your use case (size, capability, latency, quality).

Other open-weight families

Expanding to additional open-weight families is on the roadmap Backlog. Today the focus stays on Gemma 4 to keep delivery on it fast and stable.

See also