Decision Tree
Not sure where to start? Follow this guide:π I need the best all-around model
π I need the best all-around model
β Llama 3.3 70B (
llama-3.3-70b)The most popular open source model. Great at coding, reasoning, writing, and general tasks. 128K context window. Ultra-fast via Kyma.π» I need help with coding
π» I need help with coding
β Qwen 3 32B (
qwen-3-32b)Top coding model. Excellent at code generation, debugging, math, and multilingual tasks. 32K context.π§ I need the highest quality / complex reasoning
π§ I need the highest quality / complex reasoning
β Qwen 3 235B (
qwen-3-235b-cerebras)The largest model available on Kyma. 235 billion parameters. Best for complex analysis, research, and nuanced reasoning. Still ultra-fast.π I need to process long documents (>100K tokens)
π I need to process long documents (>100K tokens)
β Llama 4 Scout (
llama-4-scout)512K context window β can process entire books, codebases, or large datasets in a single request.πΌοΈ I need vision / multimodal capabilities
πΌοΈ I need vision / multimodal capabilities
β Gemma 4 31B (
gemma-4-31b)Googleβs newest open model. Supports image understanding. 128K context.β‘ I need the fastest possible response
β‘ I need the fastest possible response
β Llama 3.1 8B (
llama-3.1-8b)Smallest model, lowest latency. Perfect for classification, extraction, or simple Q&A where speed matters most.π€ I'm building an AI agent with tool use
π€ I'm building an AI agent with tool use
β Kimi K2 (
kimi-k2)Purpose-built for agentic coding and tool use. Excels at multi-step reasoning and function calling.Full Model Comparison
| Model | Context | Speed | Best For | Quality |
|---|---|---|---|---|
| β Llama 3.3 70B | 128K | Fast | General, code, reasoning | βββββ |
| β Qwen 3 32B | 32K | Fast | Code, math, multilingual | βββββ |
| β Qwen 3 235B | 32K | Fast | Complex reasoning | βββββ |
| β Gemma 4 31B | 128K | Medium | Multimodal, vision | ββββ |
| π₯ Llama 4 Scout | 512K | Fast | Long documents | ββββ |
| π₯ Kimi K2 | 128K | Fast | Agentic coding | ββββ |
| π₯ GPT-OSS 120B | 128K | Medium | Writing, general | ββββ |
| π₯ Gemma 4 26B MoE | 128K | Fast | Efficient general | ββββ |
| Gemini 3 Flash | 1M | Medium | Ultra-long context | ββββ |
| Gemini 2.5 Flash | 1M | Medium | Long context | βββ |
| GPT-OSS 20B | 128K | Fast | Simple tasks | βββ |
| Gemma 3 27B | 128K | Medium | General | βββ |
| Llama 3.1 8B | 8K | Fastest | Quick tasks | βββ |
| Llama 3.1 8B (alt) | 8K | Fastest | Quick tasks | βββ |
Use by task
| Task | Recommended Model | Why |
|---|---|---|
| Chatbot | Llama 3.3 70B | Best general quality + fast |
| Code generation | Qwen 3 32B | Top coding benchmarks |
| Code review | Kimi K2 | Built for agentic code tasks |
| Summarization | Llama 4 Scout | 512K context for long docs |
| Data extraction | Llama 3.1 8B | Fast + structured output |
| Creative writing | GPT-OSS 120B | Strong writing capabilities |
| Research analysis | Qwen 3 235B | Highest reasoning quality |
| Image understanding | Gemma 4 31B | Multimodal support |