Kyma currently serves 50 models across language, image, video, and audio. All models are verified working and accessible through the sameDocumentation Index
Fetch the complete documentation index at: https://docs.kymaapi.com/llms.txt
Use this file to discover all available pages before exploring further.
/v1 API. Language models are pay-per-token; image and video are flat per-call or pay-per-second depending on SKU; audio is per-character (TTS), per-minute (transcription / understand), or flat per call (music, voice clone, voice design). Live list and pricing available via GET /v1/models and GET /v1/credits/pricing.
How to choose quickly
- Start with
qwen-3.6-plusif you want the best default for general work, coding, and reasoning. - Use
kimi-k2.6for tool-heavy agents, long coding sessions, and image-aware workflows. - Use
deepseek-v4-profor top reasoning and complex coding with 1M context. - Use
deepseek-v4-flashwhen you want V4-family quality at the cheapest price. - Use
gemini-2.5-flashwhen you need 1M context or cheap long-context throughput. - Use
qwen-3-32bwhen latency matters and you still want strong coding quality. - Use
glm-5.1when you need a long-running coding agent for repo-scale engineering work.
Filter the catalog
The liveGET /v1/models endpoint now supports capability filters so agents can select models programmatically instead of hardcoding a shortlist.
Recommended Models
Qwen 3.6 Plus
#1 most popular. Closed-weight, highest quality overall. 131K context.
DeepSeek V4 Flash
Best value V4. 1M context, MIT, native reasoning. $0.19/M input.
DeepSeek V4 Pro
Top reasoning. 1.6T MoE flagship, 1M context, complex coding.
Kimi K2.5
Best for agents. Multimodal agentic model. 262K context.
Capability Guide
| Need | Best first pick | Why |
|---|---|---|
| General default | qwen-3.6-plus | Best overall quality, strong multilingual reasoning |
| Tool-heavy agents | kimi-k2.6 | Strong tool use, long context, multimodal |
| Top reasoning | deepseek-v4-pro | 1.6T MoE flagship, 1M context, native reasoning |
| Best value | deepseek-v4-flash | V4-tier quality at the lowest price, 1M context |
| Long-running coding agents | glm-5.1 | Better fit for repo-scale engineering and multi-step execution |
| Fast coding | qwen-3-32b | Lower latency while staying strong on code and math |
| 1M context | gemini-2.5-flash | Cheapest long-context option on Kyma |
| Vision | gemma-4-31b | Reliable image + text workflows |
| Image generation | recraft-v4 | #1 HF Arena. Design-quality default for brand and illustration. |
| Video generation | kling-3-pro | Premium cinematic clips, hero brand video |
Tier 1 — Highest Quality
| Model ID | Name | Context | Speed | Best For |
|---|---|---|---|---|
qwen-3.6-plus | Qwen 3.6 Plus | 131K | Medium | General, #1 traffic |
deepseek-v4-pro | DeepSeek V4 Pro | 1M | Medium | Top reasoning, complex coding |
deepseek-v4-flash | DeepSeek V4 Flash | 1M | Fast | Best value, long context |
deepseek-v3 | DeepSeek V3 | 160K | Medium | Previous-gen flagship, stable |
deepseek-r1 | DeepSeek R1 | 64K | Slow | Reasoning, analysis |
kimi-k2.6 | Kimi K2.6 | 262K | Medium | Agentic coding, multimodal |
gemma-4-31b | Gemma 4 31B | 128K | Medium | Multimodal, vision |
qwen-3-32b | Qwen 3 32B | 32K | Fast | Code, math, multilingual |
llama-3.3-70b | Llama 3.3 70B | 128K | Fast | General, most popular open model |
minimax-m2.5 | MiniMax M2.5 | 196K | Medium | Agentic coding (SWE-bench 80.2%) |
glm-5.1 | GLM 5.1 | 203K | Medium | Long-running coding agents, repo-scale engineering |
Tier 2 — High Quality
| Model ID | Name | Context | Speed | Best For |
|---|---|---|---|---|
minimax-m2.7 | MiniMax M2.7 | 205K | Medium | Agentic coding, productivity, debugging |
gpt-oss-120b | GPT-OSS 120B | 128K | Medium | Writing, general intelligence |
qwen-3-coder | Qwen 3 Coder | 131K | Medium | Code generation |
gemini-2.5-flash | Gemini 2.5 Flash | 1M | Fast | Long context |
gemini-3-flash | Gemini 3 Flash | 1M | Fast | Ultra-long context |
glm-4.5-air | GLM 4.5 Air | 131K | Fast | Cheap agentic bulk tasks |
glm-4.7-flash | GLM 4.7 Flash | 203K | Fast | Cheap long-context throughput |
Image Generation
Async endpoint atPOST /v1/images/generations — pay per image, no token billing. See the Image Generation guide for prompting tips and full examples.
| Model | Best For | Cost / image | Input |
|---|---|---|---|
recraft-v4 | Default — design-quality, brand assets | $0.054 | text |
recraft-v4-pro | 4MP print-ready design | $0.338 | text |
recraft-v4-vector | Native SVG — logos, icons | $0.108 | text |
recraft-v4-vector-pro | 4MP SVG, print-ready | $0.405 | text |
flux-2-pro | Photoreal, multi-reference blend | 0.101 | text + image(s) |
flux-kontext-pro | Image edit, inpaint, refine | $0.054 | text + image |
ideogram-v3 | Typography, logos, packaging | $0.108 | text |
Video Generation
Async endpoint atPOST /v1/videos/generations — pay per second of generated footage. See the Video Generation guide for prompting tips and the full per-SKU breakdown.
| Model | Best For | Cost / sec | 5s clip | Audio | Input |
|---|---|---|---|---|---|
kling-2.5-pro | Budget cinematic, b-roll | $0.0945 | $0.4725 | — | text + image |
kling-3-pro | Premium cinematic, hero video | $0.1512 | $0.7560 | — | text + image |
kling-3-pro-audio | Cinematic w/ diegetic sound | $0.2268 | $1.1340 | native | text + image |
seedance-2-pro | Action, multi-shot, social | $0.40959 | $2.04795 | bundled | text + image |
seedance-2-fast | Social shorts, rapid iteration | $0.326565 | $1.63283 | bundled | text + image |
For the live canonical list, use
GET /v1/models. Disabled models are intentionally omitted from this page.Models are updated regularly. Use
GET /v1/models for the latest list.