Skip to main content

Overview

gemini-3-flash is the newer long-context Gemini option in Kyma. It is best treated as a newer, more premium long-context choice rather than the default safe pick.

Specs

FieldValue
Model IDgemini-3-flash
Best forLong-context reasoning, multimodal analysis
Context window1M
Max output tokens8K
Input modalitiesText, image, audio, video
Output modalitiesText
Tool callingYes
Structured outputsYes
Prompt cachingYes
SpeedFast
Cost bandPremium
Release stagePreview

Use this when

  • You want the newer Gemini generation.
  • You need 1M context plus stronger reasoning than the cheaper long-context path.
  • You can tolerate preview-stage behavior in exchange for newer capability.

Pick something else when

  • You want the safer long-context default: use gemini-2.5-flash.
  • You want the best overall default: use qwen-3.6-plus.
  • You need the strongest coding agent behavior: use kimi-k2.5.

Example

from openai import OpenAI

client = OpenAI(base_url="https://kymaapi.com/v1", api_key="ky-...")

response = client.chat.completions.create(
    model="gemini-3-flash",
    messages=[{"role": "user", "content": "Analyze this large mixed-media research dump and identify the strongest conclusions."}]
)