Gemma 4 26B A4B (free) chat
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference — delivering near-31B quality at...
Capabilities
Context Window 262k tokens
Max Output 32k tokens
Inputs
Outputs
Pricing (per 1M tokens)
Input $-
Output $-
Cache Read -
Cache Write -
Supported Parameters
include_reasoningmax_tokensreasoningresponse_formatseedtemperaturetool_choicetoolstop_p