Gemma 4 26B A4B (free) chat

openrouter
google
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference — delivering near-31B quality at...

Capabilities

Context Window 262k tokens
Max Output 32k tokens
Inputs
Outputs

Pricing (per 1M tokens)

Input $-
Output $-
Cache Read -
Cache Write -

Supported Parameters

include_reasoningmax_tokensreasoningresponse_formatseedtemperaturetool_choicetoolstop_p