ERNIE 4.5 21B A3B chat

openrouter
A sophisticated text-based Mixture-of-Experts (MoE) model featuring 21B total parameters with 3B activated per token, delivering exceptional multimodal understanding and generation through heterogeneous MoE structures and modality-isolated routing. Supporting an extensive 131K token context length, the model achieves efficient inference via multi-expert parallel collaboration and quantization, while advanced post-training techniques including SFT, DPO, and UPO ensure optimized performance across diverse applications with specialized routing and balancing losses for superior task handling.

Capabilities

Context Window 120k tokens
Max Output 8k tokens
Inputs
Outputs

Pricing (per 1M tokens)

Input $0.07
Output $0.28
Cache Read -
Cache Write -

Supported Parameters

frequency_penaltymax_tokenspresence_penaltyrepetition_penaltyseedstoptemperaturetool_choicetoolstop_ktop_p