Z.ai: GLM 4.7 Flash chat
As a 30B-class SOTA model, GLM-4.7-Flash offers a new option that balances performance and efficiency. It is further optimized for agentic coding use cases, strengthening coding capabilities, long-horizon task planning,...
Capabilities
Context Window 202k tokens
Max Output 16k tokens
Inputs
Outputs
Pricing (per 1M tokens)
Input $0.06
Output $0.40
Cache Read $0.01
Cache Write -
Supported Parameters
frequency_penaltyinclude_reasoninglogit_biasmax_tokensmin_ppresence_penaltyreasoningrepetition_penaltyresponse_formatseedstopstructured_outputstemperaturetool_choicetoolstop_ktop_p