Z.ai: GLM 5 Turbo chat
GLM-5 Turbo is a new model from Z.ai designed for fast inference and strong performance in agent-driven environments such as OpenClaw scenarios. It is deeply optimized for real-world agent workflows involving long execution chains, with improved complex instruction decomposition, tool use, scheduled and persistent execution, and overall stability across extended tasks.
Capabilities
Context Window 202k tokens
Max Output 131k tokens
Inputs
Outputs
Pricing (per 1M tokens)
Input $1.20
Output $4.00
Cache Read $0.24
Cache Write -
Supported Parameters
include_reasoningmax_tokensreasoningresponse_formattemperaturetool_choicetoolstop_p