Forked from zai-org/glm-4.7-flash
Description
GLM 4.7 Flash is a 30B A3B MoE model form Z.ai. It supports a context length of 128k tokens and achieves strong performance on coding benchmarks among models of similar scale.
Capabilities
Minimum system memory
Tags
Last updated
Updated 20 days agobyREADME
Custom Fields
Special features defined by the model author
Enable Thinking
: boolean
(default=true)
Controls whether the model will think before replying
Clear Thinking
: boolean
(default=false)
Controls whether thinking content is cleared from history
Parameters
Custom configuration options included with this model
Sources
The underlying model files this model uses