GLM 4.7
Description
GLM 4.7 is extremely slow it took 2 minutes to respond to a Hi
Plugins
Oh My opencode
OpenCode version
im in the desktop version
Steps to reproduce
No response
Screenshot and/or share link
No response
Operating System
Windows 11
Terminal
Windows Terminal
This issue might be a duplicate of existing issues. Please check:
- #7692: [Bug] JSON Parse Error with Zhipu GLM-4.7: Stream chunks are concatenated incorrectly - may cause response delays
- #7779: GLM 4.7 thinking process not properly formatted - missing leading tags - could affect performance
- #6074: Differences between the free and paid versions of GLM 4.7 - discusses response speed differences
Feel free to ignore if none of these address your specific case.
Because it's overloaded. Are you using any plugina of note? Oh oh-my-opencode this is typical and another model should be configured.
Make sure you limit the context to max 100K.
Hitting the same. Random inexplicable delays. Lots of HTTP timeouts.
Getting 1303 errors very frequently: https://docs.z.ai/api-reference/api-code
High frequency usage of this API, please reduce frequency or contact customer service to increase limits
I have no plugins. I am getting this with very little context usage.
Using glm-4.7-free
I am using z.ai Coding Plan with GLM-4.7 and kind of frustrated of the speed. I guess it is ok for the price I pay, but compared to snappy opus 4.5 (which begun to shrink their quotas agressively and everybody is mad) it is unbearable.
@lezhnev74 there must be something wrong, latency or context overload, because I have quite good experience with the model, it's pretty comparable with sonnet 4.5 in terms of speed, a bit less capable. The only thing I had was occasional looping where the model keeps on spitting out the same chinese character. I also had my terminal screwed up by chinese unicode characters, however by setting a codepage on the terminal this got solved as well.
this isnt really something on our end, providers get hammered