7 stories tagged with #glm, in publish-time order across the WeSearch catalog. Tag pages update as new stories ingest.
⌘ RSS feed for this tag → or search "Glm"
Scaling Pain of Coding Agent Serving: Lessons from Debugging GLM-5 at Scale
I switched from OpenAI to z.ai for codiai coding review ng and I'm genuinely happy with it — honest review
I've been using z.ai for about three months now. Here's what I've learned. What z.ai...…
Abliterlitics: Benchmarks and Tensor Comparison for Heretic, Abliterlix, Huiui, HauhauCS for GLM 4.7 Flash
"Use Claude Code for FREE" is a Trap
Why free AI coding via Nvidia NIM and OpenRouter is a trap. The Cheap-Intelligent-Fast trilemma, 40 RPM rate limits, Opus 4.7 vs GPT-5.5 vs MiniMax M2.7 benchmarks, and why your fi…
Used a Claude Code skill to fine-tune Qwen3-1.7B from 327 noisy traces, matches GLM-5
Had 327 production traces from a restaurant-reservation agent I wanted to retrain. The plan was to fine-tune a smaller self-hostable model so I could ditch the frontier-API bill. T…
Anthropic's Claude remote uses GLM-4.7
I just noticed this after a bug wasn't getting fixed. If you start a Claude code remote environment the default model (hidden on mobile) is glm 4.7 I assumed anthropic only used th…
GLM 5.1 Locally: 40tps, 2000+ pp/s
After some sglang patching and countless experiments, managed to get reap-ed nvfp4 version running stable and FAST on 4 x RTX 6000 Pros (limited to 350W). Very happy with performan…