Z.ai debuts open source GLM-4.6V, a native tool-calling vision model for multimodal reasoning

Carl Franzen December 9, 2025 Credit: VentureBeat made with Flux 2 on Fal.aiChinese AI startup Zhipu AI aka Z.ai has released its GLM-4.6V series, a new generation of open-source vision-language models (VLMs) optimized for multimodal reasoning, frontend automation, and high-efficiency deployment. The release includes two models in “large” and “small” sizes: GLM-4.6V (106B), a larger 106-billion parameter model aimed at cloud-scale inferenceGLM-4.6V-Flash (9B), a smaller model of only 9 billion parameters designed for low-latency, local applicationsRecall that generally speaking, models with more parameters — or internal settings governing their behavior, i.e. weights and biases — are more powerful, performant, and capable of performing at a higher general level across more varied…

Read more on VentureBeat