GLM-4.6V is GLM's newest open-source multimodal model with a 128k context window. It features native function calling, bridging visual perception with executable actions for complex agentic workflows like web search and coding.
GLM-4.6V is a significant iteration for the GLM multimodal series. It scales the training context window to 128k and hits SOTA visual understanding for its size.
The biggest update here is the native Function Calling. For the first time in the GLM architecture, tool use is integrated directly into the visual model. This effectively bridges the gap from "visual perception" to "executable action."
It can automatically generate high-quality image-text interleaved content and handle complete workflows independently, like viewing products, comparing prices, and generating shopping lists. The frontend replication and visual interaction capabilities are also impressive, which significantly shortens the path from design to code for developers.
“Open-source multimodal model with native tool use”
GLM-4.6V launched on Product Hunt on December 9th, 2025 and earned 247 upvotes and 11 comments, placing #4 on the daily leaderboard. GLM-4.6V is GLM's newest open-source multimodal model with a 128k context window. It features native function calling, bridging visual perception with executable actions for complex agentic workflows like web search and coding.
On the analytics side, GLM-4.6V competes within Open Source, Artificial Intelligence and Development — topics that collectively have 540.3k followers on Product Hunt. The dashboard above tracks how GLM-4.6V performed against the three products that launched closest to it on the same day.
Who hunted GLM-4.6V?
GLM-4.6V was hunted by Zac Zuo. A “hunter” on Product Hunt is the community member who submits a product to the platform — uploading the images, the link, and tagging the makers behind it. Hunters typically write the first comment explaining why a product is worth attention, and their followers are notified the moment they post. Around 79% of featured launches on Product Hunt are self-hunted by their makers, but a well-known hunter still acts as a signal of quality to the rest of the community. See the full all-time top hunters leaderboard to discover who is shaping the Product Hunt ecosystem.
Hi everyone!
GLM-4.6V is a significant iteration for the GLM multimodal series. It scales the training context window to 128k and hits SOTA visual understanding for its size.
The biggest update here is the native Function Calling. For the first time in the GLM architecture, tool use is integrated directly into the visual model. This effectively bridges the gap from "visual perception" to "executable action."
It can automatically generate high-quality image-text interleaved content and handle complete workflows independently, like viewing products, comparing prices, and generating shopping lists. The frontend replication and visual interaction capabilities are also impressive, which significantly shortens the path from design to code for developers.
Try it on Z.ai or find the open weights on HF.