Jump to content

Draft:GLM

From Wikipedia, the free encyclopedia
GLM
Developer(s)Zhipu AI
Initial releaseAugust 2022; 2 years ago (2022-08)
Stable release
4-0414 / April 2025; 0 months ago (2025-04)
Repositorygithub.com/THUDM/GLM-4
Written inPython
Operating system
TypeChatbot
LicenseMIT, ChatGLM License, GLM-4 License
Websitechat.z.ai

GLM (also called ChatGLM or General Language Model) is a family of open-source bilingual Large Language Models developed by Z.ai (Zhipu AI) and The Knowledge Engineering Group & Data Mining (THUDM) at Tsinghua University. In April 2025, the latest GLM model family, GLM-4-0414, achieved competitive performance on benchmarks with leading models such as OpenAI's GPT-4o and DeepSeek's R1.[1]

History

[edit]

The first version of GLM, GLM-130B, was released in August 2022.[2] While not open-sourced, the model was made available to researchers through an application for non-commercial research purposes.

The first publicly-available open-access GLM model was released in February 2023 with 2B and 10B variants, designed for both blank infilling and text generation.[3]

ChatGLM

[edit]

In March 2023, ChatGLM 6B was open-sourced on Hugging Face as a conversational language model inspired by ChatGPT.[4] In May 2023, VisualGLM, a 6B parameter vision language model based on ChatGLM, was released.

In June 2023, ChatGLM 2 6B was released with improved performance and a variant with a 32K context window.[5]

In October 2023, ChatGLM 3 6B was released with both chat and base variants, with added support for function calling and improved performance.[6]

GLM-4

[edit]

In June 2024, the GLM-4 model family was released with 9 billion parameters. The model family includes GLM-4-Chat, GLM-4-Chat-1M (with a 1M token context window), and GLM-4V-9B (a vision language model). The model was competitive with other open models, outperforming Mistral 7B but falling short of contemporaneous commercial models such as GPT-4 and Claude 3 Opus.[7]

GLM-4-0414

[edit]

In April 2025, Zhipu AI released GLM-4-0414, a model based on GLM-4 with greatly improved performance. The model was trained on 15T tokens of web and synthetic data, with performance competitive with DeepSeek V3 and GPT-4o.[8][9] The following variants were made available:

  • GLM-4-32B-0414: A chat model with 32B parameters fine-tuned from the GLM-4 Base 0414 model.[10]
  • GLM-4-32B-Base-0414: A base model with 32B parameters.[11]
  • GLM-Z1-32B-0414: A reasoning model similar to OpenAI's o1 based on GLM-4.[12]
  • GLM-Z1-32B-Rumination-0414: A reasoning model based on GLM-Z1-32B but with support for "rumination," similar to OpenAI's Deep Research.[13]
  • GLM-4-9B-0414: A chat model with 9B parameters.[14]
  • GLM-Z1-9B-0414: A variant of the GLM-Z1 model based on the 9B chat model.[15]

Notably, the GLM-4-9B-0414 base model was not released.[16]

Licenses

[edit]

The GLM and ChatGLM models have been licensed under a wide variety of licenses. The original GLM-130B model is licensed under a non-commercial license prohibiting commercial use.[17] The GLM 2B and 10B did not include a license for the model, but the code itself was licensed under the MIT license.[18] The ChatGLM models are licensed under a restrictive revocable license.[19] The GLM-4 models are licensed under a revocable license that requires attribution.[20] The GLM-4-0414 model family is licensed under the MIT license, a permissive, OSI-approved license.[21]

References

[edit]
  1. ^ Razzaq, Asif (2025-04-14). "THUDM Releases GLM 4: A 32B Parameter Model Competing Head-to-Head with GPT-4o and DeepSeek-V3". MarkTechPost. Retrieved 2025-04-28.
  2. ^ "GitHub - THUDM/GLM-130B at 99509f09cce5ac32aee8ebf78e18f4399319dba4". GitHub. Retrieved 2025-04-29.
  3. ^ "THUDM/glm-10b". huggingface.co. Retrieved 2025-04-28.
  4. ^ "THUDM/chatglm-6b". huggingface.co. Retrieved 2025-04-28.
  5. ^ "THUDM/chatglm2-6b". huggingface.co. Retrieved 2025-04-29.
  6. ^ "THUDM/chatglm3-6b". huggingface.co. Retrieved 2025-04-29.
  7. ^ "THUDM/glm-4-9b-chat". huggingface.co. 2024-08-12. Retrieved 2025-04-29.
  8. ^ "THUDM/GLM-4-32B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  9. ^ Razzaq, Asif (2025-04-14). "THUDM Releases GLM 4: A 32B Parameter Model Competing Head-to-Head with GPT-4o and DeepSeek-V3". MarkTechPost. Retrieved 2025-04-29.
  10. ^ "THUDM/GLM-4-32B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  11. ^ "THUDM/GLM-4-32B-Base-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  12. ^ "THUDM/GLM-Z1-32B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  13. ^ "THUDM/GLM-Z1-Rumination-32B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  14. ^ "THUDM/GLM-4-9B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  15. ^ "THUDM/GLM-Z1-9B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.
  16. ^ "THUDM/GLM-4-9B-0414 · Base model". huggingface.co. 2025-04-14. Retrieved 2025-04-29.
  17. ^ "GLM-130B Application Form". Google Docs. Retrieved 2025-04-29.
  18. ^ THUDM/GLM, Z.ai & THUKEG, 2025-04-27, retrieved 2025-04-29
  19. ^ "MODEL_LICENSE · THUDM/chatglm-6b at main". huggingface.co. 2023-07-08. Retrieved 2025-04-29.
  20. ^ "LICENSE · THUDM/glm-4-9b at main". huggingface.co. 2024-06-08. Retrieved 2025-04-29.
  21. ^ "THUDM/GLM-4-32B-0414". huggingface.co. 2025-04-15. Retrieved 2025-04-29.