GLM-130B

Open Bilingual 130B Parameter Pre-Trained Model – Strong English and Chinese Language Understanding from Tsinghua KEG
Last Updated: January 14, 2026
By Zelili AI

About This AI

GLM-130B is an open bilingual (English and Chinese) bidirectional dense pre-trained language model with 130 billion parameters, developed by Tsinghua KEG and collaborators.

Released in 2022 as part of the General Language Model (GLM) series, it was pre-trained using the GLM algorithm on a massive corpus of roughly 400 billion tokens (200 billion each for English and Chinese).

The model demonstrates strong few-shot learning capabilities across various tasks including question answering, sentiment analysis, natural language inference, and more, often matching or approaching the performance of much larger closed models at the time.

It supports bilingual understanding with high proficiency in both languages, making it suitable for cross-lingual applications and Chinese-dominant tasks where many Western models fall short.

The project was conceived in late 2021, trained on sponsored clusters starting May 2022, and open-sourced under a permissive license with weights available on Hugging Face and GitHub.

While superseded by newer models in the GLM/ChatGLM series (like GLM-4), GLM-130B remains a landmark open model for its scale and bilingual focus, influencing subsequent research in large-scale pre-training.

It requires significant hardware (e.g., multiple high-end GPUs) for inference due to its size but can be quantized or run with optimizations.

The Hugging Face Space (zai-org/GLM-130B) provides a demo interface, though it may be limited or archived given the model’s age and evolution to newer GLM versions.

Key Features

  1. Bilingual pre-training: Balanced high-quality training on English and Chinese corpora for strong performance in both languages
  2. 130 billion parameters: Dense bidirectional architecture enabling advanced few-shot and zero-shot learning
  3. General Language Model algorithm: Uses GLM pre-training objective for better masked modeling and generation
  4. Few-shot learning excellence: Competitive results on SuperGLUE, MMLU-like tasks, and cross-lingual benchmarks at release
  5. Open weights and code: Fully available on Hugging Face and GitHub for research and fine-tuning
  6. High Chinese proficiency: Superior handling of Chinese language tasks compared to many English-centric models
  7. Support for downstream tasks: Adaptable for QA, classification, translation, summarization, and more

Price Plans

  1. Free ($0): Fully open-source model weights, code, and documentation available on Hugging Face and GitHub; no costs for download or local use
  2. Cloud Inference (Variable): Potential paid hosted inference via third-party platforms if available (not native)

Pros

  1. Landmark open model: One of the first massive open bilingual LLMs, inspiring the GLM/ChatGLM family
  2. Strong bilingual capabilities: Excellent English-Chinese balance and cross-lingual transfer
  3. Permissive open-source license: Allows free research, commercial fine-tuning, and deployment
  4. Influential in Chinese AI: Helped advance open large model development in China
  5. Proven few-shot performance: Matched closed models of similar era on diverse benchmarks

Cons

  1. Outdated by 2026 standards: Superseded by newer GLM-4, Llama, Qwen, etc., with better efficiency and capabilities
  2. Extremely resource-intensive: 130B size requires massive GPU memory for full inference
  3. Limited modern usage: Most users have migrated to smaller/faster successors like ChatGLM-6B or GLM-4
  4. No native chat tuning: Base pre-trained model; lacks instruction/chat fine-tuning of later versions
  5. Demo space may be inactive: Hugging Face Space (zai-org/GLM-130B) shows limited or no recent activity
  6. No recent updates: Original 2022 release with no major follow-up maintenance

Use Cases

  1. Bilingual research: Studying cross-lingual transfer and Chinese NLP with a large open model
  2. Fine-tuning experiments: Adapting for domain-specific tasks in English or Chinese
  3. Historical benchmarking: Comparing performance of 2022-era open LLMs
  4. Academic studies: Analyzing pre-training scaling laws and bilingual data effects
  5. Legacy applications: Running in environments where newer models are too resource-heavy

Target Audience

  1. AI researchers in NLP: Studying large-scale bilingual pre-training and GLM architecture
  2. Chinese language model developers: Building upon early open bilingual foundations
  3. Academic institutions: Using for papers, theses, and experiments on open LLMs
  4. Open-source enthusiasts: Exploring historical milestones in massive open models

How To Use

  1. Download from Hugging Face: Access model at THUDM/GLM-130B or similar repo
  2. Install dependencies: Use transformers library with trust_remote_code=True
  3. Load model: model = AutoModel.from_pretrained('THUDM/glm-130b', trust_remote_code=True)
  4. Run inference: Use pipeline or custom generation for text completion/filling
  5. Quantize if needed: Apply 8-bit or 4-bit quantization for reduced memory
  6. Fine-tune locally: Use PEFT or full fine-tuning on domain data
  7. Demo via space: Try the Hugging Face Space zai-org/GLM-130B if active

How we rated GLM-130B

  • Performance: 4.0/5
  • Accuracy: 4.2/5
  • Features: 3.8/5
  • Cost-Efficiency: 5.0/5
  • Ease of Use: 3.5/5
  • Customization: 4.5/5
  • Data Privacy: 5.0/5
  • Support: 3.8/5
  • Integration: 4.2/5
  • Overall Score: 4.1/5

GLM-130B integration with other tools

  1. Hugging Face Transformers: Native support via AutoModel and pipeline for easy inference
  2. GitHub Repository: Original code, training details, and evaluation scripts from Tsinghua KEG
  3. Local Hardware: Runs on multi-GPU setups with frameworks like DeepSpeed or Megatron
  4. Research Frameworks: Compatible with PEFT, LoRA, and other fine-tuning libraries
  5. Older Demo Spaces: Hugging Face Space zai-org/GLM-130B for online testing if still functional

Best prompts optimised for GLM-130B

  1. GLM-130B is a base pre-trained model, not instruction-tuned. Best use is for masked language modeling or continuation: Fill in the blank: The capital of France is ___
  2. As a bilingual model, prompt in Chinese: 北京是中国的___,请完成句子。
  3. For few-shot learning: Example: Input: Apple is a fruit. Output: Fruit. Input: Car is a ___ Output: Vehicle. Now: Dog is a ___
  4. English continuation: Once upon a time in a faraway land, there lived a brave knight who...
  5. Chinese generation: 请写一段关于秋天的优美描写:秋风萧瑟,落叶纷飞,...
GLM-130B was a pioneering open bilingual 130B model in 2022, offering strong English-Chinese performance and few-shot learning. Fully free and influential, it paved the way for later GLM series. By 2026 it’s outdated for most uses due to newer efficient models, but remains valuable for research and historical comparisons.

FAQs

  • What is GLM-130B?

    GLM-130B is an open bilingual (English and Chinese) 130 billion parameter pre-trained language model released by Tsinghua KEG in 2022, known for strong few-shot learning and bilingual capabilities.

  • When was GLM-130B released?

    It was conceived in late 2021, training started May 2022, and officially open-sourced on August 4, 2022.

  • Is GLM-130B free to use?

    Yes, it is completely open-source with weights available on Hugging Face and GitHub under a permissive license; no costs for download or local use.

  • Who developed GLM-130B?

    Developed by Tsinghua University’s Knowledge Engineering Group (KEG) with collaborators including Zhipu.AI.

  • What are the main strengths of GLM-130B?

    Excellent bilingual performance, strong few-shot/zero-shot learning, and competitive results on benchmarks at the time of release.

  • Is GLM-130B still relevant in 2026?

    It has been largely superseded by newer models like GLM-4, Llama 3, Qwen, etc., but remains useful for research and bilingual studies.

  • How much hardware does GLM-130B require?

    Full inference needs multiple high-end GPUs (e.g., 8x A100 40GB or equivalent); quantization or CPU offloading can reduce requirements.

  • Where can I try GLM-130B?

    Check the Hugging Face Space zai-org/GLM-130B for a demo (may be limited), or run locally via transformers library.

Newly Added Tools​

Qwen-Image-2.0

$0/Month

Qodo AI

$0/Month

Codiga

$10/Month

Tabnine

$59/Month
GLM-130B Alternatives

Cognosys AI

$0/Month

AI Perfect Assistant

$17/Month

Intern-S1-Pro

$0/Month

About Author

Hi Guys! We are a group of ML Engineers by profession with years of experience exploring and building AI tools, LLMs, and generative technologies. We analyze new tools not just as a user, but as someone who understands their technical depth and real-world value.We know how overwhelming these tools can be for most people, that’s why we break down complex AI concepts into simple, practical insights. Our goal is to help you discover these magical AI tools that actually save your time and make everyday work smarter, not harder.“We don’t just write about AI: We build, test and simplify it for you.”