ChatGLM3 is ZhipuAI & Tsinghua KEG’s third-gen conversational model suite centered on the 6B-parameter ChatGLM3-6B. It keeps the series’ smooth dialog and low deployment cost while adding native tool use (function calling), a built-in code interpreter, and agent-style workflows. The family includes base and long-context variants (8K/32K/128K). The repo ships Python APIs, CLI and web demos (Gradio/Streamlit), an OpenAI-format API server, and a compact fine-tuning kit. Quantization (4/8-bit), CPU/MPS support, and accelerator backends (TensorRT-LLM, OpenVINO, chatglm.cpp) enable lightweight local or edge deployment.
Features
- Multi-turn chat with native tools: Function Call, Code Interpreter, agent tasks.
- Long-context models: 8K, 32K, and 128K sequence lengths.
- Easy local deployment: INT4/INT8 quantization, CPU & Apple MPS support.
- Turnkey demos: CLI, Gradio/Streamlit web UIs, OpenAI-style REST server.
- Accelerator integrations: TensorRT-LLM, OpenVINO, chatglm.cpp, multi-GPU.
- Fine-tuning kit: LoRA/QLoRA/PTuning examples and scripts.
Categories
Large Language Models (LLM)License
Apache License V2.0Follow ChatGLM3
Other Useful Business Software
Gen AI apps are built with MongoDB Atlas
MongoDB Atlas is the developer-friendly database used to build, scale, and run gen AI and LLM-powered apps—without needing a separate vector database. Atlas offers built-in vector search, global availability across 115+ regions, and flexible document modeling. Start building AI apps faster, all in one place.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of ChatGLM3!