ChatGLM.cpp is a C++ implementation of the ChatGLM-6B model, enabling efficient local inference without requiring a Python environment. It is optimized for running on consumer hardware.
Features
- Provides a C++ implementation of ChatGLM-6B
- Supports running models on CPU and GPU
- Optimized for low-memory hardware and edge devices
- Allows quantization for reduced resource consumption
- Works as a lightweight alternative to Python-based inference
- Offers real-time chatbot capabilities
License
MIT LicenseFollow ChatGLM.cpp
Other Useful Business Software
Full-stack observability with actually useful AI | Grafana Cloud
Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of ChatGLM.cpp!