Run LLMs locally on Cloud Workstations
ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat
Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge
Interact with your documents using the power of GPT
Automatic question answering for local knowledge bases based on LLM
Chinese Llama-3 LLMs) developed from Meta Llama 3
AirLLM 70B inference with single 4GB GPU
Chat with any codebase in under two minutes | Fully local
Quick illustration of how one can easily read books together with LLMs
ChatGLM2-6B: An Open Bilingual Chat LLM
Open source libraries and APIs to build custom preprocessing pipelines
Qwen2.5-VL is the multimodal large language model series
Universal LLM Deployment Engine with ML Compilation
Unified framework for building enterprise RAG pipelines
Retrieval Augmented Generation (RAG) framework
local-first semantic code search engine
Accelerate local LLM inference and finetuning
GitLab automatic code review tool based on large models
Chinese and English multimodal conversational language model
GLM-4 series: Open Multilingual Multimodal Chat LMs
GPT4V-level open-source multi-modal model based on Llama3-8B
Visual Instruction Tuning: Large Language-and-Vision Assistant
Ray Aviary - evaluate multiple LLMs easily
Inference code and configs for the ReplitLM model family
Langchain Apps on Production with Jina & FastAPI