A series of math-specific large language models of our Qwen2 series
GLM-4 series: Open Multilingual Multimodal Chat LMs
Learning to Reason with Search for LLMs via Reinforcement Learning
Data Infrastructure providing an approach to multimodal AI workloads
Autoregressive Model Beats Diffusion
Neural Network architecture based on ideas of the original LSTM
TigerBot: A multi-language multi-task LLM
Run PyTorch LLMs locally on servers, desktop and mobile
StarVector is a foundation model for SVG generation
Implement CPU from scratch and play with large model deployments
Diversity-driven optimization and large-model reasoning ability
Run LLMs locally on Cloud Workstations
Unify Efficient Fine-tuning of RAG Retrieval, including Embedding
The unofficial python package that returns response of Google Bard
Visual Instruction Tuning: Large Language-and-Vision Assistant
Run Mixtral-8x7B models in Colab or consumer desktops
Open-source, high-performance Mixture-of-Experts large language model
Repo for YaYi Chinese LLMs based on LlaMA2 & BLOOM
Serving multiple LoRA finetuned LLM as one
Open-source tool to visualise your RAG
Official release of InternLM series
Framework that is dedicated to making neural data processing
Building Mixture-of-Experts from LLaMA with Continual Pre-training
Database system for building simpler and faster AI-powered application
A repository that contains models, datasets, and fine-tuning