Structured outputs for llms
Python bindings for llama.cpp
Run Local LLMs on Any Device. Open-source
Agentic, Reasoning, and Coding (ARC) foundation models
A high-throughput and memory-efficient inference and serving engine
Port of Facebook's LLaMA model in C/C++
Low-code app builder for RAG and multi-agent AI applications
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Qwen3 is the large language model series developed by Qwen team
Access large language models from the command-line
Operating LLMs in production
A guidance language for controlling large language models
PandasAI is a Python library that integrates generative AI
lightweight package to simplify LLM API calls
The unofficial python package that returns response of Google Bard
Inference code for CodeLlama models
Powerful AI language model (MoE) optimized for efficiency/performance
Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon
Database system for building simpler and faster AI-powered application
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Interact with your documents using the power of GPT
Simple, Pythonic building blocks to evaluate LLM applications
Qwen3-Coder is the code version of Qwen3
Open-source, high-performance AI model with advanced reasoning
Framework that is dedicated to making neural data processing