Full stack AI software engineer
Hypernetworks that adapt LLMs for specific benchmark tasks
Structured RAG: ingest, index, query
Specify a github or local repo, github pull request
Performance-optimized AI inference on your GPUs
Dealing with all unstructured data, such as reverse image search
Harness LLMs with Multi-Agent Programming
Agents write python code to call tools and orchestrate other agents
Run LLM prompts from your shell
Qwen3-Coder is the code version of Qwen3
Framework for building, orchestrating, and deploying AI agents
Module for automatic summarization of text documents and HTML pages
Implement CPU from scratch and play with large model deployments
Intelligent automation and multi-agent orchestration for Claude Code
OCR expert VLM powered by Hunyuan's native multimodal architecture
MCP server enabling AI agents to control and automate Windows OS
Making RAG Simpler with Small and Open-Sourced Language Models
AI-Driven Exploration in the Space of Code
New family of code large language models (LLMs)
Inference Llama 2 in one file of pure C
Collect, organize, use, and share, all in OmniBox
ChatGLM2-6B: An Open Bilingual Chat LLM
Tiny vision language model
The Modular Platform (includes MAX & Mojo)
The behavior guidance framework for customer-facing LLM agents