Showing 82 open source projects for "llm"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 1
    MNN

    MNN

    MNN is a blazing fast, lightweight deep learning framework

    MNN is a highly efficient and lightweight deep learning framework. It supports inference and training of deep learning models, and has industry leading performance for inference and training on-device. At present, MNN has been integrated in more than 20 apps of Alibaba Inc, such as Taobao, Tmall, Youku, Dingtalk, Xianyu and etc., covering more than 70 usage scenarios such as live broadcast, short video capture, search recommendation, product searching by image, interactive marketing, equity...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 2
    Kong

    Kong

    The Cloud-Native API Gateway

    Kong is a next generation cloud-native API platform for multi-cloud and hybrid organizations. When building for the web, mobile, or Internet of Things, you’ll need a common functionality to run your software, and Kong is that solution. Kong acts as a gateway, connecting microservices requests and APIs natively while also providing load balancing, logging, monitoring, authentication, rate-limiting, and so much more through plugins. Kong is highly extensible as well as platform agnostic,...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    Groq Python

    Groq Python

    The official Python Library for the Groq API

    Groq Python is the official Python SDK for the Groq REST API, giving Python developers straightforward access to Groq’s LLM, chat, audio, and other AI services. Through this library, you can call Groq’s models from Python code — for example to request chat completions, code generation, transcription, or any supported endpoint — using idiomatic Python syntax. The SDK handles authentication (via environment variable or parameter), defines proper type-safe request/response data types, and supports both synchronous and asynchronous usage patterns depending on your application needs. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    Page Agent

    Page Agent

    JavaScript in-page GUI agent. Control web interfaces

    ...Page Agent is designed to integrate seamlessly into existing web applications, making it possible to embed AI copilots into SaaS platforms without major backend changes. It supports a bring-your-own-LLM approach, allowing developers to connect their preferred language models to power the agent’s reasoning capabilities.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    KServe

    KServe

    Standardized Serverless ML Inference Platform on Kubernetes

    KServe provides a Kubernetes Custom Resource Definition for serving machine learning (ML) models on arbitrary frameworks. It aims to solve production model serving use cases by providing performant, high abstraction interfaces for common ML frameworks like Tensorflow, XGBoost, ScikitLearn, PyTorch, and ONNX. It encapsulates the complexity of autoscaling, networking, health checking, and server configuration to bring cutting edge serving features like GPU Autoscaling, Scale to Zero, and...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    DALI

    DALI

    A GPU-accelerated library containing highly optimized building blocks

    The NVIDIA Data Loading Library (DALI) is a library for data loading and pre-processing to accelerate deep learning applications. It provides a collection of highly optimized building blocks for loading and processing image, video and audio data. It can be used as a portable drop-in replacement for built-in data loaders and data iterators in popular deep learning frameworks. Deep learning applications require complex, multi-stage data processing pipelines that include loading, decoding,...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    AI Commits

    AI Commits

    A CLI that writes your git commit messages for you with AI

    AI Commits is a command-line tool that writes your git commit messages for you using an AI model. It works by running git diff to gather your staged code changes, sending that diff to an LLM (originally GPT-3, now configurable), and receiving back a concise, human-readable commit message. The tool is designed to integrate cleanly into a developer’s workflow so that generating a descriptive commit message becomes a single command rather than a chore. It supports configuration via environment variables or config files so you can set your API key, preferred model, message style, and more. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Ray

    Ray

    A unified framework for scalable computing

    Modern workloads like deep learning and hyperparameter tuning are compute-intensive and require distributed or parallel execution. Ray makes it effortless to parallelize single machine code — go from a single CPU to multi-core, multi-GPU or multi-node with minimal code changes. Accelerate your PyTorch and Tensorflow workload with a more resource-efficient and flexible distributed execution framework powered by Ray. Accelerate your hyperparameter search workloads with Ray Tune. Find the best...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    LangChain for .NET

    LangChain for .NET

    C# implementation of LangChain

    LangChain .NET is a C# implementation of the LangChain framework that enables developers to build LLM-powered applications using the .NET ecosystem. The project aims to replicate the core abstractions of LangChain, such as chains, agents, and vector stores, while adapting them to the conventions and strengths of C#. It emphasizes composability, allowing developers to build complex workflows by combining modular components that interact with language models and external data sources. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    LangChainGo

    LangChainGo

    LangChain for Go, the easiest way to write LLM-based programs in Go

    LangChainGo is a Go-based implementation of the LangChain framework, designed to help developers build applications powered by large language models using the Go programming language. It provides a modular architecture that allows developers to combine components such as language models, chains, agents, memory systems, and vector stores into flexible workflows. The framework emphasizes composability, making it easy to create complex pipelines that integrate LLMs with external data sources,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Markdown Site

    Markdown Site

    An open-source publishing framework built for AI agents and developers

    ...By leveraging Convex for backend real-time data management and Netlify for static hosting, markdown-site enables rich publishing features like SEO optimization, full-text search, analytics dashboards, and AI-indexed content tailored to LLM workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Superduper

    Superduper

    Superduper: Integrate AI models and machine learning workflows

    Superduper is a Python-based framework for building end-2-end AI-data workflows and applications on your own data, integrating with major databases. It supports the latest technologies and techniques, including LLMs, vector-search, RAG, and multimodality as well as classical AI and ML paradigms. Developers may leverage Superduper by building compositional and declarative objects that out-source the details of deployment, orchestration versioning, and more to the Superduper engine. This...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    BentoML

    BentoML

    Unified Model Serving Framework

    BentoML simplifies ML model deployment and serves your models at a production scale. Support multiple ML frameworks natively: Tensorflow, PyTorch, XGBoost, Scikit-Learn and many more! Define custom serving pipeline with pre-processing, post-processing and ensemble models. Standard .bento format for packaging code, models and dependencies for easy versioning and deployment. Integrate with any training pipeline or ML experimentation platform. Parallelize compute-intense model inference...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    rollama

    rollama

    Wrap the Ollama API, which allows you to run different LLMs

    ...It supports common LLM tasks such as text generation, annotation, and embedding creation, making it useful for tasks like document analysis and data labeling. The design mirrors familiar R workflows, allowing users to integrate AI capabilities into scripts, notebooks, and data pipelines with minimal friction. It also provides flexibility to extend functionality to any feature supported by the underlying Ollama API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Unsloth-MLX

    Unsloth-MLX

    Bringing the Unsloth experience to Mac users via Apple's MLX framework

    Unsloth-MLX offers developers the power of Unsloth’s efficient large language model fine-tuning experience on Apple Silicon Macs by wrapping Apple’s native MLX framework with an API fully compatible with Unsloth workflows. This project removes traditional barriers that prevent Mac users from prototyping and experimenting with LLM training locally by allowing the same code used in cloud GPU environments to run on M-series hardware, improving workflow continuity and reducing iteration costs. It supports loading and training Hugging Face models with fine-tuning strategies like SFT, DPO, ORPO, and GRPO and even handles exporting models to formats like GGUF for downstream use, although some limitations apply with quantized models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Open LLMs

    Open LLMs

    A list of open LLMs available for commercial use

    ...It aggregates metadata, licensing info, and often pointers to the model weights or model cards — helping users quickly compare models by size, license, domain, and capabilities. By compiling this in one place, open-llms reduces friction in exploring the LLM space, making it easier to try different models, benchmark them, or build custom applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Seldon Core

    Seldon Core

    An MLOps framework to package, deploy, monitor and manage models

    The de facto standard open-source platform for rapidly deploying machine learning models on Kubernetes. Seldon Core, our open-source framework, makes it easier and faster to deploy your machine learning models and experiments at scale on Kubernetes. Seldon Core serves models built in any open-source or commercial model building framework. You can make use of powerful Kubernetes features like custom resource definitions to manage model graphs. And then connect your continuous integration and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Gen.jl

    Gen.jl

    A general-purpose probabilistic programming system

    An open-source stack for generative modeling and probabilistic inference. Gen’s inference library gives users building blocks for writing efficient probabilistic inference algorithms that are tailored to their models, while automating the tricky math and the low-level implementation details. Gen helps users write hybrid algorithms that combine neural networks, variational inference, sequential Monte Carlo samplers, and Markov chain Monte Carlo. Gen features an easy-to-use modeling language...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    SageMaker Hugging Face Inference Toolkit

    SageMaker Hugging Face Inference Toolkit

    Library for serving Transformers models on Amazon SageMaker

    SageMaker Hugging Face Inference Toolkit is an open-source library for serving Transformers models on Amazon SageMaker. This library provides default pre-processing, predict and postprocessing for certain Transformers models and tasks. It utilizes the SageMaker Inference Toolkit for starting up the model server, which is responsible for handling inference requests. For the Dockerfiles used for building SageMaker Hugging Face Containers, see AWS Deep Learning Containers. The SageMaker Hugging...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    DocsGPT

    DocsGPT

    Private AI platform for agents, enterprise search and RAG pipelines

    ...Connect any data source (PDFs, DOCX, CSV, Excel, HTML, audio, GitHub, databases, URLs) and get accurate, hallucination-free answers with source citations. Choose your LLM: OpenAI, Anthropic, Google Gemini, or local models. Works with Qdrant, MongoDB, and Elasticsearch and more. Deploy via Docker or Kubernetes with full data sovereignty. Build embeddable chat and search widgets, automate multi-step workflows with AI agents, and integrate via Slack, Telegram, Discord, or REST API. Enterprise features include RBAC, 99.9% uptime SLA, and dedicated support. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    Tambo

    Tambo

    Add generative UI components to your AI assistant, copilot, or agent

    ...Developers use Tambo to shift UI logic toward the AI model: instead of hardcoding UI flows, the AI can decide what component to show next. Tambo also supports streaming updates (i.e. progressively rendering UI) and embedding interactions between LLM outputs and front-end state.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    MegEngine

    MegEngine

    Easy-to-use deep learning framework with 3 key features

    MegEngine is a fast, scalable and easy-to-use deep learning framework with 3 key features. You can represent quantization/dynamic shape/image pre-processing and even derivation in one model. After training, just put everything into your model and inference it on any platform at ease. Speed and precision problems won't bother you anymore due to the same core inside. In training, GPU memory usage could go down to one-third at the cost of only one additional line, which enables the DTR...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    1Panel

    1Panel

    1Panel provides an intuitive web interface and MCP Server

    1Panel is a comprehensive Linux server management dashboard and MCP server built in Go. It offers UI control over websites, containers, databases, file systems, LLMs, backups, and monitoring, streamlining typical admin workflows via web.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    TensorZero

    TensorZero

    TensorZero is an open-source stack for industrial-grade LLM apps

    tensorzero is a lightweight C++ library designed for tensor operations and numerical computing. It offers a minimal and readable implementation of core tensor functionality, making it ideal for educational purposes, lightweight applications, or those wanting to understand how tensor libraries work under the hood. With no external dependencies, tensorzero is easy to integrate into C++ projects needing basic multi-dimensional array support.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    AWS Neuron

    AWS Neuron

    Powering Amazon custom machine learning chips

    AWS Neuron is a software development kit (SDK) for running machine learning inference using AWS Inferentia chips. It consists of a compiler, run-time, and profiling tools that enable developers to run high-performance and low latency inference using AWS Inferentia-based Amazon EC2 Inf1 instances. Using Neuron developers can easily train their machine learning models on any popular framework such as TensorFlow, PyTorch, and MXNet, and run it optimally on Amazon EC2 Inf1 instances. You can...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB