Open Source Rust Artificial Intelligence Software - Page 4

Rust Artificial Intelligence Software

View 13471 business solutions

Browse free open source Rust Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Rust Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    Luminal

    Luminal

    Deep learning at the speed of light

    Luminal is a framework designed to accelerate and simplify the development of systems-level data applications by using a typed, functional, and streaming-first approach. Instead of treating data processing as a series of ad-hoc scripts, Luminal models transformations as strongly typed building blocks that can be composed into reliable, scalable pipelines. The project emphasizes correctness and performance by requiring explicit types for the data flowing through transformations, reducing runtime surprises and allowing for highly optimized execution. It is particularly well-suited for data engineering workflows where large datasets must be processed incrementally, efficiently, and deterministically. The framework also includes a runtime capable of executing pipelines across multiple backends, making it flexible in cloud and local environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    MagicAPI AI Gateway

    MagicAPI AI Gateway

    Built for demanding AI workflows

    The world's fastest AI Gateway proxy, written in Rust and optimized for maximum performance. This high-performance API gateway routes requests to various AI providers (OpenAI, GROQ) with streaming support, making it perfect for developers who need reliable and blazing-fast AI API access.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    ModelFox

    ModelFox

    ModelFox makes it easy to train, deploy, and monitor ML models

    ModelFox makes it easy to train, deploy, and monitor machine learning models. Train a model from a CSV file on the command line. Make predictions from Elixir, Go, JavaScript, PHP, Python, Ruby, or Rust. Learn about your models and monitor them in production from your browser. ModelFox makes it easy to train, deploy, and monitor machine learning models. You can install the modelfox CLI by either downloading the binary from the latest GitHub release or by building from source. Train a machine learning model by running modelfox train with the path to a CSV file and the name of the column you want to predict. The CLI automatically transforms your data into features, trains a number of linear and gradient boosted decision tree models to predict the target column, and writes the best model to a .modelfox file. If you want more control, you can provide a config file.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Monoio

    Monoio

    Rust async runtime based on io-uring

    Monoio is a Rust asynchronous runtime designed for high-performance I/O-bound servers and applications, built around native OS async I/O primitives (e.g. io_uring on Linux, epoll / kqueue on other Unix-like systems), rather than layering atop an existing runtime. Its design philosophy centers on a “thread-per-core” model where each core runs its own event loop, minimizing cross-thread synchronization needs, avoiding the overhead and complexity of task scheduling, and letting developers write efficient, low-overhead asynchronous networking or I/O code. Because tasks do not need to be Send or Sync and can make use of thread-local data safely, Monoio simplifies certain concurrency paradigms while delivering performance benefits for workloads like high-throughput network servers, proxies, or real-time services. The runtime includes abstractions for async sockets, readers/writers, TCP/UDP networking, and compatibility layers (macros, crates) to ease adoption.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Monty

    Monty

    A minimal, secure Python interpreter written in Rust for use by AI

    Monty is an experimental, security-focused Python interpreter implemented in Rust and intended for running AI-generated Python safely under strict constraints. The project’s core goal is to enable code execution in environments where untrusted or model-produced code must be tightly sandboxed to reduce risk. Rather than offering a full “general-purpose Python runtime with everything enabled,” Monty is designed to be minimal and controlled, making it easier to reason about what code can do and what it cannot. It prioritizes guardrails like resource limits and restricted capabilities, which is especially useful for agentic workflows that need to execute small pieces of Python for data transforms, validation, or tool-like computations. Because it’s written in Rust, it’s positioned to deliver a compact, portable runtime that can be embedded into larger systems that need dependable isolation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    OxidizeBot

    OxidizeBot

    High performance Twitch bot in Rust

    OxidizeBot is an open-source Twitch Bot empowering you to focus on what's important. It allows for a richer interaction between you and your chat. From a song request system to groundbreaking game modes where your viewers can interact directly with you and your game. It's written in Rust, providing an unparalleled level of reliability and performance. OxidizeBot doesn't cost you anything, and its source code is available on GitHub for anyone to tinker with! Plays music moderates your chat, plays games, you name it! You own your data. It uses your internet for the best possible latency. It's light on system resources*. And running locally means it can perform rich interactions with your games like Chaos%.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Paddler

    Paddler

    Open-source LLM load balancer and serving platform for hosting LLMs

    Paddler is an open-source LLM infrastructure platform designed to deploy, manage, and scale large language models on private infrastructure. The system acts as a specialized load balancer and serving layer for language models, enabling organizations to run inference workloads without relying on external API providers. It supports running models locally through engines such as llama.cpp while distributing requests across multiple compute nodes to improve performance and reliability. The architecture is designed with privacy and cost control in mind, making it suitable for organizations that handle sensitive data or require predictable operational costs. Paddler also includes tools for monitoring, request buffering, and autoscaling integration so that deployments can adapt dynamically to changing workloads. A built-in administrative interface allows developers and operations teams to manage models, observe system performance, and test inference endpoints.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Phantasm

    Phantasm

    Toolkits to create a human-in-the-loop approval layer

    Phantasm offers toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents' workflows in real-time, ensuring safety and reliability in AI operations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    PostgresML

    PostgresML

    The GPU-powered AI application database

    PostgresML is a complete platform in a PostgreSQL extension. Build simpler, faster, and more scalable models right inside your database. Explore the SDK and test open source models in our hosted database. Combine and automate the entire workflow from embedding generation to indexing and querying for the simplest (and fastest) knowledge-based chatbot implementation. Leverage multiple types of natural language processing and machine learning models such as vector search and personalization with embeddings to improve search results. Leverage your data with time series forecasting to garner key business insights. Build statistical and predictive models with the full power of SQL and dozens of regression algorithms. Return results and detect fraud faster with ML at the database layer. PostgresML abstracts the data management overhead from the ML/AI lifecycle by enabling users to run ML/LLM models directly on a Postgres database.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    Refact Agent

    Refact Agent

    WebUI for Fine-Tuning and Self-hosting of Open-Source LLMs

    Refact is an AI-powered code assistant designed to enhance software development workflows. It integrates with code editors and provides suggestions, refactoring assistance, and debugging insights.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Rig

    Rig

    Rust framework for building modular and scalable LLM-powered apps

    Rig is an open source Rust framework designed to help developers build modular and scalable applications powered by large language models. It provides a unified set of abstractions that allow applications to interact with many AI model providers and vector databases through a single interface. Its architecture emphasizes modularity, enabling developers to integrate only the components and integrations they need for a specific application. Rig includes built-in support for agent workflows, allowing systems to perform multi-turn reasoning, tool calling, and retrieval-based tasks within structured pipelines. It also supports capabilities such as text generation, embeddings, transcription, image generation, and audio generation depending on the provider used. Developers can integrate language models into their software with minimal boilerplate while maintaining flexibility for complex AI workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Rivet

    Rivet

    Visual AI IDE for building agents with prompt chains and graphs

    Rivet is an open source visual AI programming environment designed to help developers build complex AI agents using a node-based interface and prompt chaining workflows. It provides a desktop application that allows users to visually construct and debug AI logic as interconnected graphs, making it easier to manage sophisticated interactions between language models and external tools. Rivet also includes a TypeScript library that enables these visual graphs to be executed and integrated directly into applications, bridging the gap between prototyping and production use. Rivet supports multiple large language model providers and integrates with services such as embeddings and transcription systems, allowing developers to create richer AI-powered features. Its architecture emphasizes composability, where different components like prompts, APIs, and data processing steps can be combined into reusable pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Rust Docs MCP Server

    Rust Docs MCP Server

    Prevents outdated Rust code suggestions from AI assistants

    The Rust Docs MCP Server fetches documentation for specified Rust crates, generates embeddings for the content, and provides an MCP tool to answer questions about the crate based on the documentation context. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Rust Telegram Bot Library

    Rust Telegram Bot Library

    Rust Library for creating a Telegram Bot

    A library for writing your own Telegram bots.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    SIG Rust

    SIG Rust

    Rust language bindings for TensorFlow

    SIG Rust provides idiomatic Rust bindings for TensorFlow, making it possible for developers to work with TensorFlow functionality from within the Rust programming language. Rather than replacing TensorFlow itself, it acts as an integration layer that connects Rust applications to the TensorFlow C API. The repository is designed for developers who want Rust’s performance, safety, and systems programming strengths while still accessing TensorFlow’s machine learning capabilities. It includes setup instructions that explain how the crate can automatically download or compile the required TensorFlow shared libraries, which lowers the barrier to getting started. The project also supports environments where TensorFlow is already installed, giving developers more flexibility in how they configure their systems. Documentation, community discussion resources, and versioned releases indicate that the repository is maintained as a serious language binding.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    Savant

    Python Computer Vision & Video Analytics Framework With Batteries Incl

    Savant is an open-source, high-level framework for building real-time, streaming, highly efficient multimedia AI applications on the Nvidia stack. It helps to develop dynamic, fault-tolerant inference pipelines that utilize the best Nvidia approaches for data center and edge accelerators. Savant is built on DeepStream and provides a high-level abstraction layer for building inference pipelines. It is designed to be easy to use, flexible, and scalable. It is a great choice for building smart CV and video analytics applications for cities, retail, manufacturing, and more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    SmartGPT

    SmartGPT

    A program that provides LLMs with ability to complete complex tasks

    SmartGPT is an experimental autonomous agent framework built to help large language models tackle complex tasks with minimal or no additional user input. It works by decomposing larger objectives into smaller steps and gathering information from the internet and other outside sources as needed. The project is written in Rust and emphasizes modularity, allowing developers to compose different “Autos” depending on the workflow they want to build. Its architecture separates responsibility between a dynamic agent that reasons about what to do next and a static agent that plans and executes tool chains in a defined order. The repository describes this approach as a way to improve flexibility and consistency compared with simpler agent loops, while still acknowledging that the project is highly experimental and not focused on backward compatibility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Spice.ai OSS

    Spice.ai OSS

    A self-hostable CDN for databases

    Spice is a portable runtime offering developers a unified SQL interface to materialize, accelerate, and query data from any database, data warehouse, or data lake. Spice connects, fuses, and delivers data to applications, machine-learning models, and AI backends, functioning as an application-specific, tier-optimized Database CDN. The Spice runtime, written in Rust, is built-with industry-leading technologies such as Apache DataFusion, Apache Arrow, Apache Arrow Flight, SQLite, and DuckDB. Spice makes it easy and fast to query data from one or more sources using SQL. You can co-locate a managed dataset with your application or machine learning model, and accelerate it with Arrow in-memory, SQLite/DuckDB, or with attached PostgreSQL for fast, high-concurrency, low-latency queries. Accelerated engines give you flexibility and control over query cost and performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to integrate embedding capabilities into applications without managing model internals directly. Text Embeddings Inference is optimized for throughput and low latency, enabling it to handle large volumes of requests reliably. It also emphasizes ease of deployment, often using containerization and configurable runtime options to adapt to different infrastructure setups.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Tokenizers

    Tokenizers

    Fast State-of-the-Art Tokenizers optimized for Research and Production

    Fast State-of-the-art tokenizers, optimized for both research and production. Tokenizers provides an implementation of today’s most used tokenizers, with a focus on performance and versatility. These tokenizers are also used in Transformers. Train new vocabularies and tokenize, using today’s most used tokenizers. Extremely fast (both training and tokenization), thanks to the Rust implementation. Takes less than 20 seconds to tokenize a GB of text on a server’s CPU. Easy to use, but also extremely versatile. Designed for both research and production. Full alignment tracking. Even with destructive normalization, it’s always possible to get the part of the original sentence that corresponds to any token. Does all the pre-processing: Truncation, Padding, add the special tokens your model needs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Tokscale

    Tokscale

    A CLI tool for tracking token usage from OpenCode, Claude Code

    Tokscale is a CLI and terminal UI tool that tracks token usage and estimated cost across multiple AI coding assistants and development workflows. It treats tokens like a measurable resource, helping developers understand how much “AI energy” they are consuming over time and where it is being spent. The tool aggregates usage across supported platforms and presents it through interactive views that let users filter, sort, and explore trends without leaving the terminal. Tokscale also includes rich visualizations such as contribution-graph style summaries and daily breakdowns, making it easy to spot spikes, habits, and long-term patterns. For cost estimation, it can map usage to pricing data so developers can see what their activity implies financially across different models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    VectorChord

    VectorChord

    Scalable, fast, and disk-friendly vector search in Postgres

    VectorChord is an open-source vector database built for local and edge deployment. It supports efficient vector indexing and retrieval using ANN (approximate nearest neighbor) algorithms and is optimized for integration with LLM and AI applications. VectorChord is lightweight and can be embedded in a variety of environments for fast semantic search.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Weld

    Weld

    High-performance runtime for data analytics applications

    Weld is a programming language and runtime designed to improve the performance of data-intensive applications by optimizing computations across multiple libraries. Instead of optimizing individual functions independently, Weld introduces an intermediate representation that allows different frameworks to share optimization opportunities. This approach reduces data movement between libraries and enables the system to generate highly optimized machine code for parallel execution. Weld is particularly useful for workloads involving large-scale data processing in frameworks such as NumPy, Spark, and TensorFlow. The language includes built-in constructs for expressing data-parallel operations, enabling efficient execution on modern hardware architectures. By combining operations from multiple libraries into a single optimized execution plan, Weld can significantly improve performance in analytics and machine learning pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Whatlang-RS

    Whatlang-RS

    Natural language detection library for Rust

    Whatlang-RS is a Rust-based language detection library optimized for speed and accuracy, supporting a wide range of languages with probabilistic models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    agentgateway

    agentgateway

    Next Generation Agentic Proxy for AI Agents and MCP servers

    Agentgateway is an open-source “data plane” built specifically for agentic AI connectivity, focusing on how agents talk to other agents and to tools across different frameworks and environments. It presents itself as a complete connectivity solution that adds drop-in security, observability, and governance to agent-to-agent and agent-to-tool communication without requiring you to rebuild your agent stack. The project supports interoperable protocols designed for this ecosystem, including Agent2Agent (A2A) and Model Context Protocol (MCP), which helps standardize how tools and agents interoperate. It is designed for performance and scale, implemented in Rust and engineered to handle large throughput and multi-tenant deployments. Operationally, it emphasizes safety and control with an RBAC system tuned for MCP/A2A use cases, plus the ability to update configuration dynamically via xDS without downtime.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB