Open Source Rust Artificial Intelligence Software - Page 5

Rust Artificial Intelligence Software

View 13471 business solutions

Browse free open source Rust Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Rust Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    aigc

    aigc

    An e-book about the real-world application of LLM

    "Building Large Language Model Applications: Application Development and Architecture Design" is an open source e-book about the real-world application of LLM. It introduces the basics and applications of large language models, as well as how to build your own models. These include writing, developing, and managing prompts, exploring what the best large language models can bring, and pattern and architecture design for LLM application development.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    bloop

    bloop

    bloop is a fast code search engine written in Rust

    Bloop is an AI-powered code search tool designed to help developers quickly find relevant code snippets, documentation, and usage examples within large repositories. It provides natural language search capabilities and AI-enhanced recommendations for improving code discovery.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    dfdx

    dfdx

    Deep learning in Rust, with shape checked tensors and neural networks

    Deep learning in Rust, with shape-checked tensors and neural networks. Ergonomics & safety focused deep learning in Rust.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    gptcommit

    gptcommit

    A git prepare-commit-msg hook for authoring commit messages with GPT-3

    A git prepare-commit-msg hook for authoring commit messages with GPT-3. With this tool, you can easily generate clear, comprehensive and descriptive commit messages letting you focus on writing code. To use gptcommit, simply run git commit as you normally would. The hook will automatically generate a commit message for you using a large language model like GPT. If you're not satisfied with the generated message, you can always edit it before committing. By default, gptcommit uses the GPT-3 model. Please ensure you have sufficient credits in your OpenAI account to use it. Commit messages are a key channel for developers to communicate their work with others, especially in code reviews. When making complex code changes, it can be tedious to thoroughly document the contents of each change. I often felt the impulse to just title my commit “fix bug” and move on. Surfacing these changes with gptcommit helps the author and reviewer by bringing attention to these additional changes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    gptee

    gptee

    LLMs done the UNIX-y way

    Output from a language model using standard input as the prompt. Now supporting GPT3.5 chat completions! gptee was designed for use within shell scripts and other programs and also works in interactive shells. You can compose commands and execute them in a script. Proceed with caution before running arbitrary shell scripts. Using a chat completion model (like gpt-3.5-turbo), you can then inject a system message with -s or --system messages. For davinci and other non-chat models, the output is prefixed to the prompt. Compose shell commands like you would in a script. Try with a custom model. By default gptee uses gpt-3.5-turbo.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    hora

    hora

    Efficient approximate nearest neighbor search algorithm collections

    hora is an open-source high-performance vector similarity search library designed for large-scale machine learning and information retrieval systems. The project focuses on approximate nearest neighbor search, a fundamental technique used in modern AI applications such as recommendation systems, image search, and semantic search engines. Hora implements multiple efficient indexing algorithms that allow systems to rapidly search through high-dimensional vectors produced by machine learning models. These vectors are commonly generated by neural networks to represent images, text, audio, or other data types in a mathematical embedding space. The library is written in Rust and emphasizes performance, safety, and efficient memory management, making it suitable for production-grade applications requiring low latency and high throughput.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    llm

    llm

    An ecosystem of Rust libraries for working with large language models

    llm is an ecosystem of Rust libraries for working with large language models - it's built on top of the fast, efficient GGML library for machine learning. The primary entry point for developers is the llm crate, which wraps the llm-base and the supported model crates. Documentation for the released version is available on Docs.rs. For end-users, there is a CLI application, llm-cli, which provides a convenient interface for interacting with supported models. Text generation can be done as a one-off based on a prompt, or interactively, through REPL or chat modes. The CLI can also be used to serialize (print) decoded models, quantize GGML files, or compute the perplexity of a model. It can be downloaded from the latest GitHub release or by installing it from crates.io.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    llm-chain

    llm-chain

    Rust crate for building chains in large language models

    We offer a collection of Rust crates packed with features that make working with Large Language Models easy and seamless. With llm-chain, you can focus on building powerful AI applications. Create reusable and easily customizable prompt templates for consistent and structured interactions with LLMs. Build powerful chains of prompts that allow you to execute more complex tasks, step by step, leveraging the full potential of LLMs. Provides seamless integration with LLaMa models, enabling natural language understanding and generation tasks with Facebook's research models. Incorporates support for Stanford's Alpaca models, expanding the range of available language models for advanced AI applications. Enhance your AI agents' capabilities by giving them access to various tools, such as running Bash commands, executing Python scripts, or performing web searches.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    lumen

    lumen

    Beautiful git diff viewer, generate commits with AI

    Lumen is an open-source command-line developer tool that enhances Git workflows by combining advanced diff visualization with AI-powered code assistance. The tool provides an ergonomic interface for reviewing code changes directly in the terminal, offering syntax-highlighted diffs and structured output to make change analysis easier. In addition to displaying differences between commits, Lumen integrates AI services that can explain code changes, generate commit messages, and assist with Git operations. The platform supports multiple AI providers, allowing developers to connect to models from services such as OpenAI, Claude, Groq, or locally hosted inference engines. It also includes interactive exploration features that allow users to search through commits and understand the history of changes in a repository. Because it runs entirely from the command line, the tool integrates seamlessly into existing Git workflows without requiring graphical interfaces or additional IDE plugins.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and efficiency, leveraging hardware acceleration across CPUs, GPUs, and specialized accelerators to deliver low-latency inference both on-device and in server environments. One of its key strengths is its flexibility, as it supports multiple backends and allows developers to configure execution providers depending on available hardware. ort also includes advanced capabilities such as model compilation and optimization, reducing startup time and improving runtime performance in production systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    pgvecto.rs

    pgvecto.rs

    Vector database plugin for Postgres, written in Rust

    pgvecto.rs is a Postgres extension that provides vector similarity search functions. It is written in Rust and based on pgrx. It is currently under heavy development, please take care when using it in production. pgvecto.rs is a Postgres extension, which means that you can use it directly within your existing database. This makes it easy to integrate into your existing workflows and applications. pgvecto.rs supports filtering. You can set conditions when searching or retrieving points. This is the missing feature of other postgres extensions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    pipeless

    pipeless

    A computer vision framework to create and deploy apps in minutes

    Pipeless is an open-source computer vision framework to create and deploy applications without the complexity of building and maintaining multimedia pipelines. It ships everything you need to create and deploy efficient computer vision applications that work in real-time in just minutes. Pipeless is inspired by modern serverless technologies. It provides the development experience of serverless frameworks applied to computer vision. You provide some functions that are executed for new video frames and Pipeless takes care of everything else. You can easily use industry-standard models, such as YOLO, or load your custom model in one of the supported inference runtimes. Pipeless ships some of the most popular inference runtimes, such as the ONNX Runtime, allowing you to run inference with high performance on CPU or GPU out-of-the-box. You can deploy your Pipeless application with a single command to edge and IoT devices or the cloud.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    rssbot

    rssbot

    Lightweight Telegram RSS notification bot

    Lightweight Telegram RSS notification bot. Lightweight Telegram RSS bot for message notifications. You can download the precompiled program directly from Releases (with the Chinese version), the Linux version is musl static link, no other dependencies are required. It should be noted that the RSS records that have been pushed will not be retained. If the converted database is used directly, the old RSS will be pushed repeatedly.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    rust-bert

    rust-bert

    Rust native ready-to-use NLP pipelines and transformer-based models

    rust-bert is a Rust-based implementation of transformer-based natural language processing models that provides ready-to-use pipelines for tasks such as text classification, summarization, and question answering. The project ports many capabilities of the Hugging Face Transformers ecosystem into the Rust programming language. It allows developers to run state-of-the-art NLP models like BERT, GPT-2, and DistilBERT directly within Rust applications while maintaining high performance and memory efficiency. The library integrates with Rust machine learning infrastructure using crates such as tch-rs and ONNX Runtime for model execution. It also includes tokenization utilities, model architectures, and task-specific pipelines that simplify the development of NLP applications. Because Rust is known for its safety and performance, this project enables developers to deploy modern NLP models in production systems written in Rust.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. Shimmy focuses on performance and simplicity, using efficient runtime components to minimize memory usage and startup time compared to heavier inference frameworks. It supports modern model formats such as GGUF and SafeTensors and can automatically discover models stored locally or in common directories used by other AI tools. Advanced capabilities include CPU offloading for Mixture-of-Experts models and GPU acceleration, enabling large models to run on consumer hardware with limited VRAM.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    uzu

    uzu

    A high-performance inference engine for AI models

    uzu is a high-performance inference engine designed to run artificial intelligence models efficiently on Apple Silicon hardware. Written primarily in Rust and leveraging Apple’s Metal framework, the project focuses on maximizing performance when executing large language models and other AI workloads on devices such as Mac computers with M-series chips. The engine implements a hybrid architecture in which model layers can be executed either as custom GPU kernels or through Apple’s MPSGraph API, allowing it to balance performance and compatibility depending on the workload. By utilizing Apple’s unified memory architecture, uzu reduces memory copying overhead and improves inference throughput for local AI workloads. The system includes a simple high-level API that enables developers to run models, create inference sessions, and generate outputs with minimal configuration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    yek

    yek

    Serialize repositories into LLM-ready context w/ smart prioritization

    Yek is a Rust-based CLI tool designed to serialize text-based files from a repository or directory into a single structured output for large language model use. It scans projects using .gitignore rules to exclude irrelevant files and automatically filters out binary or oversized content. Yek prioritizes files based on Git history, placing more important content later in the output to align with how language models process context. Yek supports multiple directories, individual files, and glob patterns, making it flexible for different workflows. It can stream output when piped or save results to a temporary file, depending on usage. Configuration is handled through a yek.yaml file, allowing users to define ignore rules and priority settings. By consolidating code and documents into a single, ordered format, Yek simplifies preparing repositories for AI-driven analysis, debugging, or automation tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB