Showing 27 open source projects for "fast linux"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 1
    VectorChord

    VectorChord

    Scalable, fast, and disk-friendly vector search in Postgres

    VectorChord is an open-source vector database built for local and edge deployment. It supports efficient vector indexing and retrieval using ANN (approximate nearest neighbor) algorithms and is optimized for integration with LLM and AI applications. VectorChord is lightweight and can be embedded in a variety of environments for fast semantic search.
    Downloads: 23 This Week
    Last Update:
    See Project
  • 2
    ZeroClaw

    ZeroClaw

    Fast, small, and fully autonomous AI assistant infrastructure

    ZeroClaw is a Rust-native autonomous AI agent framework engineered for teams and developers who need highly efficient, secure, and modular AI automation infrastructure that can run reliably in both production and self-hosted environments. It is designed around a trait-based architecture so that model providers, communication channels, memory systems, and tooling integrations can be swapped or extended without rewriting core components, giving engineers flexibility and long-term...
    Downloads: 36 This Week
    Last Update:
    See Project
  • 3
    Microsandbox

    Microsandbox

    Secure local-first microVM sandbox for running untrusted code fast

    Microsandbox is an open source platform designed to securely execute untrusted code in isolated environments using lightweight virtualization techniques. It focuses on combining strong security guarantees with fast startup times by leveraging hardware-level microVM isolation instead of relying solely on traditional containers or full virtual machines. It aims to solve the common tradeoffs between speed, isolation, and control that developers encounter when running untrusted workloads. It...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 4
    Spice.ai OSS

    Spice.ai OSS

    A self-hostable CDN for databases

    Spice is a portable runtime offering developers a unified SQL interface to materialize, accelerate, and query data from any database, data warehouse, or data lake. Spice connects, fuses, and delivers data to applications, machine-learning models, and AI backends, functioning as an application-specific, tier-optimized Database CDN. The Spice runtime, written in Rust, is built-with industry-leading technologies such as Apache DataFusion, Apache Arrow, Apache Arrow Flight, SQLite, and DuckDB....
    Downloads: 32 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Tokenizers

    Tokenizers

    Fast State-of-the-Art Tokenizers optimized for Research and Production

    Fast State-of-the-art tokenizers, optimized for both research and production. Tokenizers provides an implementation of today’s most used tokenizers, with a focus on performance and versatility. These tokenizers are also used in Transformers. Train new vocabularies and tokenize, using today’s most used tokenizers. Extremely fast (both training and tokenization), thanks to the Rust implementation. Takes less than 20 seconds to tokenize a GB of text on a server’s CPU. Easy to use, but also...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    bloop

    bloop

    bloop is a fast code search engine written in Rust

    Bloop is an AI-powered code search tool designed to help developers quickly find relevant code snippets, documentation, and usage examples within large repositories. It provides natural language search capabilities and AI-enhanced recommendations for improving code discovery.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 7
    MagicAPI AI Gateway

    MagicAPI AI Gateway

    Built for demanding AI workflows

    The world's fastest AI Gateway proxy, written in Rust and optimized for maximum performance. This high-performance API gateway routes requests to various AI providers (OpenAI, GROQ) with streaming support, making it perfect for developers who need reliable and blazing-fast AI API access.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    Every Code (often referred to simply as Code) is a fast, local AI-powered coding agent designed to run directly in the terminal environment. It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with...
    Downloads: 13 This Week
    Last Update:
    See Project
  • 9
    Qdrant

    Qdrant

    Vector Database for the next generation of AI applications

    Qdrant is a vector similarity engine & vector database. It deploys as an API service providing search for the nearest high-dimensional vectors. With Qdrant, embeddings or neural network encoders can be turned into full-fledged applications for matching, searching, recommending, and much more! Provides the OpenAPI v3 specification to generate a client library in almost any programming language. Alternatively, utilize ready-made client for Python or other programming languages with additional...
    Downloads: 67 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 10
    Lightpanda Browser

    Lightpanda Browser

    Lightpanda: the headless browser designed for AI and automation

    Lightpanda is an open-source headless browser designed specifically for automation, artificial intelligence workflows, and large-scale web interaction tasks. Unlike traditional browsers that include full graphical rendering engines meant for human users, Lightpanda is built from scratch to operate entirely in headless mode, focusing only on the components required for programmatic web interaction. This design allows it to execute JavaScript and interact with web pages while avoiding the...
    Downloads: 56 This Week
    Last Update:
    See Project
  • 11
    OpenChat for Linux

    OpenChat for Linux

    OpenChat for Linux — a fast, lightweight desktop client for ChatGPT

    OpenChat for Linux is a desktop client for ChatGPT / OpenAI Chat designed specifically for Linux. It’s built with Tauri (Rust) for low resource usage and stability, and it uses a “message window” approach (keeps a small active slice of the conversation and loads more as you scroll) so long chats don’t bog down or crash the app. Downloads are available in common Linux formats (AppImage, Debian package, tarball), with additional packaging manifests for Flatpak, Snap, RPM, AUR, and Nix.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 12
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    Extractous

    Extractous

    Fast and efficient unstructured data extraction

    Extractous is a Rust-based unstructured data extraction library focused on fast local parsing of documents and other content-heavy files. Its purpose is to extract text and metadata efficiently from formats such as PDF, Word, HTML, email archives, images, and more, without depending on external APIs or separate parsing servers. The project emphasizes performance and low memory usage, and its maintainers describe it as a local-first alternative to heavier extraction stacks. For broader format...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    AIChat

    AIChat

    All-in-one LLM CLI tool featuring Shell Assistant

    AIChat is a lightweight terminal-based chatbot powered by GPT models, enabling AI-driven conversations directly from the command line.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 15
    Daft

    Daft

    Distributed DataFrame for Python designed for the cloud

    Daft is a framework for ETL, analytics and ML/AI at scale. Its familiar Python Dataframe API is built to outperform Spark in performance and ease of use. Daft plugs directly into your ML/AI stack through efficient zero-copy integrations with essential Python libraries such as Pytorch and Ray. It also allows requesting GPUs as a resource for running models. Daft runs locally with a lightweight multithreaded backend. When your local machine is no longer sufficient, it scales seamlessly to run...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 16
    Lance

    Lance

    Modern columnar data format for ML and LLMs implemented in Rust

    Lance is a columnar data format that is easy and fast to version, query and train on. It’s designed to be used with images, videos, 3D point clouds, audio and of course tabular data. It supports any POSIX file systems, and cloud storage like AWS S3 and Google Cloud Storage.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 17
    Zed

    Zed

    High-performance, multiplayer code editor from the creators of Atom

    Zed is a next-generation code editor designed for high-performance collaboration with humans and AI. Written from scratch in Rust to efficiently leverage multiple CPU cores and your GPU. Integrate upcoming LLMs into your workflow to generate, transform, and analyze code. Chat with teammates, write notes together, and share your screen and project. Multibuffers compose excerpts from across the codebase in one editable surface. Evaluate code inline via Jupyter runtimes and collaboratively edit...
    Downloads: 29 This Week
    Last Update:
    See Project
  • 18
    HelixDB

    HelixDB

    Graph-vector database for building unified AI backends fast

    HelixDB is an open source database designed to unify multiple data storage paradigms into a single platform tailored for AI-driven applications. It combines graph and vector data models, allowing developers to manage relationships and embeddings within the same system without relying on separate services. HelixDB is built from scratch in Rust and uses LMDB as its storage engine, enabling high performance and low-latency query execution. HelixDB also supports additional data formats such as...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 19
    CocoIndex

    CocoIndex

    ETL framework to index data for AI, such as RAG

    CocoIndex is an open-source framework designed for building powerful, local-first semantic search systems. It lets users index and retrieve content based on meaning rather than keywords, making it ideal for modern AI-based search applications. CocoIndex leverages vector embeddings and integrates with various models and frameworks, including OpenAI and Hugging Face, to provide high-quality semantic understanding. It’s built for transparency, ease of use, and local control over your search...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 20
    SemTools

    SemTools

    Semantic search and document parsing tools for the command line

    SemTools is an open-source command-line toolkit designed for document parsing, semantic indexing, and semantic search workflows. The project focuses on enabling developers and AI agents to process large document collections and extract meaningful semantic representations that can be searched efficiently. Built with Rust for performance and reliability, the toolchain provides fast processing of text and structured documents while maintaining low system overhead. SemTools can parse documents,...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 21
    Forge Code

    Forge Code

    AI enabled pair programmer for Claude, GPT, O Series, Grok, Deepseek

    Forge is a modern, open-source tool that brings AI-powered code assistance directly into your terminal workflow, effectively turning your shell into a “pair programmer”, without ever leaving your development environment. Written in Rust (with a command-line interface), Forge integrates with your existing shell (bash, zsh, fish, etc.) or IDE-agnostic workflows, allowing you to interact with your codebase, command-line tools, and version control as usual, but with the added support of large...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 22
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 24
    Code2Prompt

    Code2Prompt

    Convert codebases into structured prompts optimized for LLM analysis

    code2prompt is an open source command line tool designed to convert an entire codebase into a structured prompt that can be easily used with large language models. It analyzes a project directory, gathers relevant source files, and formats them into a single prompt that includes the source tree and code content. This approach helps developers quickly provide full project context to AI models without manually copying files or assembling prompts. code2prompt is built in Rust and focuses on...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    DeepProve

    DeepProve

    Framework to prove inference of ML models blazingly fast

    DeepProve is an advanced cryptographic framework designed to verify machine learning model inference using zero-knowledge proofs, enabling trustless validation of AI computations without exposing underlying data. The project focuses on zkML, a rapidly emerging field that combines machine learning with zero-knowledge cryptography to ensure both privacy and correctness. It supports neural network architectures such as multilayer perceptrons and convolutional neural networks, allowing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB