Showing 433 open source projects for "fast linux"

View related business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 1
    Fast MCP

    Fast MCP

    A Ruby Implementation of the Model Context Protocol

    Fast MCP is a lightweight framework designed to simplify the development and deployment of servers that implement the Model Context Protocol. The Model Context Protocol enables AI assistants and applications to connect with external tools, services, and data sources through a standardized interface. Fast-mcp provides developers with a streamlined toolkit for building MCP servers that expose application functionality to AI agents. The framework focuses on ease of use, allowing developers to...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 2
    fast-stable-diffusion

    fast-stable-diffusion

    Fast-stable-diffusion + DreamBooth

    fast-stable-diffusion is a community-curated GitHub repository that provides Colab notebooks and integration examples for running Stable Diffusion and associated UIs like AUTOMATIC1111, ComfyUI, and DreamBooth directly on Google Colab environments. Rather than being a standalone packaged application, this project offers ready-to-use interactive notebooks that install and launch full-feature Stable Diffusion web UIs inside Colab without requiring complex local setups or GPU installations....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    course.fast.ai

    course.fast.ai

    The fast.ai course notebooks

    course22 is the official repository containing the notebooks, slides, and supporting materials for the 2022 edition of the fast.ai course Practical Deep Learning for Coders. The repository serves as the core educational resource for the course, providing learners with hands-on exercises and coding tutorials that accompany each lecture. The project emphasizes learning deep learning through experimentation rather than purely theoretical study, encouraging students to build models and analyze...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Zvec

    Zvec

    A lightweight, lightning-fast, in-process vector database

    Zvec is an open-source, lightweight, in-process vector database designed to embed directly into applications and serve fast similarity search workloads without the overhead of a separate server process. Developed by Alibaba’s Tongyi Lab, it positions itself as the “SQLite of vector databases” by being easy to integrate, minimal in dependencies, and capable of handling high throughput with low latency on edge devices or small systems. Zvec excels at approximate nearest neighbor search and...
    Downloads: 14 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    ZeroClaw

    ZeroClaw

    Fast, small, and fully autonomous AI assistant infrastructure

    ZeroClaw is a Rust-native autonomous AI agent framework engineered for teams and developers who need highly efficient, secure, and modular AI automation infrastructure that can run reliably in both production and self-hosted environments. It is designed around a trait-based architecture so that model providers, communication channels, memory systems, and tooling integrations can be swapped or extended without rewriting core components, giving engineers flexibility and long-term...
    Downloads: 38 This Week
    Last Update:
    See Project
  • 6
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 7
    VectorChord

    VectorChord

    Scalable, fast, and disk-friendly vector search in Postgres

    VectorChord is an open-source vector database built for local and edge deployment. It supports efficient vector indexing and retrieval using ANN (approximate nearest neighbor) algorithms and is optimized for integration with LLM and AI applications. VectorChord is lightweight and can be embedded in a variety of environments for fast semantic search.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 8
    MCP Everything Search

    MCP Everything Search

    An MCP server that provides fast file searching capabilities

    Everything Search MCP Server is an MCP server that provides fast file searching capabilities across Windows, macOS, and Linux. On Windows, it utilizes the Everything SDK; on macOS, it leverages the built-in mdfind command; and on Linux, it uses the locate or plocate command. ​
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    PicoClaw

    PicoClaw

    Ultra-Efficient AI Assistant in Go

    ...The project’s goals include broad platform support (including Linux, macOS, and multiple CPU architectures), rapid startup times that make the assistant feel responsive, and integration with popular messaging platforms via gateways or bots.
    Downloads: 113 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    CatBoost

    CatBoost

    High-performance library for gradient boosting on decision trees

    CatBoost is a fast, high-performance open source library for gradient boosting on decision trees. It is a machine learning method with plenty of applications, including ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. CatBoost offers superior performance over other GBDT libraries on many datasets, and has several superb features. It has best in class prediction speed, supports both numerical and categorical features, has a fast and scalable...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 11
    Ninjabot

    Ninjabot

    A fast cryptocurrency platform for trading bot in Go

    A fast cryptocurrency trading bot framework implemented in Go. Ninjabot permits users to create and test custom strategies for spot markets. Ninjabot is an open-source platform that provides tools to implement custom strategies and backtests for trading cryptocurrencies in Go. Ninjabot CLI provides utilities commands to support backtesting and bot development. Currently, we only support Binance exchange. If you want to include support for other exchanges, you need to implement a new struct...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 12
    spaCy

    spaCy

    Industrial-strength Natural Language Processing (NLP)

    spaCy is a library built on the very latest research for advanced Natural Language Processing (NLP) in Python and Cython. Since its inception it was designed to be used for real world applications-- for building real products and gathering real insights. It comes with pretrained statistical models and word vectors, convolutional neural network models, easy deep learning integration and so much more. spaCy is the fastest syntactic parser in the world according to independent benchmarks, with...
    Downloads: 74 This Week
    Last Update:
    See Project
  • 13
    EnvPool

    EnvPool

    C++-based high-performance parallel environment execution engine

    EnvPool is a fast, asynchronous, and parallel RL environment library designed for scaling reinforcement learning experiments. Developed by SAIL at Singapore, it leverages C++ backend and Python frontend for extremely high-speed environment interaction, supporting thousands of environments running in parallel on a single machine. It's compatible with Gymnasium API and RLlib, making it suitable for scalable training pipelines.
    Downloads: 20 This Week
    Last Update:
    See Project
  • 14
    FastSD CPU

    FastSD CPU

    Fast stable diffusion on CPU and AI PC

    FastSD CPU is an optimized fork of Stable Diffusion designed to run efficiently on CPUs and devices without dedicated GPUs by leveraging Latent Consistency Models and Adversarial Diffusion Distillation techniques that accelerate inference. It focuses on bringing fast text-to-image generation to mainstream hardware like desktop CPUs, lower-end laptops, or edge devices without requiring high-end graphics processors. The repository contains multiple interfaces including a desktop GUI for simple...
    Downloads: 41 This Week
    Last Update:
    See Project
  • 15
    Microsandbox

    Microsandbox

    Secure local-first microVM sandbox for running untrusted code fast

    Microsandbox is an open source platform designed to securely execute untrusted code in isolated environments using lightweight virtualization techniques. It focuses on combining strong security guarantees with fast startup times by leveraging hardware-level microVM isolation instead of relying solely on traditional containers or full virtual machines. It aims to solve the common tradeoffs between speed, isolation, and control that developers encounter when running untrusted workloads. It...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 16
    Spice.ai OSS

    Spice.ai OSS

    A self-hostable CDN for databases

    Spice is a portable runtime offering developers a unified SQL interface to materialize, accelerate, and query data from any database, data warehouse, or data lake. Spice connects, fuses, and delivers data to applications, machine-learning models, and AI backends, functioning as an application-specific, tier-optimized Database CDN. The Spice runtime, written in Rust, is built-with industry-leading technologies such as Apache DataFusion, Apache Arrow, Apache Arrow Flight, SQLite, and DuckDB....
    Downloads: 31 This Week
    Last Update:
    See Project
  • 17
    PyTorch

    PyTorch

    Open source machine learning framework

    PyTorch is a Python package that offers Tensor computation (like NumPy) with strong GPU acceleration and deep neural networks built on tape-based autograd system. This project allows for fast, flexible experimentation and efficient production. PyTorch consists of torch (Tensor library), torch.autograd (tape-based automatic differentiation library), torch.jit (a compilation stack [TorchScript]), torch.nn (neural networks library), torch.multiprocessing (Python multiprocessing), and...
    Downloads: 124 This Week
    Last Update:
    See Project
  • 18
    llama.cpp

    llama.cpp

    Port of Facebook's LLaMA model in C/C++

    The llama.cpp project enables the inference of Meta's LLaMA model (and other models) in pure C/C++ without requiring a Python runtime. It is designed for efficient and fast model execution, offering easy integration for applications needing LLM-based capabilities. The repository focuses on providing a highly optimized and portable implementation for running large language models directly within C/C++ environments.
    Downloads: 193 This Week
    Last Update:
    See Project
  • 19
    Chainlit

    Chainlit

    Build Python LLM apps in minutes

    Chainlit is an open-source Python package that makes it incredibly fast to build and share LLM apps. Integrate the Chainlit API in your existing code to spawn a ChatGPT-like interface in minutes! Integrate seamlessly with an existing code base or start from scratch in minutes. Understand the intermediary steps that produced an output at a glance. Deep dive into prompts in the Prompt Playground to understand where things went wrong and iterate. Invite your teammates, create annotated datasets...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 20
    Keras

    Keras

    Python-based neural networks API

    Python Deep Learning library
    Downloads: 16 This Week
    Last Update:
    See Project
  • 21
    Tokenizers

    Tokenizers

    Fast State-of-the-Art Tokenizers optimized for Research and Production

    Fast State-of-the-art tokenizers, optimized for both research and production. Tokenizers provides an implementation of today’s most used tokenizers, with a focus on performance and versatility. These tokenizers are also used in Transformers. Train new vocabularies and tokenize, using today’s most used tokenizers. Extremely fast (both training and tokenization), thanks to the Rust implementation. Takes less than 20 seconds to tokenize a GB of text on a server’s CPU. Easy to use, but also...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    bloop

    bloop

    bloop is a fast code search engine written in Rust

    Bloop is an AI-powered code search tool designed to help developers quickly find relevant code snippets, documentation, and usage examples within large repositories. It provides natural language search capabilities and AI-enhanced recommendations for improving code discovery.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 23
    FLAML

    FLAML

    A fast library for AutoML and tuning

    FLAML is a lightweight Python library that finds accurate machine learning models automatically, efficiently and economically. It frees users from selecting learners and hyperparameters for each learner. For common machine learning tasks like classification and regression, it quickly finds quality models for user-provided data with low computational resources. It supports both classical machine learning models and deep neural networks. It is easy to customize or extend. Users can find their...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    vLLM

    vLLM

    A high-throughput and memory-efficient inference and serving engine

    vLLM is a fast and easy-to-use library for LLM inference and serving. High-throughput serving with various decoding algorithms, including parallel sampling, beam search, and more.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 25
    Rill

    Rill

    Fast SQL-based BI tool for real-time dashboards and analytics

    Rill is an operational BI tool that turns raw datasets into fast, interactive dashboards using SQL and a code-first approach. It helps data teams move from data lake to insight quickly, without the complexity of traditional BI systems. With an embedded in-memory database powered by DuckDB or ClickHouse, queries run in milliseconds, enabling real-time exploration and analysis. Rill supports local and remote data sources such as CSV, Parquet, S3, and GCS, making it flexible across...
    Downloads: 5 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB