Showing 25 open source projects for "hardware configuration"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Project NOMAD

    Project NOMAD

    Self-contained, offline survival computer with tools, knowledge, & AI

    Project N.O.M.A.D. (Node for Offline Media, Archives, and Data) is a self-contained, offline-first knowledge and education server designed to run on local hardware. It combines AI, educational resources, and critical information tools into a single platform accessible through a web browser. Built as a containerized system managed via Docker, it orchestrates multiple applications through a centralized “Command Center” interface. NOMAD enables users to access offline Wikipedia, AI chat, maps,...
    Downloads: 107 This Week
    Last Update:
    See Project
  • 2
    lmstudio.js

    lmstudio.js

    LM Studio TypeScript SDK

    ...The library exposes the same capabilities used internally by the LM Studio desktop application, allowing external apps to load models, run inference, and build autonomous AI workflows. It is designed to simplify the creation of local AI tools by handling complex concerns such as dependency management, hardware compatibility, and model configuration. The SDK introduces an agent-style API that can execute multi-step tool-using workflows through a single call, enabling more advanced automation scenarios. Applications built with the SDK can run anywhere LM Studio is available, whether in foreground or headless mode. Overall, lmstudio-js acts as a developer bridge between local LLM infrastructure and modern JavaScript applications.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    mergekit

    mergekit

    Tools for merging pretrained large language models

    ...This approach allows researchers to combine specialized models into a more versatile system capable of performing multiple tasks. mergekit implements a variety of merging algorithms and strategies that control how model parameters are blended together during the merging process. The library is designed to operate efficiently even in environments with limited hardware resources by using memory-efficient processing methods that can run entirely on CPUs. It also provides configuration-driven workflows that allow users to experiment with different merging strategies without modifying source code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    kokoro-onnx

    kokoro-onnx

    TTS with kokoro and onnx runtime

    kokoro-onnx is a text-to-speech toolkit that wraps the Kokoro neural TTS model in an easy-to-use ONNX Runtime interface, so you can generate speech from Python with minimal setup. It focuses on running efficiently on commodity hardware, including macOS with Apple Silicon, while still delivering near real-time performance for many use cases. The project ships prebuilt model files and a simple example script, so you can go from installation to producing an audio.wav file in just a few steps. It supports multiple languages and voices, with a curated voice list and configuration via a VOICES file hosted alongside the models. ...
    Downloads: 91 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    AI Runner

    AI Runner

    Offline inference engine for art, real-time voice conversations

    AI Runner is an offline inference engine designed to run a collection of AI workloads on your own machine, including image generation for art, real-time voice conversations, LLM-powered chatbots and automated workflows. It is implemented as a desktop-oriented Python application and emphasizes privacy and self-hosting, allowing users to work with text-to-speech, speech-to-text, text-to-image and multimodal models without sending data to external services. At the core of its LLM stack is a...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 6
    Humanoid-Gym

    Humanoid-Gym

    Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real

    Humanoid-Gym is a reinforcement learning framework designed to train locomotion and control policies for humanoid robots using high-performance simulation environments. The system is built on top of NVIDIA Isaac Gym, which allows large-scale parallel simulation of robotic environments directly on GPU hardware. Its primary goal is to enable efficient training of humanoid robots in simulation while enabling policies to transfer effectively to real-world hardware without additional training....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    uzu

    uzu

    A high-performance inference engine for AI models

    ...By utilizing Apple’s unified memory architecture, uzu reduces memory copying overhead and improves inference throughput for local AI workloads. The system includes a simple high-level API that enables developers to run models, create inference sessions, and generate outputs with minimal configuration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Colossal-AI

    Colossal-AI

    Making large AI models cheaper, faster and more accessible

    The Transformer architecture has improved the performance of deep learning models in domains such as Computer Vision and Natural Language Processing. Together with better performance come larger model sizes. This imposes challenges to the memory wall of the current accelerator hardware such as GPU. It is never ideal to train large models such as Vision Transformer, BERT, and GPT on a single GPU or a single machine. There is an urgent demand to train models in a distributed environment....
    Downloads: 0 This Week
    Last Update:
    See Project
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 10
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models. Its README emphasizes real-world adoption across startups, research groups, and large companies, signaling a focus on...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 11
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 12
    OuteTTS

    OuteTTS

    Interface for OuteTTS models

    OuteTTS is an interface library for running OuteTTS text-to-speech models across a range of backends, making it easier to deploy the same model on different hardware and runtimes. It provides a high-level Interface API that wraps model configuration, speaker handling, and audio generation so you can focus on integrating speech into your application rather than wiring up low-level engines. The project supports multiple backends including llama.cpp (Python bindings and server), Hugging Face Transformers, ExLlamaV2, VLLM and a JavaScript interface via Transformers.js, allowing it to run on CPUs, NVIDIA CUDA GPUs, AMD ROCm, Vulkan-capable GPUs, and Apple Metal. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Moltworker

    Moltworker

    Run OpenClaw on Cloudflare Workers

    ...The project includes web UIs, Cloudflare configuration files, and scripts to authenticate devices, manage environment secrets, and enable optional services like persistent R2 storage for chat history.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    llama.vscode is a Visual Studio Code extension that provides AI-assisted coding features powered primarily by locally running language models. The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code completion, conversational chat assistance, and AI-assisted code editing directly within the IDE. Developers can select and manage models through a configuration interface that automatically downloads and runs the required models locally. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Deep Java Library (DJL)

    Deep Java Library (DJL)

    An engine-agnostic deep learning framework in Java

    ...Because DJL is deep learning engine agnostic, you don't have to make a choice between engines when creating your projects. You can switch engines at any point. To ensure the best performance, DJL also provides automatic CPU/GPU choice based on hardware configuration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    HPCC Systems

    HPCC Systems

    End-to-end big data in a massively scalable supercomputing platform.

    HPCC Systems® (www.hpccsystems.com) from LexisNexis® Risk Solutions is a proven, open source solution for Big Data insights that can be implemented by businesses of all sizes. With HPCC Systems, developers can design applications with Big Data at their core, enabling businesses to better analyze and understand data at scale, improving business time to results and decisions. HPCC Systems offers a consistent data-centric programming language, two processing platforms and a single, complete...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    gpu_poor

    gpu_poor

    Calculate token/s & GPU memory requirement for any LLM

    gpu_poor is an open-source tool designed to help developers determine whether their hardware is capable of running a specific large language model and to estimate the performance they can expect from it. The project focuses on calculating GPU memory requirements and predicted inference speed for different models, hardware configurations, and quantization strategies. By analyzing factors such as model size, context length, batch size, and GPU specifications, the system estimates how much VRAM...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    State of Open Source AI

    State of Open Source AI

    Clarity in the current fast-paced mess of Open Source innovation

    ...Because the AI domain moves quickly, part of the aim is to make the content maintainable and updateable by the community. The structure includes chapters or sections about model formats, evaluation benchmarks, hardware/backends, MLOps systems, alignment and safety issues, and open datasets. The repository contains the text (in Markdown or similar), configuration for build or publishing (static site or e-book), and contributor guidelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Lightning-Hydra-Template

    Lightning-Hydra-Template

    PyTorch Lightning + Hydra. A very user-friendly template

    Convenient all-in-one technology stack for deep learning prototyping - allows you to rapidly iterate over new models, datasets and tasks on different hardware accelerators like CPUs, multi-GPUs or TPUs. A collection of best practices for efficient workflow and reproducibility. Thoroughly commented - you can use this repo as a reference and educational resource. Not fitted for data engineering - the template configuration setup is not designed for building data processing pipelines that depend on each other. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    llm

    llm

    An ecosystem of Rust libraries for working with large language models

    llm is an ecosystem of Rust libraries for working with large language models - it's built on top of the fast, efficient GGML library for machine learning. The primary entry point for developers is the llm crate, which wraps the llm-base and the supported model crates. Documentation for the released version is available on Docs.rs. For end-users, there is a CLI application, llm-cli, which provides a convenient interface for interacting with supported models. Text generation can be done as a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    ...Unlike earlier object detection systems that only scale depth or width, this architecture scales multiple aspects of the neural network including structure, resolution, and channel configuration. This scaling strategy enables the model to adapt to different hardware environments while maintaining a strong balance between speed and detection accuracy. The repository includes multiple model variants such as YOLOv4-tiny, YOLOv4-CSP, and large-scale configurations designed for high-performance detection tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22

    manjaro talking

    enjoy the simplicity, with accessibility!

    This project is more intendet for advanced users, because it uses the architect installer, see https://wiki.manjaro.org/index.php?title=Installation_with_Manjaro_Architect If you have problems installing drivers, open up manjaro settings manager>hardware configuration and look for your drivers there. The above method requires ocr/sighted assistance. For easier navigation and for the method that works on both desktop and architect images, use mhwd-tui Hope you enjoy! Do not hesidate to send me your feetback at stasp@poczta.sytes.net or some where on SF. Thank you for checking out this project! ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    exchange-core

    exchange-core

    Ultra-fast matching engine written in Java based on LMAX Disruptor

    ...Goldman Sachs GS Collections), Real Logic Agrona, OpenHFT Chronicle-Wire, LZ4 Java, and Adaptive Radix Trees. Designed for high scalability and pauseless 24/7 operation under high-load conditions and providing low-latency responses. Single order book configuration is capable to process 5M operations per second on 10-years old hardware (Intel® Xeon® X5690) with moderate latency degradation. HFT optimized. Priority is a limit-order-move operation mean latency (currently ~0.5µs). Cancel operation takes ~0.7µs, placing new order ~1.0µs. Disk journaling and journal replay support, state snapshots (serialization) and restore operations, LZ4 compression. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Swift AI

    Swift AI

    The Swift machine learning library

    ...Swift AI includes a collection of common tools used for artificial intelligence and scientific applications. A flexible, fully-connected neural network with support for deep learning. Optimized specifically for Apple hardware, using advanced parallel processing techniques. We've created some example projects to demonstrate the usage of Swift AI. Each resides in their own repository and can be built with little or no configuration. Each module now contains its own documentation. We recommend that you read the docs carefully for detailed instructions on using the various components of Swift AI. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    oooark
    Project migrated: https://github.com/arktools This library aims at providing a uniform method for simulation, prototyping, and testing of autonomous robots. It also adheres to the object-oriented paradigm to ease system configuration and development. http://www.youtube.com/watch?v=-wQVrM5SL2o&fe
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB