• MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    PicoLM

    PicoLM

    Run a 1-billion parameter LLM on a $10 board with 256MB RAM

    ...The project focuses on enabling efficient local inference by optimizing memory usage, computation, and system dependencies so that relatively large models can operate on devices with minimal RAM. It is written primarily in C and designed with a minimalist architecture that removes unnecessary dependencies and external libraries. The runtime is capable of running language models with billions of parameters on devices with only a few hundred megabytes of memory, which is significantly lower than typical LLM infrastructure requirements. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    NullClaw

    NullClaw

    Fastest, smallest, and fully autonomous AI assistant infrastructure

    NullClaw is the smallest fully autonomous AI assistant infrastructure, built entirely in Zig as a single static binary with zero runtime dependencies. At just 678 KB with ~1 MB peak RAM usage, it boots in under 2 milliseconds and runs on virtually any hardware, including low-cost ARM boards. Despite its size, it delivers a complete AI stack with 22+ model providers, 18+ communication channels, integrated tools, hybrid memory, and sandboxed runtime support. Its architecture is fully modular, using vtable interfaces that allow providers, channels, tools, memory backends, and runtimes to be swapped without code changes. ...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 3
    MESHROOM

    MESHROOM

    3D reconstruction software

    ...Automatically estimate fisheye circle or manually edit it. Take advantage of motorized-head file. Easy to integrate in your Renderfarm System. Add specific rules to select the most suitable machines regarding CPU, RAM, GPU requirements of each Node.
    Downloads: 106 This Week
    Last Update:
    See Project
  • 4
    llmfit

    llmfit

    157 models, 30 providers, one command to find what runs on hardware

    llmfit is a terminal-based utility that helps developers determine which large language models can realistically run on their local hardware by analyzing system resources and model requirements. The tool automatically detects CPU, RAM, GPU, and VRAM specifications, then ranks available models based on performance factors such as speed, quality, and memory fit. It provides both an interactive terminal user interface and a traditional CLI mode, enabling flexible workflows for different user preferences. llmfit also supports advanced configurations including multi-GPU setups, mixture-of-experts architectures, and dynamic quantization recommendations. ...
    Downloads: 43 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    gensim

    gensim

    Topic Modelling for Humans

    Gensim is a Python library for topic modeling, document indexing, and similarity retrieval with large corpora. The target audience is the natural language processing (NLP) and information retrieval (IR) community.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    InvokeAI

    InvokeAI

    InvokeAI is a leading creative engine for Stable Diffusion models

    ...It provides a streamlined process with various new features and options to aid the image generation process. It runs on Windows, Mac and Linux machines, and runs on GPU cards with as little as 4 GB or RAM. InvokeAI is a leading creative engine built to empower professionals and enthusiasts alike. Generate and create stunning visual media using the latest AI-driven technologies. InvokeAI offers an industry leading Web Interface, interactive Command Line Interface, and also serves as the foundation for multiple commercial products. This fork is supported across Linux, Windows and Macintosh. ...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 7
    Datasets

    Datasets

    Hub of ready-to-use datasets for ML models

    ...We also feature a deep integration with the Hugging Face Hub, allowing you to easily load and share a dataset with the wider NLP community. There are currently over 2658 datasets, and more than 34 metrics available. Datasets naturally frees the user from RAM memory limitation, all datasets are memory-mapped using an efficient zero-serialization cost backend (Apache Arrow). Smart caching: never wait for your data to process several times.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    GPU Hot

    GPU Hot

    Real-time NVIDIA GPU dashboard

    GPU Hot is an open-source, lightweight monitoring dashboard designed to provide real-time visibility into NVIDIA GPU performance across single machines or entire clusters. The project offers a self-hosted web interface that streams hardware metrics directly from GPU servers, enabling developers, ML engineers, and system administrators to observe GPU utilization and system behavior in real time through a browser. The dashboard collects and displays a wide range of performance metrics...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    OnnxStream

    OnnxStream

    Lightweight inference library for ONNX files, written in C++

    The challenge is to run Stable Diffusion 1.5, which includes a large transformer model with almost 1 billion parameters, on a Raspberry Pi Zero 2, which is a microcomputer with 512MB of RAM, without adding more swap space and without offloading intermediate results on disk. The recommended minimum RAM/VRAM for Stable Diffusion 1.5 is typically 8GB. Generally, major machine learning frameworks and libraries are focused on minimizing inference latency and/or maximizing throughput, all of which at the cost of RAM usage. ...
    Downloads: 14 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    StudioOllamaUI

    StudioOllamaUI

    StudioOllamaUI is a local, portable interface for Ollama

    ...Unzip, run, and that's it. It doesn't clutter your registry or leave traces on your disk. AI for Everyone: No expensive GPU? No problem. Optimized to run smoothly on your CPU and RAM. Total Privacy: Everything stays on your machine. No data leaves for the cloud, and no hidden files are left on your system.
    Leader badge
    Downloads: 22 This Week
    Last Update:
    See Project
  • 11
    Glint Translator
    ...Features • 3 Translation Modes: Fluent (parallel), Area (overlay), Full Screen (smart detection) • Speaker detection with color-coding • Glint AI custom terminology control • Game-based profile system • Advanced settings with 50+ parameters for fine-tuned control • Share and import custom profiles (.glint) between users • Low CPU/RAM usage, optimized for Windows 10/11 Live Subtitle (Real-Time Voice Translation) Real-time speech-to-text translation for games, movies, and voice chats. Automatically detects audio, converts speech to text, and translates it instantly. Example: They speak German → you see Turkish AI Model Support • Google Gemini: 2.5 Flash, 2.5 Pro • OpenAI: GPT-4o, GPT-4 Turbo
    Downloads: 13 This Week
    Last Update:
    See Project
  • 12
    SPM - Monitoring  system

    SPM - Monitoring system

    Monitoring Tool for your IT Environment

    ...With SPM Monitoring Systems, you can monitor your network, servers, applications, and services with ease, and receive timely alerts when issues arise. Host Availability Monitoring. Agent-Based Monitoring: CPU,RAM,Storage,Network,EventLogs etc. SNMP Monitoring. Web Site Monitoring. Notification System: E-Mail, SMS, Push, Telegram Bot (also control via telegram bot) and Sound Notifications. Logging. Web Console. Control by Telegram Bot. Agents for Windows and Linux. SPM Monitoring Systems is a reliable and efficient solution for monitoring your IT environment. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    Warlock-Studio

    Warlock-Studio

    AI Suite for upscaling, interpolating & restoring images/videos

    v6.0. Warlock-Studio is a Windows application that uses Real-ESRGAN, BSRGAN, IRCNN, GFPGAN, RealESRNet, RealESRAnime and RIFE Artificial Intelligence models to upscale, restore faces, interpolate frames and reduce noise in images and videos. the application supports GPU acceleration (including multi-GPU setups) and offers batch processing for large workloads. It includes drag-and-drop handling for single or multiple files, optional pre-resize functions, and an automatic tiling system...
    Downloads: 33 This Week
    Last Update:
    See Project
  • 14
    TurboPilot

    TurboPilot

    Open source large-language-model based code completion engine

    TurboPilot is a self-hosted copilot clone that uses the library behind llama.cpp to run the 6 Billion Parameter Salesforce Codegen model in 4GiB of RAM. It is heavily based and inspired by on the fauxpilot project. This is a proof of concept right now rather than a stable tool. Autocompletion is quite slow in this version of the project. Feel free to play with it, but your mileage may vary.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Language Models

    Language Models

    Explore large language models in 512MB of RAM

    languagemodels is a lightweight Python library designed to simplify experimentation with large language models while maintaining extremely low hardware requirements. The project focuses on enabling developers and students to explore language model capabilities without needing expensive GPUs or large cloud infrastructures. By using small and optimized models, the library allows LLM inference to run in environments with limited resources, sometimes requiring only a few hundred megabytes of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    VAMS

    VAMS

    Virtual Assistant Maintenance System

    Virtual Assistant Maintenance System also knowns as VAMS is an AI software application, that helps users with some computer maintenance issues. Application Requirements: Operating System: Windows 8.1/10 /11 Processor: Intel Core i5 or equivalent RAM: 4GB or higher Free Disk Space: 500MB
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    LLaMA.go

    LLaMA.go

    llama.go is like llama.cpp in pure Golang

    ...The code of the project is based on the legendary ggml.cpp framework of Georgi Gerganov written in C++ with the same attitude to performance and elegance. Both models store FP32 weights, so you'll needs at least 32Gb of RAM (not VRAM or GPU RAM) for LLaMA-7B. Double to 64Gb for LLaMA-13B.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Alpaca.cpp

    Alpaca.cpp

    Locally run an Instruction-Tuned Chat-Style LLM

    Run a fast ChatGPT-like model locally on your device. This combines the LLaMA foundation model with an open reproduction of Stanford Alpaca a fine-tuning of the base model to obey instructions (akin to the RLHF used to train ChatGPT) and a set of modifications to llama.cpp to add a chat interface. Download the zip file corresponding to your operating system from the latest release. The weights are based on the published fine-tunes from alpaca-lora, converted back into a PyTorch checkpoint...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 19
    Darknet YOLO

    Darknet YOLO

    Real-Time Object Detection for Windows and Linux

    This is YOLO-v3 and v2 for Windows and Linux. YOLO (You only look once) is a state-of-the-art, real-time object detection system of Darknet, an open source neural network framework in C. YOLO is extremely fast and accurate. It uses a single neural network to divide a full image into regions, and then predicts bounding boxes and probabilities for each region. This project is a fork of the original Darknet project.
    Downloads: 42 This Week
    Last Update:
    See Project
  • 20
    Turi Create

    Turi Create

    Simplifies the development of custom machine learning models

    ...Turi Create supports macOS 10.12+, Linux (with glibc 2.10+), Windows 10 (via WSL). Turi Create requires Python 2.7, 3.5, 3.6, 3.7, 3.8. Also, x86_64 architecture, and at least 4 GB of RAM. We recommend using virtualenv to use, install, or build Turi Create. The package User Guide and API Docs contain more details on how to use Turi Create. If you want to build Turi Create from source, see BUILD.md. Turi Create does not require a GPU, but certain models can be accelerated 9-13x by utilizing a GPU.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Snips NLU

    Snips NLU

    Snips Python library to extract meaning from text

    Snips NLU is a Natural Language Understanding python library that allows to parse sentences written in natural language, and extract structured information. It’s the library that powers the NLU engine used in the Snips Console that you can use to create awesome and private-by-design voice assistants. The exact output is a bit richer, the point here is to give a glimpse on what kind of information can be extracted. Behind every chatbot and voice assistant lies a common piece of technology:...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    PyTorch-BigGraph

    PyTorch-BigGraph

    Generate embeddings from large-scale graph-structured data

    PyTorch-BigGraph (PBG) is a system for learning embeddings on massive graphs—think billions of nodes and edges—using partitioning and distributed training to keep memory and compute tractable. It shards entities into partitions and buckets edges so that each training pass only touches a small slice of parameters, which drastically reduces peak RAM and enables horizontal scaling across machines. PBG supports multi-relation graphs (knowledge graphs) with relation-specific scoring functions, negative sampling strategies, and typed entities, making it suitable for link prediction and retrieval. Its training loop is built for throughput: asynchronous I/O, memory-mapped tensors, and lock-free updates keep GPUs and CPUs fed even at extreme scale. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    The Edge Machine Learning library

    The Edge Machine Learning library

    Machine learning algorithms for edge devices

    Machine learning models for edge devices need to have a small footprint in terms of storage, prediction latency, and energy. One instance of where such models are desirable is resource-scarce devices and sensors in the Internet of Things (IoT) setting. Making real-time predictions locally on IoT devices without connecting to the cloud requires models that fit in a few kilobytes.These algorithms can train models for classical supervised learning problems with memory requirements that are...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Navigation Shortest Map Dhaval Kadia

    Navigation Shortest Map Dhaval Kadia

    Shortest Path Navigator with Directions and Distance

    Programmed on Language C. A nano-Infant of GOOGLE-MAP Optimized to Consume some MB instead of 50 GB of RAM This was Programmed Quarter Year ago in 2014
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25

    Darkbot

    The IRC's Talking Robot

    [ Please read https://sourceforge.net/p/darkbot/news/2014/01/darkbots-revitalization/ ] Darkbot is a portable IRC chat robot written in the C language that can be taught responses to user inquiries, and even have conversations with them. Darkbot was originally created by Jason Hamilton as an aid for help channels on Intenet Relay Chat.
    Leader badge
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB