• Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    ComfyUI-WanVideoWrapper

    ComfyUI-WanVideoWrapper

    ComfyUI wrapper nodes for WanVideo and related models

    ...The project supports complex node-based pipelines where users can control sampling, conditioning, and frame continuity across generated sequences. It also enables extended video generation by linking outputs between iterations, allowing for longer and more coherent animations. Additionally, the wrapper often includes optimizations for performance, such as low VRAM configurations and multi-stage sampling strategies.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    ComfyUI

    ComfyUI

    The most powerful and modular diffusion model GUI, api and backend

    ...This UI will let you design and execute advanced stable diffusion pipelines using a graph/nodes/flowchart-based interface. We are a team dedicated to iterating and improving ComfyUI, supporting the ComfyUI ecosystem with tools like node manager, node registry, cli, automated testing, and public documentation. Open source AI models will win in the long run against closed models and we are only at the beginning. Our core mission is to advance and democratize AI tooling. We believe that the future of AI tooling is open-source and community-driven.
    Downloads: 175 This Week
    Last Update:
    See Project
  • 3
    ComfyUI-Copilot

    ComfyUI-Copilot

    AI assistant for ComfyUI workflow generation, debugging, and tuning

    ComfyUI-Copilot is an AI-powered assistant designed to extend the capabilities of ComfyUI by simplifying and automating complex workflow development tasks. It functions as a custom node integrated directly into the ComfyUI environment, allowing users to interact with workflows through natural language and intelligent suggestions. ComfyUI-Copilot focuses on reducing the complexity of building node-based pipelines for generative AI tasks such as image generation, making it more accessible to both beginners and experienced users. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    ComfyUI-LTXVideo

    ComfyUI-LTXVideo

    LTX-Video Support for ComfyUI

    ComfyUI-LTXVideo is a bridge between ComfyUI’s node-based generative workflow environment and the LTX-Video multimedia processing framework, enabling creators to orchestrate complex video tasks within a visual graph paradigm. Instead of writing code to apply effects, transitions, edits, and data flows, users can assemble nodes that represent video inputs, transformations, and outputs, letting them prototype and automate video production pipelines visually.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    ComfyUI-3D-Pack

    ComfyUI-3D-Pack

    An extensive node suite that enables ComfyUI to process 3D inputs

    ComfyUI-3D-Pack is an extension package for the ComfyUI visual AI workflow environment that enables users to generate and manipulate 3D assets using advanced machine learning techniques. ComfyUI itself is a node-based interface for designing and executing generative AI pipelines, and this extension expands its capabilities by introducing nodes specifically designed for working with three-dimensional data. The package allows the platform to process inputs such as meshes and UV textures and integrate them into generative workflows similar to those used for image and video generation. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 6
    SwarmUI

    SwarmUI

    Modular AI image and video generation web UI with extensible tools

    ...SwarmUI is built to accommodate both beginners and advanced users by offering a simple “Generate” interface alongside more advanced workflow tools that expose deeper configuration options. It integrates with underlying systems like node-based workflows, enabling flexible and customizable pipelines for complex generation tasks. SwarmUI also emphasizes scalability, originally inspired by the idea of coordinating multiple GPUs to work together for large batch or grid-based image generation. SwarmUI includes a variety of built-in tools such as image editing, prompt handling, and automation features.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 7
    ComfyUI-HunyuanVideoWrapper

    ComfyUI-HunyuanVideoWrapper

    ComfyUI wrapper nodes for HunyuanVideo

    The ComfyUI-HunyuanVideoWrapper project is a ComfyUI extension that integrates Hunyuan-based multimodal video generation models into node-based workflows. It allows users to generate or manipulate video content by combining text prompts with one or more input images, enabling flexible conditioning of outputs. The system introduces specialized nodes such as text-image encoders that allow multiple image inputs to be referenced directly within prompts. This makes it possible to guide generation using both visual and textual context simultaneously. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    MESHROOM

    MESHROOM

    3D reconstruction software

    ...Take advantage of motorized-head file. Easy to integrate in your Renderfarm System. Add specific rules to select the most suitable machines regarding CPU, RAM, GPU requirements of each Node.
    Downloads: 116 This Week
    Last Update:
    See Project
  • 9
    Memobase

    Memobase

    Fast backend for long-term AI user memory via structured profiles

    ...The system focuses on three principal performance metrics: high search performance, reduced large language model (LLM) costs through batch processing techniques, and low latency with minimal SQL operations. Memobase supports integration with existing LLM workflows via APIs and SDKs (including Python, Node, and Go), making it easy to adopt within diverse application stacks.
    Downloads: 2 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    TensorRT LLM

    TensorRT LLM

    TensorRT LLM provides users with an easy-to-use Python API

    ...It provides a Python-based API built on top of PyTorch that allows developers to define, customize, and deploy LLMs efficiently across a variety of hardware configurations, from single GPUs to large multi-node clusters. The library focuses on maximizing throughput and minimizing latency through advanced techniques such as quantization, custom attention kernels, and optimized memory management strategies. It includes support for cutting-edge inference methods like speculative decoding and inflight batching, enabling real-time and large-scale AI applications. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    GPUStack

    GPUStack

    Performance-optimized AI inference on your GPUs

    GPUStack is an open-source GPU cluster management platform designed to simplify the deployment and operation of artificial intelligence models across heterogeneous hardware environments. The system aggregates GPU resources from multiple machines into a unified cluster so developers and administrators can run large language models and other AI workloads efficiently across distributed infrastructure. Instead of requiring complex orchestration systems such as Kubernetes, GPUStack provides a...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Hivemind

    Hivemind

    Decentralized deep learning in PyTorch. Built to train models

    ...Its intended usage is training one large model on hundreds of computers from different universities, companies, and volunteers. Distributed training without a master node: Distributed Hash Table allows connecting computers in a decentralized network. Fault-tolerant backpropagation: forward and backward passes succeed even if some nodes are unresponsive or take too long to respond. Decentralized parameter averaging: iteratively aggregate updates from multiple workers without the need to synchronize across the entire network. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    ...This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. Copyright (c) 2022, NVIDIA CORPORATION. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Claude Code SDK Python

    Claude Code SDK Python

    Python SDK for Claude Agent

    ...It is designed to integrate with local Python workflows and allow developers to embed Claude Code capabilities directly in their applications or scripts. The repo is MIT-licensed and includes documentation and installation instructions (requires Python 3.10+, Node installation of Claude Code). Example usage shows how to stream responses, parse structured message blocks, or create persistent client sessions.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 15
    TensorFlow

    TensorFlow

    TensorFlow is an open source library for machine learning

    ...The platform can be easily deployed on multiple CPUs, GPUs and Google's proprietary chip, the tensor processing unit (TPU). TensorFlow expresses its computations as dataflow graphs, with each node in the graph representing an operation. Nodes take tensors—multidimensional arrays—as input and produce tensors as output. The framework allows for these algorithms to be run in C++ for better performance, while the multiple levels of APIs let the user determine how high or low they wish the level of abstraction to be in the models produced. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 16
    Text Generation Inference

    Text Generation Inference

    Large Language Model Text Generation Inference

    Text Generation Inference is a high-performance inference server for text generation models, optimized for Hugging Face's Transformers. It is designed to serve large language models efficiently with optimizations for performance and scalability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    AReal

    AReal

    Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible

    AReaL is an open source, fully asynchronous reinforcement learning training system. AReal is designed for large reasoning and agentic models. It works with models that perform reasoning over multiple steps, agents interacting with environments. It is developed by the AReaL Team at Ant Group (inclusionAI) and builds upon the ReaLHF project. Release of training details, datasets, and models for reproducibility. It is intended to facilitate reproducible RL training on reasoning / agentic tasks,...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    MobileLLM

    MobileLLM

    MobileLLM Optimizing Sub-billion Parameter Language Models

    MobileLLM is a lightweight large language model (LLM) framework developed by Facebook Research, optimized for on-device deployment where computational and memory efficiency are critical. Introduced in the ICML 2024 paper “MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases”, it focuses on delivering strong reasoning and generalization capabilities in models under one billion parameters. The framework integrates several architectural innovations—SwiGLU...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    FullTClash

    FullTClash

    General proxy performance testing tool based on Clash using Telegram

    Back end part useClash project(It can also be called nowmihomo)The relevant code is used as the outing agent. The front end part uses Telegram API as the interactive interface, which needs to be used in conjunction with Telegram, that is, a Telegram robot (bot), FullTClash bot is a Telegram robot (hereinafter referred to as bot) carrying its test tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    TreeQuest

    TreeQuest

    A Tree Search Library with Flexible API for LLM Inference-Time Scaling

    TreeQuest, developed by SakanaAI, is a versatile Python library implementing adaptive tree search algorithms—such as AB‑MCTS—for enhancing inference-time performance of large language models (LLMs). It allows developers to define custom state-generation and scoring functions (e.g., via LLMs), and then efficiently explores possible answer trees during runtime. With support for multi-LLM collaboration, checkpointing, and mixed policies, TreeQuest enables smarter, trial‑and‑error question...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    LangGraph

    LangGraph

    Build resilient language agents as graphs

    LangGraph is a library for building stateful, multi-actor applications with LLMs, used to create agent and multi-agent workflows. Compared to other LLM frameworks, it offers these core benefits: cycles, controllability, and persistence. LangGraph allows you to define flows that involve cycles, essential for most agentic architectures, differentiating it from DAG-based solutions. As a very low-level framework, it provides fine-grained control over both the flow and state of your application,...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    Fast3R

    Fast3R

    Fast3R: Towards 3D Reconstruction of 1000+ Images in One Forward Pass

    Fast3R is Meta AI’s official CVPR 2025 release for “Towards 3D Reconstruction of 1000+ Images in One Forward Pass.” It represents a next-generation feedforward 3D reconstruction model capable of producing dense point clouds and camera poses for hundreds to thousands of images or video frames in a single inference pass—eliminating the need for slow, iterative structure-from-motion pipelines. Built on PyTorch Lightning and extending concepts from DUSt3R and Spann3r, Fast3R unifies multi-view...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    NVIDIA NeMo

    NVIDIA NeMo

    Toolkit for conversational AI

    ...Conversational AI architectures are typically large and require a lot of data and compute for training. NeMo uses PyTorch Lightning for easy and performant multi-GPU/multi-node mixed-precision training. Supported models: Jasper, QuartzNet, CitriNet, Conformer-CTC, Conformer-Transducer, Squeezeformer-CTC, Squeezeformer-Transducer, ContextNet, LSTM-Transducer (RNNT), LSTM-CTC. NGC collection of pre-trained speech processing models.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Apache Hamilton

    Apache Hamilton

    Helps data scientists define testable self-documenting dataflows

    Apache Hamilton is an open-source Python framework designed to simplify the creation and management of dataflows used in analytics, machine learning pipelines, and data engineering workflows. The framework enables developers to define data transformations as simple Python functions, where each function represents a node in a dataflow graph and its parameters define dependencies on other nodes. Hamilton automatically analyzes these functions and constructs a directed acyclic graph representing the pipeline, allowing the system to execute transformations in the correct order. This approach encourages modular, testable, and maintainable data pipelines because each transformation is isolated and easily unit tested. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Cube Studio

    Cube Studio

    Cube Studio open source cloud native one-stop machine learning

    Cube Studio is an open-source, cloud-native end-to-end machine learning and AI platform designed to support the full lifecycle of AI development — from data preparation and interactive notebook coding to distributed training, model tuning, and deployment in production-ready environments. It provides a unified interface where teams can manage data sources, track datasets, and build pipelines using drag-and-drop workflow orchestration, making it accessible for both engineers and data...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB