Showing 404 open source projects for "cloud"

View related business solutions
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    MiniMax-MCP

    MiniMax-MCP

    Official MiniMax Model Context Protocol (MCP) server

    MiniMax-MCP is the official Model Context Protocol (MCP) server for accessing MiniMax’s multimodal generative APIs from MCP-compatible clients. It acts as a bridge between tools like Claude Desktop, Cursor, Windsurf, OpenAI Agents, and the MiniMax platform, exposing capabilities such as text-to-speech, voice cloning, image generation, text-to-image, video generation, image-to-video, text-to-video, and music generation. The server is written in Python and distributed under the MIT license,...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Qwen-Audio

    Qwen-Audio

    Chat & pretrained large audio language model proposed by Alibaba Cloud

    Qwen-Audio is a large audio-language model developed by Alibaba Cloud, built to accept various types of audio input (speech, natural sounds, music, singing) along with text input, and output text. There is also an instruction-tuned version called Qwen-Audio-Chat which supports conversational interaction (multi-round), audio + text input, creative tasks and reasoning over audio. It uses multi-task training over many different audio tasks (30+), and achieves strong multi-benchmarks performance without task-specific fine‐tuning. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    GPU Puzzles

    GPU Puzzles

    Solve puzzles. Learn CUDA

    ...The project emphasizes experimentation and problem solving, encouraging learners to discover GPU programming techniques through trial and exploration. It can be run in cloud environments such as Google Colab, making it easy for beginners to start experimenting without configuring local GPU hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    OpenAI Forward

    OpenAI Forward

    An efficient forwarding service designed for LLMs

    ...Its main purpose is to make model access more manageable and efficient by adding operational controls such as request rate limiting, token rate limiting, caching, logging, routing, and key management around existing LLM endpoints. The project can proxy both local and cloud-hosted language model services, which makes it useful for teams that want a single control layer regardless of whether they are using something like LocalAI or a hosted provider compatible with OpenAI-style APIs. A major emphasis of the repository is asynchronous performance, using tools such as uvicorn, aiohttp, and asyncio to support high-throughput forwarding workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    LLM-Finetuning

    LLM-Finetuning

    LLM Finetuning with peft

    ...The project focuses on parameter-efficient fine-tuning methods such as LoRA and QLoRA, which allow large models to be adapted to new tasks without requiring full retraining. Instead of requiring specialized hardware or complex training pipelines, many examples are designed to run in cloud notebook environments such as Google Colab. The repository includes step-by-step notebooks demonstrating how to fine-tune models such as LLaMA, Falcon, OPT, Vicuna, and GPT-NeoX. These tutorials show how developers can adapt pretrained models for tasks such as chatbots, classification, and instruction following. The project also illustrates how low-precision training techniques and adapter-based methods reduce memory requirements while maintaining strong model performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Machine Learning Systems

    Machine Learning Systems

    Introduction to Machine Learning Systems

    Machine Learning Systems is an open educational repository that serves as the source and learning stack for the Machine Learning Systems textbook, a project focused on teaching how to engineer AI systems that work reliably in real-world environments. Rather than concentrating only on model training, the material emphasizes the broader discipline of AI engineering, covering efficiency, reliability, deployment, and evaluation across the full lifecycle of intelligent systems. The repository...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    SERA CLI

    SERA CLI

    A tool to use the Ai2 Open Coding Agents Soft-Verified Agents

    ...It provides a convenient interface for deploying, testing, and using SERA models without needing to write scaffold code from scratch, acting as both a proxy and utility wrapper to simplify workflows that involve large agent models. Through sera-cli, users can connect to local or cloud-hosted SERA deployments, including via Modal for quick GPU provisioning and model caching, which helps accelerate experiments. The project is targeted at practitioners and researchers in the AI space who need a flexible but powerful CLI interface for model invocation, endpoint configuration, and integration with development pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Habit Tracker

    Habit Tracker

    Habit Tracker for the AI Coding Workshop

    Habit Tracker is a personal habit-tracking web application designed to help users build and maintain daily habits through intuitive UI and analytics that visualize progress over time. It runs locally with a FastAPI backend (Python) and a React frontend, storing all data in a lightweight SQLite database so there’s no need for user accounts or cloud storage, which keeps habit data fully private and self-contained. The app provides streak tracking and completion rates for each habit, giving users feedback on consistency and motivation by showing how often habits are completed and where they may be lagging. A calendar view lets users see a monthly grid of their habit history with color-coded days to highlight patterns and encourage daily engagement. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    PasteGuard

    PasteGuard

    Masks sensitive data and secrets before they reach AI

    ...PasteGuard supports two primary modes: mask mode, which anonymizes data and still uses external APIs; and route mode, which forwards sensitive requests to a local LLM inference engine while sending the rest to the cloud. It can be self-hosted via Docker, works with a wide range of SDKs and tools, and includes a browser extension for automatic protection in everyday AI chats.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    Mosec

    Mosec

    A high-performance ML model serving framework, offers dynamic batching

    Mosec is a high-performance and flexible model-serving framework for building ML model-enabled backend and microservices. It bridges the gap between any machine learning models you just trained and the efficient online service API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Determined

    Determined

    Determined, deep learning training platform

    ...Interpret your experiment results using the Determined UI and TensorBoard, and reproduce experiments with artifact tracking. Deploy your model using Determined's built-in model registry. Easily share on-premise or cloud GPUs with your team. Determined’s cluster scheduling offers first-class support for deep learning and seamless spot instance support. Check out examples of how you can use Determined to train popular deep learning models at scale.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    BentoML

    BentoML

    Unified Model Serving Framework

    BentoML simplifies ML model deployment and serves your models at a production scale. Support multiple ML frameworks natively: Tensorflow, PyTorch, XGBoost, Scikit-Learn and many more! Define custom serving pipeline with pre-processing, post-processing and ensemble models. Standard .bento format for packaging code, models and dependencies for easy versioning and deployment. Integrate with any training pipeline or ML experimentation platform. Parallelize compute-intense model inference...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ProxyAI

    ProxyAI

    The leading open-source AI copilot for JetBrains

    ProxyAI is an open-source AI-powered coding assistant designed primarily for JetBrains IDEs, offering a highly customizable alternative to tools like GitHub Copilot while maintaining flexibility across multiple AI providers and deployment environments. It allows developers to connect to a wide range of language models, including cloud-based services and locally hosted models, enabling both online and fully offline workflows depending on user preferences. The platform emphasizes deep integration with the developer’s environment, providing context-aware assistance by referencing files, folders, Git history, and even external documentation during interactions. ProxyAI enhances productivity by enabling natural language-driven code editing, intelligent autocompletion, and automated generation of commit messages, all within the IDE interface. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Microsandbox

    Microsandbox

    Secure local-first microVM sandbox for running untrusted code fast

    ...It aims to solve the common tradeoffs between speed, isolation, and control that developers encounter when running untrusted workloads. It provides a local-first and self-hosted approach, allowing users to maintain full ownership of their execution environment without depending on external cloud services. Microsandbox is particularly geared toward AI agent workflows, offering integrations that enable automated systems to safely run generated code and commands. It also supports standard container images, making it compatible with existing development ecosystems and tooling.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Ploomber

    Ploomber

    The fastest way to build data pipelines

    Ploomber is an open-source framework designed to simplify the development and deployment of data science and machine learning pipelines. It allows developers to transform exploratory data analysis workflows into production-ready pipelines without rewriting large portions of code. The system integrates with common development environments such as Jupyter Notebook, VS Code, and PyCharm, enabling data scientists to continue working with familiar tools while building scalable workflows. Ploomber...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    ...The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code completion, conversational chat assistance, and AI-assisted code editing directly within the IDE. Developers can select and manage models through a configuration interface that automatically downloads and runs the required models locally. The extension also supports agent-style coding workflows, where AI tools can perform more complex tasks such as analyzing project context or editing multiple files.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    RAG from Scratch

    RAG from Scratch

    Demystify RAG by building it from scratch

    RAG From Scratch is an educational open-source project designed to teach developers how retrieval-augmented generation systems work by building them step by step. Instead of relying on complex frameworks or cloud services, the repository demonstrates the entire RAG pipeline using transparent and minimal implementations. The project walks through key concepts such as generating embeddings, building vector databases, retrieving relevant documents, and integrating the retrieved context into language model prompts. Each example is written with detailed explanations so that developers can understand the internal mechanics of semantic search and context-aware language generation. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    ...The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows developers to use powerful open-source models such as Llama, Phi, DeepSeek, and Qwen while maintaining privacy and avoiding external service dependencies. The system supports contextual queries where the AI analyzes files within a directory and generates answers based on project documentation or source code. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    dataline

    dataline

    AI data analysis and visualization on CSV, Postgres, MySQL, Snowflake

    ...Once connected, users can generate tables, charts, and reports automatically based on queries produced by the AI engine. The platform is designed with a privacy-first architecture that stores data locally on the user’s device rather than sending it to external cloud services by default. It can also hide sensitive data from language models during processing, ensuring that only necessary metadata is used for query generation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    OmAgent

    OmAgent

    Build multimodal language agents for fast prototype and production

    OmAgent is an open-source Python framework designed to simplify the development of multimodal language agents that can reason, plan, and interact with different types of data sources. The framework provides abstractions and infrastructure for building AI agents that operate on text, images, video, and audio while maintaining a relatively simple interface for developers. Instead of forcing developers to implement complex orchestration logic manually, the system manages task scheduling, worker...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    OpenVINO Model Server

    OpenVINO Model Server

    A scalable inference server for models optimized with OpenVINO

    OpenVINO™ Model Server is a high-performance inference serving system designed to host and serve machine learning models that have been optimized with the OpenVINO toolkit. It’s implemented in C++ for scalability and efficiency, making it suitable for both edge and cloud deployments where inference workloads must be reliable and high throughput. The server exposes model inference via standard network protocols like REST and gRPC, allowing any client that speaks those protocols to request predictions remotely, abstracting away the complexity of where and how the model runs. It supports model deployment in diverse environments including Docker, bare-metal machines, and Kubernetes clusters, and is especially useful in microservices architectures where AI services need to scale independently. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    TraceRoot

    TraceRoot

    Find the Root Cause in Your Code's Trace

    ...It offers interactive trace exploration with zoomable log clusters, span and latency views, and code-linked insights. Lightweight SDKs for Python and TypeScript enable seamless instrumentation using OpenTelemetry, with support for both self-hosted and cloud deployment. Human-in-the-loop interaction is central: developers can guide reasoning by selecting relevant spans or logs, then verify agent reasoning through traceable context.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Phidata

    Phidata

    Build multi-modal Agents with memory, knowledge, tools and reasoning

    ...It includes features like real-time monitoring, agent evaluations, and performance optimization tools, ensuring the reliability and scalability of AI solutions. Phidata also allows developers to bring their own cloud infrastructure, offering flexibility for custom setups. The platform provides robust support for enterprises, including security features, agent guardrails, and automated DevOps for smoother deployment processes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    FATE

    FATE

    An industrial grade federated learning framework

    ...Supporting various federated learning scenarios, FATE now provides a host of federated learning algorithms, including logistic regression, tree-based algorithms, deep learning and transfer learning. FATE became open-source in February 2019. FATE TSC was established to lead FATE open-source community, with members from major domestic cloud computing and financial service enterprises. FedAI is a community that helps businesses and organizations build AI models effectively and collaboratively, by using data in accordance with user privacy protection, data security, data confidentiality and government regulations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ProjectLibre - Project Management

    ProjectLibre - Project Management

    #1 alternative to Microsoft Project : Project Management & Gantt Chart

    ProjectLibre project management software: #1 free alternative to Microsoft Project w/ 7.8M+ downloads in 193 countries. ProjectLibre is a replacement of MS Project & includes Gantt Chart, Network Diagram, WBS, Earned Value etc. This site downloads our FOSS desktop app. 🌐 Try the Cloud: http://www.projectlibre.com/register/trial We also offer ProjectLibre Cloud—a subscription, AI-powered SaaS for teams & enterprises. Cloud supports multi-project management w/ role-based access, central resource pool, Dashboard, Portfolio View 💡 The AI Cloud version can generate full project plans (tasks, durations, dependencies) from a natural language prompt — in any language...
    Leader badge
    Downloads: 16,480 This Week
    Last Update:
    See Project