Showing 150 open source projects for "run code"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 1
    CodinIT.dev

    CodinIT.dev

    Free, local, open-source AI app builder

    CodinIT.dev is a free, local, open source AI app builder that lets you go from idea to full-stack application entirely on your machine, no coding required, just chat with AI. You can build unlimited apps with real-time previews, instant undo, and responsive, frictionless workflows. Deep Supabase integration means you can create UI and backend logic in one cohesive environment, while the model-agnostic architecture lets you connect to any AI, whether cloud-based (Gemini 3 Pro, GPT-5,...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 2
    Shinkai: Local AI Agents

    Shinkai: Local AI Agents

    Shinkai allows you to create advanced AI (local) agents effortlessly

    Shinkai is a free, open-source AI platform that lets anyone create powerful AI agents without coding. These agents can collaborate with each other, handle complex tasks, and operate in decentralized crypto environments. Key Features: - No-Code Agent Creation - Build specialized agents (trading bots, sentiment trackers, etc.) with simple descriptions - Multi-Agent Collaboration - Agents work together to solve complex problems - Crypto Integration - Built-in support for decentralized payments and transactions - Flexible AI Models - Choose from cloud models (GPT-4, Claude) or run locally - Universal Compatibility - Works with Model Context Protocol (MCP) for cross-platform integration - Local Security - Crypto keys and computations stay on your device Shinkai transforms AI from single-task tools into collaborative, autonomous systems that can operate in decentralized networks while maintaining privacy and security.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    Mixtral offloading

    Mixtral offloading

    Run Mixtral-8x7B models in Colab or consumer desktops

    ...The project implements techniques that allow model components to be dynamically moved between CPU memory and GPU memory during inference, significantly reducing the amount of GPU VRAM required to run the model. This approach takes advantage of the sparse activation properties of mixture-of-experts architectures, where only a subset of expert networks are used for each token during generation. By selectively loading and caching the required experts, the system avoids keeping the entire model in GPU memory at once. The repository includes notebooks and code examples that demonstrate how to run large language models on consumer hardware such as personal GPUs or cloud notebook environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    OpenAssistant

    OpenAssistant

    Chat-based assistant that understands tasks

    ...We are in the early stages of development, working from established research in applying RLHF to large language models. Open Assistant is a project organized by LAION and individuals around the world interested in bringing this technology to everyone. The code and models are licensed under the Apache 2.0 license. Open Assistant will be free to use and modify. There will be versions which will be runnable on consumer hardware. You do not need to run the project locally unless you are contributing to the development process. The website link above will take you to the public website where you can use the data collection app.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    AutoPR

    AutoPR

    Run AI-powered workflows over your codebase

    AutoPR is an AI-driven tool for automating pull request (PR) generation and review processes. It streamlines code contributions by suggesting fixes, generating pull requests, and reviewing code using AI models, reducing manual overhead for developers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    ReplitLM

    ReplitLM

    Inference code and configs for the ReplitLM model family

    ReplitLM is a family of open-source language models developed by Replit for assisting with programming tasks such as code generation and completion. The project includes model checkpoints, configuration files, and example code that enable developers to run and experiment with the models locally or within machine learning frameworks. These models are designed specifically for coding workflows and are trained on large datasets of source code covering many programming languages and development environments. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Asteroid

    Asteroid

    The PyTorch-based audio source separation toolkit for researchers

    ...Recipes provide an easy way to reproduce results with data preparation, system design, training and evaluation in a single script. This is an essential tool for the community! The default logger is TensorBoard in all the recipes. From the recipe folder, you can run the following to visualize the logs of all your runs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    GPT-Code UI

    GPT-Code UI

    An open source implementation of OpenAI's ChatGPT Code interpreter

    An open source implementation of OpenAI's ChatGPT Code interpreter. Simply ask the OpenAI model to do something and it will generate & execute the code for you. You can put a .env in the working directory to load the OPENAI_API_KEY environment variable. For Azure OpenAI Services, there are also other configurable variables like deployment name. See .env.azure-example for more information. Note that model selection on the UI is currently not supported for Azure OpenAI Services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    LLaMA

    LLaMA

    Inference code for Llama models

    “Llama” is the repository from Meta (formerly Facebook/Meta Research) containing the inference code for LLaMA (Large Language Model Meta AI) models. It provides utilities to load pre-trained LLaMA model weights, run inference (text generation, chat, completions), and work with tokenizers. Tokenizer utilities, download scripts, shell helpers to fetch model weights with correct licensing/permissions. Includes example scripts for chat completions and text completions to show how to call the models in code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 10
    fastquant

    fastquant

    Backtest and optimize your ML trading strategies with only 3 lines

    fastquant is a Python library designed to simplify quantitative financial analysis and algorithmic trading strategy development. The project focuses on making backtesting accessible by providing a high-level interface that allows users to test investment strategies with only a few lines of code. It integrates historical market data sources and trading frameworks so that users can quickly build experiments without constructing complex data pipelines. The framework enables users to test common...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    StoryTeller

    StoryTeller

    Multimodal AI Story Teller, built with Stable Diffusion, GPT, etc.

    A multimodal AI story teller, built with Stable Diffusion, GPT, and neural text-to-speech (TTS). Given a prompt as an opening line of a story, GPT writes the rest of the plot; Stable Diffusion draws an image for each sentence; a TTS model narrates each line, resulting in a fully animated video of a short story, replete with audio and visuals. To develop locally, install dev dependencies and install pre-commit hooks. This will automatically trigger linting and code quality checks before each...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Neural Network Intelligence

    Neural Network Intelligence

    AutoML toolkit for automate machine learning lifecycle

    Neural Network Intelligence is an open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning. NNI (Neural Network Intelligence) is a lightweight but powerful toolkit to help users automate feature engineering, neural architecture search, hyperparameter tuning and model compression. The tool manages automated machine learning (AutoML) experiments, dispatches and runs experiments'...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ChatGLM Efficient Tuning

    ChatGLM Efficient Tuning

    Fine-tuning ChatGLM-6B with PEFT

    ChatGLM-Efficient-Tuning is a hands-on toolkit for fine-tuning ChatGLM-family models with parameter-efficient methods on everyday hardware. It wraps techniques like LoRA and prompt-tuning into simple training scripts so you can adapt a large model to your domain without full retraining. The project exposes practical switches for quantization and mixed precision, allowing bigger models to fit into limited VRAM. It includes examples for instruction tuning and dialogue datasets, making it...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    mindflow

    mindflow

    AI-powered CLI git wrapper, boilerplate code generator, chat history

    I-powered CLI git wrapper, boilerplate code generator, chat history manager, and code search engine to streamline your dev workflow. The ChatGPT-powered swiss army knife for the modern developer! We provide an AI-powered CLI git wrapper, boilerplate code generator, code search engine, a conversation history manager, and much more! Configure the model used for generating responses by running mf config and selecting either GPT 3.5 Turbo (default) or GPT 4. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Horovod

    Horovod

    Distributed training framework for TensorFlow, Keras, PyTorch, etc.

    Horovod was originally developed by Uber to make distributed deep learning fast and easy to use, bringing model training time down from days and weeks to hours and minutes. With Horovod, an existing training script can be scaled up to run on hundreds of GPUs in just a few lines of Python code. Horovod can be installed on-premise or run out-of-the-box in cloud platforms, including AWS, Azure, and Databricks. Horovod can additionally run on top of Apache Spark, making it possible to unify data processing and model training into a single pipeline. Once Horovod has been configured, the same infrastructure can be used to train models with any framework, making it easy to switch between TensorFlow, PyTorch, MXNet, and future frameworks as machine learning tech stacks continue to evolve. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Repo of Tree of Thoughts (ToT)

    Repo of Tree of Thoughts (ToT)

    Implementation of "Tree of Thoughts

    Language models are increasingly being deployed for general problem-solving across a wide range of tasks, but are still confined to token-level, left-to-right decision-making processes during inference. This means they can fall short in tasks that require exploration, strategic lookahead, or where initial decisions play a pivotal role. To surmount these challenges, we introduce a new framework for language model inference, Tree of Thoughts (ToT), which generalizes over the popular Chain of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    TensorFlow Documentation

    TensorFlow Documentation

    TensorFlow documentation

    An end-to-end platform for machine learning. TensorFlow makes it easy to create ML models that can run in any environment. Learn how to use the intuitive APIs through interactive code samples.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    minimalRL-pytorch

    minimalRL-pytorch

    Implementations of basic RL algorithms with minimal lines of codes

    minimalRL is a lightweight reinforcement learning repository that implements several classic algorithms using minimal PyTorch code. The project is designed primarily as an educational resource that demonstrates how reinforcement learning algorithms work internally without the complexity of large frameworks. Each algorithm implementation is contained within a single file and typically ranges from about 100 to 150 lines of code, making it easy for learners to inspect the entire implementation at once. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    LM Human Preferences

    LM Human Preferences

    Code for the paper Fine-Tuning Language Models from Human Preferences

    ...The repository includes scripts to train the reward model (learning to rank or score pairs of outputs), and to fine-tune a policy (a language model) with reinforcement learning (or related techniques) guided by that reward model. The code is provided “as is” and explicitly says it may no longer run out-of-the-box due to dependencies or dataset migrations. It was tested on the smallest GPT-2 (124M parameters) under a specific environment (TensorFlow 1.x, specific CUDA / cuDNN combinations). It includes utilities for launching experiments, sampling from policies, and simple experiment orchestration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Img2Txt

    Img2Txt

    Img2Txt - Extract Text From Images using AI

    Important: If you are sharing this program. Please Include the official Download Link What is Img2Txt? Img2Txt is a Python-based application packaged using PyInstaller that utilizes the power of pytesseract, an AI-powered optical character recognition (OCR) library, to extract text from images and convert it into plain text. The application features a simple and modern user-friendly interface created using customtkinter, allowing users to easily process images and obtain the text...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Bot on Anything

    Bot on Anything

    Large model-based chatbot builder that can quickly integrate AI models

    Bot on Anything is a versatile open-source AI chatbot builder that lets developers connect large language models such as ChatGPT, Claude, and Gemini to virtually any messaging platform, website, or interface with minimal configuration. At its heart, the project abstracts away the glue logic between AI model APIs and disparate application “channels,” enabling the same bot logic to run in Slack, Telegram, Gmail, enterprise tools, web UIs, or command-line terminals. Configuration is handled simply through a central JSON file where you define which model and which application channel you want to glue together, so developers can create sophisticated AI assistants without rewriting integration code from scratch. The architecture emphasizes reusability and extensibility, allowing the addition of new model backends or new channels with relative ease. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    NOW

    NOW

    No-code tool for creating a neural search solution in minutes

    One line to host them all. Bootstrap your multimodal search case in minutes. NOW gives the world access to multimodal neural search with just one command. NOW supports various formats for uploading your dataset to your search application. You may either choose a demo dataset hosted by NOW, or use your own custom dataset, to build an application. NOW can support your custom data in the form of a DocumentArray, as a path to a local folder, or S3 bucket. You can choose a demo dataset to get...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Mars Framework

    Mars Framework

    Mars is a tensor-based unified framework for large-scale data

    ...The project provides a tensor-based execution model that extends the capabilities of tools such as NumPy, pandas, and scikit-learn so that large datasets can be processed in parallel without rewriting code for distributed environments. Its architecture automatically divides large computational tasks into smaller chunks that can be executed across multiple nodes in a cluster, allowing complex analytics, machine learning workflows, and data transformations to run efficiently at scale. Mars is particularly useful for workloads that exceed the memory capacity of a single machine or require high levels of parallel processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    CleanRL

    CleanRL

    High-quality single file implementation of Deep Reinforcement Learning

    CleanRL is a Deep Reinforcement Learning library that provides high-quality single-file implementation with research-friendly features. The implementation is clean and simple, yet we can scale it to run thousands of experiments using AWS Batch. CleanRL is not a modular library and therefore it is not meant to be imported. At the cost of duplicate code, we make all implementation details of a DRL algorithm variant easy to understand, so CleanRL comes with its own pros and cons. You should consider using CleanRL if you want to 1) understand all implementation details of an algorithm's variant or 2) prototype advanced features that other modular DRL libraries do not support (CleanRL has minimal lines of code so it gives you great debugging experience and you don't have to do a lot of subclassing like sometimes in modular DRL libraries).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Minimal text diffusion

    Minimal text diffusion

    A minimal implementation of diffusion models for text generation

    A minimal implementation of diffusion models of text: learns a diffusion model of a given text corpus, allowing to generate text samples from the learned model. The main idea was to retain just enough code to allow training a simple diffusion model and generating samples, remove image-related terms, and make it easier to use. To train a model, run scripts/train.sh. By default, this will train a model on the simple corpus. However, you can change this to any text file using the --train_data argument. Note that you may have to increase the sequence length (--seq_len) if your corpus is longer than the simple corpus. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB