Showing 641 open source projects for "ai code agent"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    LLaMA

    LLaMA

    Inference code for Llama models

    “Llama” is the repository from Meta (formerly Facebook/Meta Research) containing the inference code for LLaMA (Large Language Model Meta AI) models. It provides utilities to load pre-trained LLaMA model weights, run inference (text generation, chat, completions), and work with tokenizers. Tokenizer utilities, download scripts, shell helpers to fetch model weights with correct licensing/permissions. Includes example scripts for chat completions and text completions to show how to call the models in code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    PromethAI

    PromethAI

    Open-source framework that gives you AI Agents

    PromethAI-Backend is a backend framework for AI-driven automation and knowledge extraction. It is designed to integrate with large language models (LLMs) to provide AI-enhanced workflows, including content generation, summarization, and data analysis.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Free-Auto-GPT

    Free-Auto-GPT

    Free AutoGPT enables autonomous AI tasks without paid APIs

    Free Auto GPT is an open source project that delivers a simplified version of an autonomous AI agent capable of completing tasks independently. It allows users to run an AutoGPT-style system without relying on paid OpenAI APIs, making it more accessible for experimentation and personal use. Free Auto GPT can take a goal, break it into smaller steps, and execute actions in a loop to achieve results with minimal human input. Designed for ease of use, the project focuses on removing cost barriers while still demonstrating how autonomous agents function. ...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 4
    Lightning-Hydra-Template

    Lightning-Hydra-Template

    PyTorch Lightning + Hydra. A very user-friendly template

    ...Not fitted for data engineering - the template configuration setup is not designed for building data processing pipelines that depend on each other. PyTorch Lightning, a lightweight PyTorch wrapper for high-performance AI research. Think of it as a framework for organizing your PyTorch code. Hydra, a framework for elegantly configuring complex applications. The key feature is the ability to dynamically create a hierarchical configuration by composition and override it through config files and the command line.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 5
    ChatGLM Efficient Tuning

    ChatGLM Efficient Tuning

    Fine-tuning ChatGLM-6B with PEFT

    ChatGLM-Efficient-Tuning is a hands-on toolkit for fine-tuning ChatGLM-family models with parameter-efficient methods on everyday hardware. It wraps techniques like LoRA and prompt-tuning into simple training scripts so you can adapt a large model to your domain without full retraining. The project exposes practical switches for quantization and mixed precision, allowing bigger models to fit into limited VRAM. It includes examples for instruction tuning and dialogue datasets, making it...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    mindflow

    mindflow

    AI-powered CLI git wrapper, boilerplate code generator, chat history

    I-powered CLI git wrapper, boilerplate code generator, chat history manager, and code search engine to streamline your dev workflow. The ChatGPT-powered swiss army knife for the modern developer! We provide an AI-powered CLI git wrapper, boilerplate code generator, code search engine, a conversation history manager, and much more! Configure the model used for generating responses by running mf config and selecting either GPT 3.5 Turbo (default) or GPT 4. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Bard

    Bard

    Python SDK/API for reverse engineered Google Bard

    The Bard repository from acheong08 is an unofficial client library and set of tools designed to interact programmatically with Google Bard, Google’s conversational AI service, allowing developers to automate queries, integrate Bard into their own applications, and experiment with scripting around Bard-style prompts without relying on a formal, official API. Because Google doesn’t provide a standard public API for Bard, this project works by reverse-engineering the web interactions that power Bard’s interface and exposing them through code-friendly interfaces, enabling chat automation, custom workflows, and bot creation. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 8
    ParlAI

    ParlAI

    A framework for training and evaluating AI models

    ParlAI is a comprehensive research platform for building, training, and evaluating dialogue agents across a wide variety of tasks and datasets. It provides a unified interface—agents, teachers, and worlds—so the same model can be trained on multi-turn chit-chat, question answering, task-oriented dialogue, retrieval, or safety-focused datasets without changing core code. The library integrates tightly with PyTorch and supports both generative and retrieval-augmented models, along with...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 9
    langchain-prefect

    langchain-prefect

    Tools for using Langchain with Prefect

    Large Language Models (LLMs) are interesting and useful  -  building apps that use them responsibly feels like a no-brainer. Tools like Langchain make it easier to build apps using LLMs. We need to know details about how our apps work, even when we want to use tools with convenient abstractions that may obfuscate those details. Prefect is built to help data people build, run, and observe event-driven workflows wherever they want. It provides a framework for creating deployments on a whole...
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    unit-minions

    unit-minions

    AI R&D Efficiency Improvement Research: Do-It-Yourself Training LoRA

    "AI R&D Efficiency Improvement Research: Do-It-Yourself Training LoRA", including Llama (Alpaca LoRA) model, ChatGLM (ChatGLM Tuning) related Lora training. Training content: user story generation, test code generation, code-assisted generation, text to SQL, text generation code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    DiT (Diffusion Transformers)

    DiT (Diffusion Transformers)

    Official PyTorch Implementation of "Scalable Diffusion Models"

    DiT (Diffusion Transformer) is a powerful architecture that applies transformer-based modeling directly to diffusion generative processes for high-quality image synthesis. Unlike CNN-based diffusion models, DiT represents the diffusion process in the latent space and processes image tokens through transformer blocks with learned positional encodings, offering scalability and superior sample quality. The model architecture parallels large language models but for image tokens—each block...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Bot on Anything

    Bot on Anything

    Large model-based chatbot builder that can quickly integrate AI models

    ...Configuration is handled simply through a central JSON file where you define which model and which application channel you want to glue together, so developers can create sophisticated AI assistants without rewriting integration code from scratch. The architecture emphasizes reusability and extensibility, allowing the addition of new model backends or new channels with relative ease. It supports switching between multiple AI models and targets within the same project.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    VALL-E

    VALL-E

    PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)

    We introduce a language modeling approach for text to speech synthesis (TTS). Specifically, we train a neural codec language model (called VALL-E) using discrete codes derived from an off-the-shelf neural audio codec model, and regard TTS as a conditional language modeling task rather than continuous signal regression as in previous work. During the pre-training stage, we scale up the TTS training data to 60K hours of English speech which is hundreds of times larger than existing systems....
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    Multi-Agent Particle Envs

    Multi-Agent Particle Envs

    Code for a multi-agent particle environment used in a paper

    ...Scenarios are designed to model cooperative, competitive, and mixed interactions among agents, making it useful for testing algorithms in multi-agent settings. The project includes built-in scenarios such as navigation to landmarks, cooperative tasks, and adversarial setups. Although archived, its concepts and code structure remain foundational for more advanced libraries like PettingZoo, which extended and maintained this environment.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 15
    AEA Framework

    AEA Framework

    A framework for autonomous economic agent (AEA) development

    agents-aea by Fetch.ai is a framework for building autonomous economic agents (AEAs) that can act independently, communicate, and transact on decentralized networks. It focuses on enabling AI-driven agents to participate in digital marketplaces and ecosystems.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    Img2Txt

    Img2Txt

    Img2Txt - Extract Text From Images using AI

    Important: If you are sharing this program. Please Include the official Download Link What is Img2Txt? Img2Txt is a Python-based application packaged using PyInstaller that utilizes the power of pytesseract, an AI-powered optical character recognition (OCR) library, to extract text from images and convert it into plain text. The application features a simple and modern user-friendly interface created using customtkinter, allowing users to easily process images and obtain the text...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    artikelschreiber

    artikelschreiber

    Frontend and Backend Code for ArtikelSchreiber.com and UNAIQUE.NET

    Frontend and Backend Code for ArtikelSchreiber.com and UNAIQUE.NET Text Generator deutsch - Dein KI Text Generator kostenlos mit Künstlicher Intelligenz The Software as a Service can be found here: SEO Optimizer: Ghost Writer - Hausarbeiten schreiben mit KI and KI Text Generator This product includes software developed by Sebastian Enger, M.Sc. Copyright (c) 2023, Sebastian Enger, M.Sc. All rights reserved. Frontend and Backend Source Code for...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    DeepMozart

    DeepMozart

    Audio generation using diffusion models

    Audio generation using diffusion models in PyTorch. The code is based on the audio-diffusion-pytorch repository.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ConvNeXt V2

    ConvNeXt V2

    Code release for ConvNeXt V2 model

    ConvNeXt V2 is an evolution of the ConvNeXt architecture that co-designs convolutional networks alongside self-supervised learning. The V2 version introduces a fully convolutional masked autoencoder (FCMAE) framework where parts of the image are masked and the network reconstructs the missing content, marrying convolutional inductive bias with powerful pretraining. A key innovation is a new Global Response Normalization (GRN) layer added to the ConvNeXt backbone, which enhances feature...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    CodeContests, developed by Google DeepMind, is a large-scale competitive programming dataset designed for training and evaluating machine learning models on code generation and problem solving. This dataset played a central role in the development of AlphaCode, DeepMind’s model for solving programming problems at a human-competitive level, as published in Science. CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder,...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 21
    minGPT

    minGPT

    A minimal PyTorch re-implementation of the OpenAI GPT

    minGPT is a minimalist, educational re-implementation of the GPT (Generative Pretrained Transformer) architecture built in PyTorch, designed by Andrej Karpathy to expose the core structure of a transformer-based language model in as few lines of code as possible. It strips away extraneous bells and whistles, aiming to show how a sequence of token indices is fed into a stack of transformer blocks and then decoded into the next token probabilities, with both training and inference supported....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    CPT

    CPT

    CPT: A Pre-Trained Unbalanced Transformer

    A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation. We replace the old BERT vocabulary with a larger one of size 51271 built from the training data, in which we 1) add missing 6800+ Chinese characters (most of them are traditional Chinese characters); 2) remove redundant tokens (e.g. Chinese character tokens with ## prefix); 3) add some English tokens to reduce OOV. Position Embeddings We extend the max_position_embeddings from 512 to 1024. We...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Minimal text diffusion

    Minimal text diffusion

    A minimal implementation of diffusion models for text generation

    A minimal implementation of diffusion models of text: learns a diffusion model of a given text corpus, allowing to generate text samples from the learned model. The main idea was to retain just enough code to allow training a simple diffusion model and generating samples, remove image-related terms, and make it easier to use. To train a model, run scripts/train.sh. By default, this will train a model on the simple corpus. However, you can change this to any text file using the --train_data...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    DialoGPT

    DialoGPT

    Large-scale pretraining for dialogue

    ...The model was trained on a massive dataset of approximately 147 million conversational exchanges extracted from Reddit discussion threads, allowing it to learn patterns of natural human conversation. DialoGPT provides multiple pretrained model sizes and includes code for training, fine-tuning, and evaluating dialogue generation models. The repository also contains scripts for preparing conversation datasets and reproducing experimental benchmarks related to conversational AI research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Video Pre-Training

    Video Pre-Training

    Learning to Act by Watching Unlabeled Online Videos

    ...The repository contains demonstration models of different widths, fine-tuned variants (e.g. for building houses or early-game tasks), and inference scripts that instantiate agents from pretrained weights. Key modules include the behavioral cloning logic, the agent wrapper, and data loading pipelines (with an accessible skeleton for loading Minecraft demonstration data). The repo also includes a run_agent.py script for testing an agent interactively, and an agent.py module encapsulating the control logic.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB