Open Source Python Artificial Intelligence Software - Page 67

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    Claude Code Bridge

    Claude Code Bridge

    Real-time multi-AI collaboration: Claude, Codex & Gemini

    Claude Code Bridge is an open-source command-line tool designed to enable real-time collaboration between multiple AI coding assistants within a unified development environment. The system allows developers to coordinate interactions between models such as Claude, Codex, and Gemini so that they can work together on programming tasks. By maintaining persistent shared context between these models, the tool reduces redundant prompts and minimizes token usage while allowing each AI system to contribute specialized capabilities. The architecture functions as a unified launcher that manages communication between multiple AI providers and coordinates their responses within the same development session. Developers can run the tool in terminal environments and integrate it with terminal multiplexers such as tmux or advanced terminal emulators.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Claude Code Hooks Mastery

    Claude Code Hooks Mastery

    Master Claude Code Hooks

    Claude Code Hooks Mastery is a trending community-centric GitHub repository aimed at helping developers master Claude Code hooks — customizable integration points that let users extend, automate, and augment workflows when using Claude Code, an agentic terminal coding assistant. Although the project itself doesn’t include a single coherent application, it functions as a curated collection of advanced hook examples, best practices, and coding patterns that show how to tailor Claude Code to specific use cases such as automated CI workflows, custom command triggers, and integrations with external tools. The repository is part of a larger ecosystem of Claude Code tooling that enables natural-language-driven coding tasks, and the hooks contained here help users go beyond default behaviors to solve real problems efficiently.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Claude Cognitive

    Claude Cognitive

    Persistent context and multi-instance coordination

    Claude Cognitive is an advanced memory and context-management extension designed to address the stateless limitations of Claude Code by giving the model a form of persistent “working memory” and multi-instance coordination. It introduces an attention-based context router that prioritizes files and content relevant to the current development discussion — tagging them as HOT, WARM, or COLD based on recency and keyword activation — so Claude Code doesn’t waste token budget rereading irrelevant code. This context routing dramatically reduces redundant token usage and accelerates large codebase interactions by focusing only on what truly matters to the current task. Additionally, Claude-Cognitive includes a pool coordinator to share state across multiple Claude Code instances, preserving what’s been learned or completed and preventing repetitive debugging or redundant exploration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    CleverHans

    CleverHans

    An adversarial example library for constructing attacks

    This repository contains the source code for CleverHans, a Python library to benchmark machine learning systems' vulnerability to adversarial examples. You can learn more about such vulnerabilities on the accompanying blog. The CleverHans library is under continual development, always welcoming contributions of the latest attacks and defenses. In particular, we always welcome help with resolving the issues currently open. Since v4.0.0, CleverHans supports 3 frameworks: JAX, PyTorch, and TF2. We are currently prioritizing implementing attacks in PyTorch, but we very much welcome contributions for all 3 frameworks. In versions v3.1.0 and prior, CleverHans supported TF1; the code for v3.1.0 can be found under cleverhans_v3.1.0/ or by checking out a prior Github release. The library focuses on providing a reference implementation of attacks against machine learning models to help with benchmarking models against adversarial examples.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Clipper is a Python based Shell and Debugger for the CLIPS Expert System (see http://clipsrules.sourceforge.net/). It is currently build on pyGTK and pyCLIPS.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    CoTracker

    CoTracker

    CoTracker is a model for tracking any point (pixel) on a video

    CoTracker is a learning-based point tracking system that jointly follows many user-specified points across a video, rather than tracking each point independently. By reasoning about all tracks together, it can maintain temporal consistency, handle mutual occlusions, and reduce identity swaps when trajectories cross. The model takes sparse point queries on one frame and predicts their sub-pixel locations and a visibility score for every subsequent frame, producing long, coherent trajectories. Its transformer-style architecture aggregates information both along time and across points, allowing it to recover tracks even after brief disappearances. The repository ships with inference scripts, pretrained weights, and simple interfaces to seed points, run tracking, and export trajectories for downstream tasks. Typical uses include correspondence building, motion analysis, dynamic SLAM priors, video editing masks, and evaluation of geometric consistency in real scenes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Coach

    Coach

    Enables easy experimentation with state of the art algorithms

    Coach is a python framework that models the interaction between an agent and an environment in a modular way. With Coach, it is possible to model an agent by combining various building blocks, and training the agent on multiple environments. The available environments allow testing the agent in different fields such as robotics, autonomous driving, games and more. It exposes a set of easy-to-use APIs for experimenting with new RL algorithms and allows simple integration of new environments to solve. Coach collects statistics from the training process and supports advanced visualization techniques for debugging the agent being trained. Coach supports many state-of-the-art reinforcement learning algorithms, which are separated into three main classes - value optimization, policy optimization, and imitation learning. Coach supports a large number of environments which can be solved using reinforcement learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Code World Model (CWM)

    Code World Model (CWM)

    Research code artifacts for Code World Model (CWM)

    CWM (Code World Model) is a 32-billion-parameter open-weights language model. It is developed by Meta for enhancing code generation and reasoning about programs. It is explicitly trained on execution traces, action-observation trajectories, and agentic interactions in controlled environments. It has been developed to better capture how code, actions, and state interact over time. The repository provides inference code, reproducibility scripts, prompt guides, and more. It has model cards, utilities, demos, and evaluation artifacts. Inference scripts and utilities for code generation tasks. Evaluation benchmarks on code, mathematics, and reasoning tasks. Demos, serving code, and evaluation pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Code-Mode

    Code-Mode

    Plug-and-play library to enable agents to call MCP and UTCP tools

    Code-Mode is a plug-and-play library that lets AI agents call tools by executing TypeScript (or via a Python wrapper) instead of making many individual function calls. Its core philosophy is that language models are very good at writing code, so rather than exposing hundreds of separate tool endpoints, you give the model a single “code execution” tool that has access to your full toolkit through code. This approach can dramatically reduce the number of tool-call iterations needed in complex workflows, turning multi-step call chains into a single code execution with internal branching and loops. The repository contains both TypeScript and Python libraries, plus a code-mode-mcp component for integrating with MCP and UTCP ecosystems. Benchmarks in the README highlight improvements in latency and token cost for scenarios involving multiple tools, showing that code execution often outperforms traditional JSON-based function calling.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    CodeGeeX4

    CodeGeeX4

    CodeGeeX4-ALL-9B, a versatile model for all AI software development

    CodeGeeX4 is the fourth-generation open source multilingual code large language model (LLM) developed by ZhipuAI. Designed as a powerful AI coding assistant, it supports over 100 programming languages and has been trained on a massive code and natural language corpus. Compared to its predecessors, CodeGeeX4 introduces improved reasoning, stronger alignment with developer needs, and better performance on real-world programming benchmarks. It supports tasks such as code completion, generation from natural language descriptions, code translation, bug fixing, and explanation. The repository provides model checkpoints, inference examples, and fine-tuning guides, making it adaptable for both research and practical software development workflows. With its open release, CodeGeeX4 aims to provide a transparent alternative to proprietary coding assistants while advancing the field of AI-assisted programming.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    CogAgent

    CogAgent

    An open sourced end-to-end VLM-based GUI Agent

    CogAgent is a 9B-parameter bilingual vision-language GUI agent model based on GLM-4V-9B, trained with staged data curation, optimization, and strategy upgrades to improve perception, action prediction, and generalization across tasks. It focuses on operating real user interfaces from screenshots plus text, and follows a strict input–output format that returns structured actions, grounded operations, and optional sensitivity annotations. The model is designed for agent-style execution rather than freeform chat, maintaining a continuous execution history across steps while requiring a fresh session for each new task. Inference supports BF16 on NVIDIA GPUs, with optional INT8 and INT4 modes available but with noted performance loss at INT4; example CLIs and a web demo illustrate bounding-box outputs and operation categories.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    CogView

    CogView

    Text-to-Image generation. The repo for NeurIPS 2021 paper

    CogView is a large-scale pretrained text-to-image transformer model, introduced in the NeurIPS 2021 paper CogView: Mastering Text-to-Image Generation via Transformers. With 4 billion parameters, it was one of the earliest transformer-based models to successfully generate high-quality images from natural language descriptions in Chinese, with partial support for English via translation. The model incorporates innovations such as PB-relax and Sandwich-LN to enable stable training of very deep transformers without NaN loss issues. CogView supports multiple tasks beyond text-to-image, including image captioning, post-selection (ranking candidate images by relevance to a prompt), and super-resolution (upscaling model-generated images). The repository provides pretrained models, inference scripts, and training examples, along with a Docker environment for reproducibility.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    CogView4

    CogView4

    CogView4, CogView3-Plus and CogView3(ECCV 2024)

    CogView4 is the latest generation in the CogView series of vision-language foundation models, developed as a bilingual (Chinese and English) open-source system for high-quality image understanding and generation. Built on top of the GLM framework, it supports multimodal tasks including text-to-image synthesis, image captioning, and visual reasoning. Compared to previous CogView versions, CogView4 introduces architectural upgrades, improved training pipelines, and larger-scale datasets, enabling stronger alignment between textual prompts and generated visual content. It emphasizes bilingual usability, making it well-suited for cross-lingual multimodal applications. The model also supports fine-tuning and downstream customization, extending its applicability to creative content generation, human–computer interaction, and research on vision-language alignment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Colossal-AI

    Colossal-AI

    Making large AI models cheaper, faster and more accessible

    The Transformer architecture has improved the performance of deep learning models in domains such as Computer Vision and Natural Language Processing. Together with better performance come larger model sizes. This imposes challenges to the memory wall of the current accelerator hardware such as GPU. It is never ideal to train large models such as Vision Transformer, BERT, and GPT on a single GPU or a single machine. There is an urgent demand to train models in a distributed environment. However, distributed training, especially model parallelism, often requires domain expertise in computer systems and architecture. It remains a challenge for AI researchers to implement complex distributed training solutions for their models. Colossal-AI provides a collection of parallel components for you. We aim to support you to write your distributed deep learning models just like how you write your model on your laptop.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Command Line Parser GetPot

    Command Line Parser GetPot

    Tool to parse the command line and configuration files.

    Powerful command line and configuration file parsing for C++, Python, Ruby and Java (others to come). This tool provides many features, such as separate treatment for options, variables, and flags, unrecognized object detection, prefixes and much more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    CommunityBackup

    ...a work in progress.

    Setup files are quite old.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Compose

    Compose

    A machine learning tool for automated prediction engineering

    Compose is a machine learning tool for automated prediction engineering. It allows you to structure prediction problems and generate labels for supervised learning. An end user defines an outcome of interest by writing a labeling function, then runs a search to automatically extract training examples from historical data. Its result is then provided to Featuretools for automated feature engineering and subsequently to EvalML for automated machine learning. Prediction problems are structured by using a label maker and a labeling function. The label maker automatically extracts data along the time index to generate labels. The process starts by setting the first cutoff time after the minimum amount of data. Then subsequent cutoff times are spaced apart using gaps. Starting from each cutoff time, a window determines the amount of data, also referred to as a data slice, to pass into a labeling function.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Consistency Models

    Consistency Models

    Official repo for consistency models

    consistency_models is the repository for Consistency Models, a new family of generative models introduced by OpenAI that aim to generate high-quality samples by mapping noise directly into data — circumventing the need for lengthy diffusion chains. It builds on and extends diffusion model frameworks (e.g. based on the guided-diffusion codebase), adding techniques like consistency distillation and consistency training to enable fast, often one-step, sample generation. The repo is implemented in PyTorch and includes support for large-scale experiments on datasets like ImageNet-64 and LSUN variants. It also contains checkpointed models, evaluation scripts, and variants of sampling / editing algorithms described in the paper. Because consistency models reduce the number of inference steps, they are promising for real-time or low-latency generative systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ConsistencyDecoder

    ConsistencyDecoder

    Consistency Distilled Diff VAE

    ConsistencyDecoder is a Python package from OpenAI that introduces an improved decoding method for variational autoencoders (VAEs) used in Stable Diffusion pipelines. Instead of relying solely on the standard GAN or VAE decoder, this approach leverages a Consistency Distilled Diff VAE, designed to produce higher-quality and more stable outputs from encoded latents. The project provides a simple API for encoding with a Stable Diffusion VAE and decoding using the new consistency model, allowing for side-by-side comparisons with traditional decoders. It demonstrates how consistency models can enhance visual fidelity while maintaining efficiency, reducing artifacts common in GAN-decoded outputs. The repository includes installation instructions, usage examples, and visual comparisons to highlight improvements. Though compact, it provides an accessible entry point for experimenting with advanced decoding strategies in diffusion-based generative models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Consistent Depth

    Consistent Depth

    We estimate dense, flicker-free, geometrically consistent depth

    Consistent Depth is a research project developed by Facebook Research that presents an algorithm for reconstructing dense and geometrically consistent depth information for all pixels in a monocular video. The system builds upon traditional structure-from-motion (SfM) techniques to provide geometric constraints while integrating a convolutional neural network trained for single-image depth estimation. During inference, the model fine-tunes itself to align with the geometric constraints of a specific input video, ensuring stable and realistic depth maps even in less-constrained regions. This approach achieves improved geometric consistency and visual stability compared to prior monocular reconstruction methods. The project can process challenging hand-held video footage, including those with moderate dynamic motion, making it practical for real-world usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Context Engineering

    Context Engineering

    A frontier, first-principles handbook

    Context Engineering is a comprehensive, open-source project serving as a first-principles handbook for the emerging discipline of context design and optimization in AI. Moving beyond traditional prompt engineering, this repository defines and explores how to craft and provide complete context payloads — not just single prompts — to large language models so they can perform tasks more reliably and intelligently. It takes inspiration from thought leaders like Andrej Karpathy and bridges theory with practical examples, offering structured guidance on context orchestration, memory, retrieval, and state control within AI workflows. With extensive materials drawn from research, surveys, and visual explanations, the project acts as both a learning resource and a reference for practitioners looking to improve model behavior by engineering richer inputs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Context Engineering Template

    Context Engineering Template

    Context engineering is the new vibe coding

    Context Engineering Template is a comprehensive template and workflow repository designed to teach and implement context engineering, a structured approach to preparing and organizing the information necessary for AI coding assistants to complete complex tasks reliably. Instead of relying solely on short prompts, this project encourages developers to create rich, structured context files that include project rules, examples, and validation criteria so that AI systems can act more like informed collaborators and less like general-purpose generators. The repository provides templates such as CLAUDE.md for defining global project rules, INITIAL.md for feature requests, and folders for examples, PRPs, validation scripts, and settings to support systematic prompt generation and execution with tools like Claude Code. By using this template, teams can ensure consistency across AI outputs, reduce errors that stem from contextual misunderstandings, and build reusable patterns.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Continuous Claude v3

    Continuous Claude v3

    Context management for Claude Code. Hooks maintain state via ledgers

    Continuous Claude v3 is a persistent, multi-agent development environment built around the Claude Code CLI that aims to overcome the limitations of standard LLM context windows. Rather than relying on a single session’s context, Continuous Claude uses mechanisms like ledgers, YAML handoffs, and a memory system to preserve and recall state across multiple sessions, ensuring that learned insights and plans are not lost when context compaction occurs. The project orchestrates many specialized agents and skills—109 skills and 32 agents—so that complex coding tasks can be broken down, analyzed, and executed collaboratively by different components. It also includes a layered code analysis pipeline to reduce token usage and maintain relevant context efficiently. This continuous learning environment enables workflows such as bug fixing, refactoring, planning, and exploratory investigation while minimizing the need to re-explain context manually.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Controllable-RAG-Agent

    Controllable-RAG-Agent

    This repository provides an advanced RAG

    Controllable-RAG-Agent is an advanced Retrieval-Augmented Generation (RAG) system designed specifically for complex, multi-step question answering over your own documents. Instead of relying solely on simple semantic search, it builds a deterministic control graph that acts as the “brain” of the agent, orchestrating planning, retrieval, reasoning, and verification across many steps. The pipeline ingests PDFs, splits them into chapters, cleans and preprocesses text, then constructs vector stores for fine-grained chunks, chapter summaries, and book quotes to support nuanced queries. At query time, it anonymizes entities, creates a high-level plan, de-anonymizes and expands that plan into concrete retrieval or reasoning tasks, and executes them in sequence while continuously revising the plan. A key focus is hallucination control: each answer is verified against retrieved context, and responses are reworked when they are not sufficiently grounded in the source documents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    ConvNeXt

    ConvNeXt

    Code release for ConvNeXt model

    ConvNeXt is a modernized convolutional neural network (CNN) architecture designed to rival Vision Transformers (ViTs) in accuracy and scalability while retaining the simplicity and efficiency of CNNs. It revisits classic ResNet-style backbones through the lens of transformer design trends—large kernel sizes, inverted bottlenecks, layer normalization, and GELU activations—to bridge the performance gap between convolutions and attention-based models. ConvNeXt’s clean, hierarchical structure makes it efficient for both pretraining and fine-tuning across a wide range of visual recognition tasks. It achieves competitive or superior results on ImageNet and downstream datasets while being easier to deploy and train than transformers. The repository provides pretrained models, training recipes, and ablation studies demonstrating how incremental design choices collectively yield state-of-the-art performance.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB