Open Source Python Artificial Intelligence Software - Page 17

Python Artificial Intelligence Software

View 13534 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    BEIR

    BEIR

    A Heterogeneous Benchmark for Information Retrieval

    BEIR is a benchmark framework for evaluating information retrieval models across various datasets and tasks, including document ranking and question answering.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    Binary Ninja MCP

    Binary Ninja MCP

    A Binary Ninja plugin, MCP server

    The Binary Ninja MCP is a plugin and bridge that integrates Binary Ninja with Large Language Model clients via the Model Context Protocol, enhancing reverse engineering workflows with AI assistance. ​
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    Chat with LLMs Everywhere

    Chat with LLMs Everywhere

    Run PyTorch LLMs locally on servers, desktop and mobile

    TorchChat is an open-source project from the PyTorch ecosystem designed to demonstrate how large language models can be executed efficiently across different computing environments. The project provides a compact codebase that illustrates how to run conversational AI systems using PyTorch models on laptops, servers, and mobile devices. It is intended primarily as a reference implementation that shows developers how to integrate large language models into applications without requiring a large or complex infrastructure stack. TorchChat supports running models through Python interfaces as well as integrating them directly into native applications written in languages such as C or C++. The project also demonstrates how modern LLMs like LLaMA-style models can be deployed locally while maintaining good performance across different hardware platforms.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    ChatGLM2-6B

    ChatGLM2-6B

    ChatGLM2-6B: An Open Bilingual Chat LLM

    ChatGLM2-6B is the second-gen Chinese-English conversational LLM from ZhipuAI/Tsinghua. It upgrades the base model with GLM’s hybrid pretraining objective, 1.4 TB bilingual data, and preference alignment—delivering big gains on MMLU, CEval, GSM8K, and BBH. The context window extends up to 32K (FlashAttention), and Multi-Query Attention improves speed and memory use. The repo includes Python APIs, CLI & web demos, OpenAI-style/FASTAPI servers, and quantized checkpoints for lightweight local deployment on GPUs or CPU/MPS.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    ChatGLM3

    ChatGLM3

    ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat

    ChatGLM3 is ZhipuAI & Tsinghua KEG’s third-gen conversational model suite centered on the 6B-parameter ChatGLM3-6B. It keeps the series’ smooth dialog and low deployment cost while adding native tool use (function calling), a built-in code interpreter, and agent-style workflows. The family includes base and long-context variants (8K/32K/128K). The repo ships Python APIs, CLI and web demos (Gradio/Streamlit), an OpenAI-format API server, and a compact fine-tuning kit. Quantization (4/8-bit), CPU/MPS support, and accelerator backends (TensorRT-LLM, OpenVINO, chatglm.cpp) enable lightweight local or edge deployment.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 6
    ChatGPT Retrieval Plugin

    ChatGPT Retrieval Plugin

    The ChatGPT Retrieval Plugin lets you easily find personal documents

    The chatgpt-retrieval-plugin repository implements a semantic retrieval backend that lets ChatGPT (or GPT-powered tools) access private or organizational documents in natural language by combining vector search, embedding models, and plugin infrastructure. It can serve as a custom GPT plugin or function-calling backend so that a chat session can “look up” relevant documents based on user queries, inject those results into context, and respond more knowledgeably about a private knowledge base. The repo provides code for ingestion pipelines (embedding documents), APIs for querying, local server components, and privacy / PII detection modules. It also contains plugin manifest files (OpenAPI spec, plugin JSON) so that the retrieval backend can be registered in a plugin ecosystem. Because retrieval is often needed to make LLMs “know what’s in your docs” without leaking everything, this plugin aims to be a secure, flexible building block for retrieval-augmented generation (RAG) systems.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 7
    Claude Code Skills & Plugins Hub

    Claude Code Skills & Plugins Hub

    270+ Claude Code plugins with 739 agent skills

    Claude Code Plugins Plus Skills is a large open-source ecosystem of plugins and AI “skills” designed to extend the capabilities of Claude Code development agents. The repository functions as a marketplace-style collection of hundreds of plugins and specialized skills that enable Claude Code to perform complex development, automation, and operational tasks. These plugins cover a wide range of domains including DevOps automation, security testing, API debugging, infrastructure management, and AI workflow orchestration. The project also includes orchestration patterns and best practices that guide how multiple AI agents or tools can collaborate effectively in software development workflows. Developers can install plugins through a package-style plugin system and integrate them with their Claude Code environment using standardized commands.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 8
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    CodeContests, developed by Google DeepMind, is a large-scale competitive programming dataset designed for training and evaluating machine learning models on code generation and problem solving. This dataset played a central role in the development of AlphaCode, DeepMind’s model for solving programming problems at a human-competitive level, as published in Science. CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder, Codeforces, CodeChef, Aizu, and HackerEarth. Each problem includes structured metadata, problem descriptions, paired input/output test cases, and multiple correct and incorrect solutions in various programming languages. The dataset is distributed in Riegeli format using Protocol Buffers, with separate training, validation, and test splits for reproducible machine learning experiments.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    Cosmos-RL

    Cosmos-RL

    Cosmos-RL is a flexible and scalable Reinforcement Learning framework

    Cosmos-RL is a scalable reinforcement learning framework designed specifically for physical AI systems such as robotics, autonomous agents, and multimodal models. It provides a distributed training architecture that separates policy learning and environment rollout processes, enabling efficient and asynchronous reinforcement learning at scale. The framework supports multiple parallelism strategies, including tensor, pipeline, and data parallelism, allowing it to leverage large GPU clusters effectively. It is built with compatibility in mind, supporting popular model families such as LLaMA, Qwen, and diffusion-based world models, as well as integration with Hugging Face ecosystems. cosmos-rl also includes support for advanced RL algorithms, low-precision training, and fault-tolerant execution, making it suitable for large-scale production workloads.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    CowAgent

    CowAgent

    AI assistant based on large models that can actively think and plan

    CowAgent, based on the chatgpt-on-wechat project, is an open-source AI agent framework that integrates large language models into the WeChat ecosystem to create intelligent conversational assistants. It enables automated message handling by connecting WeChat accounts with AI models that can generate contextual replies, process voice messages, and produce images directly inside chats. The platform has evolved beyond a simple chatbot into a more autonomous agent capable of planning complex tasks, maintaining long-term memory, and invoking external tools to complete workflows. It supports multi-turn conversations with per-user context tracking, allowing more natural and persistent interactions across private and group chats. Developers can extend functionality through a plugin architecture and customizable rules, making it suitable for both personal assistants and enterprise automation scenarios.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    Cua

    Cua

    Open-source infrastructure for Computer-Use Agents. Sandboxes

    Cua is an open-source command-line utility and workflow orchestrator designed to help developers define, compose, and run common tasks with a unified interface, promoting consistency and reuse across projects. It introduces a declarative syntax for specifying build scripts, automation pipelines, environment setups, and project-specific commands so contributors don’t need to memorize disparate scripts or tooling across languages and ecosystems. Cua can also manage task dependencies, handle cross-platform invocations, and simplify complex workflows into simple aliases or compound commands that are easy to share in teams. By centralizing shared commands in a structured, documented config, it helps reduce errors, accelerates onboarding of new contributors, and keeps task definitions versioned with the codebase. The CLI is typically lightweight, easy to install, and designed to integrate with existing toolchains and shells without friction.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 12
    DeepLabCut

    DeepLabCut

    Implementation of DeepLabCut

    DeepLabCut™ is an efficient method for 2D and 3D markerless pose estimation based on transfer learning with deep neural networks that achieves excellent results (i.e. you can match human labeling accuracy) with minimal training data (typically 50-200 frames). We demonstrate the versatility of this framework by tracking various body parts in multiple species across a broad collection of behaviors. The package is open source, fast, robust, and can be used to compute 3D pose estimates or for multi-animals. Please see the original paper and the latest work below! This package is collaboratively developed by the Mathis Group & Mathis Lab at EPFL (releases prior to 2.1.9 were developed at Harvard University). The code is freely available and easy to install in a few clicks with Anaconda (and pypi). DeepLabCut is an open-source Python package for animal pose estimation.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 13
    DensePose

    DensePose

    A real-time approach for mapping all human pixels of 2D RGB images

    DensePose is a computer vision system that maps all human pixels in an RGB image to the 3D surface of a human body model. It extends human pose estimation from predicting joint keypoints to providing dense correspondences between 2D images and a canonical 3D mesh (such as the SMPL model). This enables detailed understanding of human shape, motion, and surface appearance directly from images or videos. The repository includes the DensePose network architecture, training code, pretrained models, and dataset tools for annotation and visualization. DensePose is widely used in augmented reality, motion capture, virtual try-on, and visual effects applications because it enables real-time 3D human mapping from 2D inputs. The model architecture builds on Mask R-CNN, using additional regression heads to predict UV coordinates that map image pixels to 3D surfaces.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    Diffgram

    Diffgram

    Training data (data labeling, annotation, workflow) for all data types

    From ingesting data to exploring it, annotating it, and managing workflows. Diffgram is a single application that will improve your data labeling and bring all aspects of training data under a single roof. Diffgram is world’s first truly open source training data platform that focuses on giving its users an unlimited experience. This is aimed to reduce your data labeling bills and increase your Training Data Quality. Training Data is the art of supervising machines through data. This includes the activities of annotation, which produces structured data; ready to be consumed by a machine learning model. Annotation is required because raw media is considered to be unstructured and not usable without it. That’s why training data is required for many modern machine learning use cases including computer vision, natural language processing and speech recognition.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 15
    ESPnet

    ESPnet

    End-to-end speech processing toolkit

    ESPnet is a comprehensive end-to-end speech processing toolkit covering a wide spectrum of tasks, including automatic speech recognition (ASR), text-to-speech (TTS), speech translation (ST), speech enhancement, speaker diarization, and spoken language understanding. It uses PyTorch as its deep learning engine and adopts a Kaldi-style data processing pipeline for features, data formats, and experimental recipes. This combination allows researchers to leverage modern neural architectures while still benefiting from the robust data preparation practices developed in the speech community. ESPnet provides many ready-to-run recipes for popular academic benchmarks, making it straightforward to reproduce published results or serve as baselines for new research. The toolkit also hosts numerous pretrained models and example configs, ranging from Transformer and Conformer architectures to various attention-based encoder-decoder models.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 16
    ElevenLabs Python

    ElevenLabs Python

    The official Python SDK for the ElevenLabs API

    elevenlabs-python is the official Python SDK for the ElevenLabs API, giving developers a convenient way to access ElevenLabs’ high-quality, lifelike voices. The library wraps the HTTP API into a typed Python client, so you can perform text-to-speech, streaming, voice cloning, voice management, and agents-related operations with simple method calls. It exposes ElevenLabs’ main models such as Eleven Multilingual v2, Eleven Flash v2.5, and Eleven Turbo v2.5, each targeting different trade-offs between latency, cost, and quality. The SDK is designed for quick setup: after installing the package and setting an API key, you can generate speech in multiple languages and play or process the resulting audio bytes. It includes helper utilities (like play and stream) so you can either play audio locally or integrate it into your own playback or networking pipeline.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 17
    EverMemOS

    EverMemOS

    Long-term memory OS for AI with structured recall and context awarenes

    EverMemOS is an open-source memory operating system built to give AI agents long-term, structured memory. It captures conversations, transforms them into organized memory units, and enables agents to recall past interactions with context and meaning. Instead of treating each prompt independently, it builds evolving user profiles, tracks preferences, and connects related events into coherent narratives. Its architecture combines memory storage, indexing, and retrieval with agent-level reasoning, allowing AI systems to make informed decisions based on prior interactions. EverMemOS goes beyond simple retrieval by actively applying stored knowledge to current tasks, improving personalization and consistency. EverMemOS uses a multi-stage memory lifecycle to convert raw dialogue into structured semantic data, supporting long-horizon reasoning and adaptive behavior across sessions.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 18
    EverydayWechat

    EverydayWechat

    Python tool that automates WeChat messages, replies, & group utilities

    EverydayWechat is a Python-based automation tool designed to enhance and automate interactions on the WeChat messaging platform. Built using Python 3 and the Itchat library, it connects to the web version of WeChat to perform various automated messaging tasks. It allows users to send scheduled messages to friends or group chats, including daily weather updates, reminders, inspirational quotes, and other personalized content. It also supports intelligent automatic replies to incoming messages by integrating with multiple chatbot services. In addition to personal messaging automation, the project includes a group assistant that can respond to queries and provide useful information within chat groups. These group utilities can retrieve data such as weather conditions, calendar details, garbage classification information, movie box office statistics, delivery tracking updates, and air quality reports.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 19
    EvoAgentX

    EvoAgentX

    Self-evolving AI agent framework for automated workflows

    EvoAgentX is an open source framework for building, evaluating, and continuously improving LLM-based agents and multi-agent workflows. It moves beyond static pipelines by introducing a self-evolving system where agents are automatically generated, tested, and optimised through iterative feedback. Developers can define goals in natural language, while the framework handles workflow creation, execution, and refinement. Its modular architecture supports layered components for agents, workflows, evaluation, and evolution, enabling flexible experimentation and scaling. EvoAgentX integrates optimisation algorithms to refine prompts, tool usage, and workflow structures over time. This allows agents to adapt dynamically instead of relying on fixed logic. It is designed for researchers and developers who want to automate complex agent systems and improve performance through continuous learning cycles, reducing manual orchestration and enabling more efficient development.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 20
    FAY

    FAY

    Framework for building AI-powered interactive digital humans and agent

    Fay is an open source framework designed to build and deploy interactive digital humans powered by large language models. It acts as a middleware layer that connects digital character technologies with conversational AI systems and business applications. Fay supports various types of digital humans, including 2.5D and 3D avatars, and can be integrated with applications running on mobile devices, PCs, web platforms, and embedded systems. Its architecture allows developers to combine different AI components such as speech recognition, text-to-speech, and large language models to create conversational digital agents. Fay provides multiple interfaces for text, voice, and digital human control, enabling developers to build interactive assistants, virtual presenters, or automated service agents. It also supports custom knowledge bases and configurable behaviors so developers can tailor the personality and responses of the digital human.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    Flower

    Flower

    Flower: A Friendly Federated Learning Framework

    A unified approach to federated learning, analytics, and evaluation. Federate any workload, any ML framework, and any programming language. Federated learning systems vary wildly from one use case to another. Flower allows for a wide range of different configurations depending on the needs of each individual use case. Flower originated from a research project at the University of Oxford, so it was built with AI research in mind. Many components can be extended and overridden to build new state-of-the-art systems. Different machine learning frameworks have different strengths. Flower can be used with any machine learning framework, for example, PyTorch, TensorFlow, Hugging Face Transformers, PyTorch Lightning, scikit-learn, JAX, TFLite, MONAI, fastai, MLX, XGBoost, Pandas for federated analytics, or even raw NumPy for users who enjoy computing gradients by hand.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    GLIDE (Text2Im)

    GLIDE (Text2Im)

    GLIDE: a diffusion-based text-conditional image synthesis model

    glide-text2im is an open source implementation of OpenAI’s GLIDE model, which generates photorealistic images from natural language text prompts. It demonstrates how diffusion-based generative models can be conditioned on text to produce highly detailed and coherent visual outputs. The repository provides both model code and pretrained checkpoints, making it possible for researchers and developers to experiment with text-to-image synthesis. GLIDE includes advanced techniques such as classifier-free guidance, which improves the quality and alignment of generated images with the input text. The project also offers sampling scripts and utilities for exploring how diffusion models can be applied to multimodal tasks. As one of the early diffusion-based text-to-image systems, glide-text2im laid important groundwork for later advances in generative AI research.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    GLM-4-Voice

    GLM-4-Voice

    GLM-4-Voice | End-to-End Chinese-English Conversational Model

    GLM-4-Voice is an open-source speech-enabled model from ZhipuAI, extending the GLM-4 family into the audio domain. It integrates advanced voice recognition and generation with the multimodal reasoning capabilities of GLM-4, enabling smooth natural interaction via spoken input and output. The model supports real-time speech-to-text transcription, spoken dialogue understanding, and text-to-speech synthesis, making it suitable for conversational AI, virtual assistants, and accessibility applications. GLM-4-Voice builds upon the bilingual strengths of the GLM architecture, supporting both Chinese and English, and is designed to handle long-form conversations with context retention. The repository provides model weights, inference demos, and setup instructions for deploying speech-enabled AI systems.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    GPT Neo

    GPT Neo

    An implementation of model parallel GPT-2 and GPT-3-style models

    An implementation of model & data parallel GPT3-like models using the mesh-tensorflow library. If you're just here to play with our pre-trained models, we strongly recommend you try out the HuggingFace Transformer integration. Training and inference is officially supported on TPU and should work on GPU as well. This repository will be (mostly) archived as we move focus to our GPU-specific repo, GPT-NeoX. NB, while neo can technically run a training step at 200B+ parameters, it is very inefficient at those scales. This, as well as the fact that many GPUs became available to us, among other things, prompted us to move development over to GPT-NeoX. All evaluations were done using our evaluation harness. Some results for GPT-2 and GPT-3 are inconsistent with the values reported in the respective papers. We are currently looking into why, and would greatly appreciate feedback and further testing of our eval harness.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 25
    GitDiagram

    GitDiagram

    AI tool that converts GitHub repositories into interactive diagrams

    GitDiagram is an open source web application designed to help developers quickly understand the structure and architecture of GitHub repositories by automatically generating interactive diagrams. It analyzes repository metadata such as the file tree and project documentation to build a visual representation of how different components of a project relate to one another. It uses an AI-powered pipeline to interpret repository structure and transform that information into system design diagrams rendered with Mermaid visualization. These diagrams provide a high-level overview of a codebase, making it easier for developers to explore unfamiliar projects or understand large and complex repositories. Users can interact with the generated diagrams by clicking components to navigate directly to related files or directories within the repository. GitDiagram combines a modern web frontend with a backend service that processes repository data and generates diagrams dynamically.
    Downloads: 4 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB