Open Source Python Artificial Intelligence Software - Page 52

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Perception Models

    Perception Models

    State-of-the-art Image & Video CLIP, Multimodal Large Language Models

    Perception Models is a state-of-the-art framework developed by Facebook Research for advanced image and video perception tasks. It introduces two primary components: the Perception Encoder (PE) for visual feature extraction and the Perception Language Model (PLM) for multimodal decoding and reasoning. The PE module is a family of vision encoders designed to excel in image and video understanding, surpassing models like SigLIP2, InternVideo2, and DINOv2 across multiple benchmarks. Meanwhile, PLM integrates with PE to power vision-language modeling, achieving results competitive with leading multimodal systems such as QwenVL2.5 and InternVL3, all while being fully reproducible with open data. The project supports a wide range of research applications, from visual recognition and dense prediction to fine-grained multimodal understanding. Additionally, it includes several large-scale open datasets for both image and video perception.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Poe Bot Protocol
    Poe is a platform for interacting with AI-based bots. That includes well-known chat bots like OpenAI's ChatGPT and Anthropic's Claude, but anyone can create their own bot by implementing the protocol described in this repository.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Prompt flow

    Prompt flow

    Build high-quality LLM apps

    Prompt flow is a suite of development tools designed to streamline the end-to-end development cycle of LLM-based AI applications, from ideation, prototyping, testing, and evaluation to production deployment and monitoring. It makes prompt engineering much easier and enables you to build LLM apps with production quality.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    PySC2

    PySC2

    StarCraft II learning environment

    PySC2 is DeepMind's Python component of the StarCraft II Learning Environment (SC2LE). It exposes Blizzard Entertainment's StarCraft II Machine Learning API as a Python RL Environment. This is a collaboration between DeepMind and Blizzard to develop StarCraft II into a rich environment for RL research. PySC2 provides an interface for RL agents to interact with StarCraft 2, getting observations and sending actions. The easiest way to get PySC2 is to use pip. That will install the pysc2 package along with all the required dependencies. virtualenv can help manage your dependencies. You may also need to upgrade pip: pip install --upgrade pip for the pysc2 install to work. If you're running on an older system you may need to install libsdl libraries for the pygame dependency.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    PyTorch Geometric Temporal

    PyTorch Geometric Temporal

    Spatiotemporal Signal Processing with Neural Machine Learning Models

    The library consists of various dynamic and temporal geometric deep learning, embedding, and Spatio-temporal regression methods from a variety of published research papers. Moreover, it comes with an easy-to-use dataset loader, train-test splitter and temporal snaphot iterator for dynamic and temporal graphs. The framework naturally provides GPU support. It also comes with a number of benchmark datasets from the epidemiological forecasting, sharing economy, energy production and web traffic management domains. Finally, you can also create your own datasets. The package interfaces well with Pytorch Lightning which allows training on CPUs, single and multiple GPUs out-of-the-box. PyTorch Geometric Temporal makes implementing Dynamic and Temporal Graph Neural Networks quite easy - see the accompanying tutorial. Head over to our documentation to find out more about installation, creation of datasets and a full list of implemented methods and available datasets.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    PyTorch Image Models

    PyTorch Image Models

    The largest collection of PyTorch image encoders / backbones

    timm (PyTorch Image Models) is a premier library hosting a vast collection of state-of-the-art image classification models and backbones such as ResNet, EfficientNet, NFNet, Vision Transformer, ConvNeXt, and more. Created by Ross Wightman and now maintained by Hugging Face, it includes pretrained weights, data loaders, augmentations, optimizers, schedulers, and reference scripts for training, evaluation, inference, and model export. It's an essential toolkit for vision research and production workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    PyTorch-BigGraph

    PyTorch-BigGraph

    Generate embeddings from large-scale graph-structured data

    PyTorch-BigGraph (PBG) is a system for learning embeddings on massive graphs—think billions of nodes and edges—using partitioning and distributed training to keep memory and compute tractable. It shards entities into partitions and buckets edges so that each training pass only touches a small slice of parameters, which drastically reduces peak RAM and enables horizontal scaling across machines. PBG supports multi-relation graphs (knowledge graphs) with relation-specific scoring functions, negative sampling strategies, and typed entities, making it suitable for link prediction and retrieval. Its training loop is built for throughput: asynchronous I/O, memory-mapped tensors, and lock-free updates keep GPUs and CPUs fed even at extreme scale. The toolkit includes evaluation metrics and export tools so learned embeddings can be used in downstream nearest-neighbor search, recommendation, or analytics. In practice, PBG’s design lets practitioners train high-quality graph embeddings.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    PyTorch-Tutorial-2nd

    PyTorch-Tutorial-2nd

    CV, NLP, LLM project applications, and advanced engineering deployment

    PyTorch-Tutorial-2nd is an open-source educational repository that provides structured tutorials for learning deep learning with the PyTorch framework. The project serves as a practical companion to a second edition of a PyTorch learning guide and is designed to help learners understand neural network concepts through hands-on coding examples. The repository covers a wide range of topics including tensor operations, neural network construction, model training workflows, and optimization strategies. It also introduces practical machine learning techniques such as convolutional neural networks, recurrent networks, and other architectures commonly used in modern AI applications. Each tutorial focuses on step-by-step implementation so learners can understand how theoretical concepts translate into working code. The materials are designed for both beginners and intermediate developers who want to gain practical experience building deep learning models using PyTorch.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    PyTorch3D

    PyTorch3D

    PyTorch3D is FAIR's library of reusable components for deep learning

    PyTorch3D is a comprehensive library for 3D deep learning that brings differentiable rendering, geometric operations, and 3D data structures into the PyTorch ecosystem. It’s designed to make it easy to build and train neural networks that work directly with 3D data such as meshes, point clouds, and implicit surfaces. The library provides fast GPU-accelerated implementations of rendering pipelines, transformations, rasterization, and lighting—making it possible to compute gradients through full 3D rendering processes. Researchers use it for tasks like shape generation, reconstruction, view synthesis, and visual reasoning. PyTorch3D also includes utilities for loading, transforming, and sampling 3D assets, so models can be trained end-to-end from 2D supervision or partial data. Its modular design allows easy extension—components like differentiable rasterizers, mesh blending, or signed distance field (SDF) modules can be swapped or combined to test new architectures quickly.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    RAGxplorer

    RAGxplorer

    Open-source tool to visualise your RAG

    RAGxplorer is an open-source visualization tool designed to help developers analyze and understand Retrieval-Augmented Generation (RAG) pipelines. Retrieval-augmented generation combines language models with external document retrieval systems in order to produce more accurate and grounded responses. However, RAG systems can be complex because they involve multiple components such as embedding models, vector databases, and retrieval algorithms. RAGxplorer provides visual tools that allow developers to inspect how documents are embedded, retrieved, and used to answer queries. The software can load documents, generate embeddings, and project them into reduced vector spaces so that users can visually explore relationships between queries and retrieved documents. It also includes interactive interfaces that show how retrieval affects the final output of the language model.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    RecBole

    RecBole

    A unified, comprehensive and efficient recommendation library

    A unified, comprehensive and efficient recommendation library. We design general and extensible data structures to unify the formatting and usage of various recommendation datasets. We implement more than 100 commonly used recommendation algorithms and provide formatted copies of 28 recommendation datasets. We support a series of widely adopted evaluation protocols or settings for testing and comparing recommendation algorithms. RecBole is developed based on Python and PyTorch for reproducing and developing recommendation algorithms in a unified, comprehensive and efficient framework for research purpose. It can be installed from pip, conda and source, and is easy to use. We have implemented more than 100 recommender system models, covering four common recommender system categories in RecBole and eight toolkits of RecBole2.0, including General Recommendation, Sequential Recommendation, Context-aware Recommendation, and Knowledge-based Recommendation and sub-packages.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Recurrent Interface Network (RIN)

    Recurrent Interface Network (RIN)

    Implementation of Recurrent Interface Network (RIN)

    Implementation of Recurrent Interface Network (RIN), for highly efficient generation of images and video without cascading networks, in Pytorch. The author unawaredly reinvented the induced set-attention block from the set transformers paper. They also combine this with the self-conditioning technique from the Bit Diffusion paper, specifically for the latents. The last ingredient seems to be a new noise function based around the sigmoid, which the author claims is better than cosine scheduler for larger images. The big surprise is that the generations can reach this level of fidelity. Will need to verify this on my own machine. Additionally, we will try adding an extra linear attention on the main branch as well as self-conditioning in the pixel space. The insight of being able to self-condition on any hidden state of the network as well as the newly proposed sigmoid noise schedule are the two main findings.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Rhino

    Rhino

    On-device Speech-to-Intent engine powered by deep learning

    Rhino is Picovoice's Speech-to-Intent engine. It directly infers intent from spoken commands within a given context of interest, in real-time. The end-to-end platform for embedding private voice AI into any software in a few lines of code. Design with no limits on top of a modular platform. Create use-case-specific voice AI models in seconds. Develop voice features with a few lines of code using intuitive and cross-platform SDKs. Deliver voice AI everywhere: on-device, mobile, web browsers, on-premise, or cloud. Measure adoption, learn, and iterate. Continuously re-design and re-train to optimize engagement. Building accurate, responsive, and private voice technology is difficult. We learned the hard way, so you don’t have to. Picovoice heavily invests in R&D to offer superior voice AI that surpasses even Big Tech in accuracy and efficiency. Picovoice researchers do not follow recent frameworks and techniques but build them.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    SMAC

    SMAC

    SMAC: The StarCraft Multi-Agent Challenge

    SMAC (StarCraft II Multi-Agent Challenge) is a benchmark environment for cooperative multi-agent reinforcement learning (MARL), based on real-time strategy (RTS) game scenarios in StarCraft II. It allows researchers to test algorithms where multiple units (agents) must collaborate to win battles against built-in game AI opponents. SMAC provides a controlled testbed for studying decentralized execution and centralized training paradigms in MARL.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    SalesGPT

    SalesGPT

    Context-aware AI Sales Agent to automate sales outreach

    This repo is an implementation of a context-aware AI Agent for Sales using LLMs and can work across voice, email and texting (SMS, WhatsApp, WeChat, Weibo, Telegram, etc.). SalesGPT is context-aware, which means it can understand what stage of a sales conversation it is in and act accordingly. Moreover, SalesGPT has access to tools, such as your own pre-defined product knowledge base, significantly reducing hallucinations.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Sandstorm

    Sandstorm

    One API call, pull Claude agent, completely sandboxed

    Sandstorm is an open-source project that wraps a powerful Claude-based AI agent within a completely sandboxed, ephemeral API service designed to make agentic AI workflows easy to deploy and scale without infrastructure complexity. The core idea is to provide “one API call” access to a robust Claude agent loop that runs inside a secure sandbox, so you can upload files, connect tools, and run long-running tasks — all managed behind a simple REST-style interface that disappears when the work is done. This approach lowers the friction of building autonomous agents by removing the need to provision servers, orchestrate distributed agents, or manage persistent tooling; agents can be spun up in parallel without manual setup and shut down when complete. The sandbox environment isolates agent execution for security and predictability, and project updates continue to harden observability, fault handling, and configuration validation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Seldon Core

    Seldon Core

    An MLOps framework to package, deploy, monitor and manage models

    The de facto standard open-source platform for rapidly deploying machine learning models on Kubernetes. Seldon Core, our open-source framework, makes it easier and faster to deploy your machine learning models and experiments at scale on Kubernetes. Seldon Core serves models built in any open-source or commercial model building framework. You can make use of powerful Kubernetes features like custom resource definitions to manage model graphs. And then connect your continuous integration and deployment (CI/CD) tools to scale and update your deployment. Built on Kubernetes, runs on any cloud and on-premises. Framework agnostic, supports top ML libraries, toolkits and languages. Advanced deployments with experiments, ensembles and transformers. Our open-source framework makes it easier and faster to deploy your machine learning models and experiments at scale on Kubernetes. The Kubeflow project is dedicated to making deployments of machine learning (ML) workflows on Kubernetes.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Self-hosted AI Package

    Self-hosted AI Package

    Run all your local AI together in one package

    Self-hosted AI Package is an open-source Docker Compose-based starter kit that makes it easy to bootstrap a full local AI and low-code development environment with commonly used open tools, empowering developers to run LLMs and AI workflows entirely on their infrastructure. The stack typically includes Ollama for running local large language models, n8n as a low-code workflow automation platform, Supabase for database and vector storage, Open WebUI for interacting with models, Flowise for agent building, and additional services like SearXNG, Neo4j, and Langfuse for search, knowledge graphs, and observability. This integrated setup allows users to experiment with RAG pipelines, automated workflows, AI agents, and project data management without relying on external hosted services, increasing flexibility and privacy. The repository comes with example workflows (such as Local RAG AI Agent workflows) and environment configurations that help streamline setup and encourage customization.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    Shap-E

    Shap-E

    Generate 3D objects conditioned on text or images

    The shap-e repository provides the official code and model release for Shap-E, a conditional generative model designed to produce 3D assets (implicit functions, meshes, neural radiance fields) from text or image prompts. The model is built with a two-stage architecture: first an encoder that maps existing 3D assets into parameterizations of implicit functions, and then a conditional diffusion model trained on those parameterizations to generate new assets. Because it works at the level of implicit functions, Shap-E can render output both as textured meshes and NeRF-style volumetric renderings. The repository contains sample notebooks (e.g. sample_text_to_3d.ipynb, sample_image_to_3d.ipynb) so users can try out text → 3D or image → 3D generation. The code is distributed under the MIT license, and includes a “model card” that documents limitations, recommended use, and ethical considerations.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    ShoppingAgent

    ShoppingAgent

    Custom Chinese chatbot with Seq2Seq, GPT, and agent features

    ShoppingAgent is an open source Chinese conversational AI system that allows users to build and train their own chatbot using custom datasets. It provides multiple implementations of chatbot architectures, including traditional Seq2Seq models as well as newer GPT-style approaches, reflecting the evolution of conversational AI techniques. ShoppingAgent is structured to support experimentation across different deep learning frameworks such as TensorFlow, PyTorch, and MindSpore, giving developers flexibility in how they train and deploy models. In addition to core chatbot functionality, the project introduces agent-based capabilities, enabling practical use cases like automated workflows and task-oriented assistants. It also includes support for small language models and local training scripts, making it accessible for users with limited computational resources. ShoppingAgent can be applied to scenarios such as customer service, question answering, and casual conversation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Siamese and triplet learning

    Siamese and triplet learning

    Siamese and triplet networks with online triplet mining in PyTorch

    Siamese and triplet learning is a PyTorch implementation of Siamese and triplet neural network architectures designed for learning embedding representations in machine learning tasks. These types of networks learn to map images into a compact feature space where the distance between vectors reflects the similarity between inputs. Such embeddings are commonly used in applications like face recognition, image similarity search, and few-shot learning. The repository demonstrates how to train these models using contrastive loss and triplet loss functions, which encourage embeddings of similar samples to be close while pushing dissimilar samples farther apart. It includes data loaders, training scripts, neural network architectures, and evaluation metrics that allow researchers to experiment with different embedding learning strategies. The project also implements online pair and triplet mining techniques to efficiently generate training examples during model training.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Simple StyleGan2 for Pytorch

    Simple StyleGan2 for Pytorch

    Simplest working implementation of Stylegan2

    Simple Pytorch implementation of Stylegan2 that can be completely trained from the command-line, no coding needed. You will need a machine with a GPU and CUDA installed. You can also specify the location where intermediate results and model checkpoints should be stored. You can increase the network capacity (which defaults to 16) to improve generation results, at the cost of more memory. By default, if the training gets cut off, it will automatically resume from the last checkpointed file. Once you have finished training, you can generate images from your latest checkpoint. If a previous checkpoint contained a better generator, (which often happens as generators start degrading towards the end of training), you can load from a previous checkpoint with another flag. A technique used in both StyleGAN and BigGAN is truncating the latent values so that their values fall close to the mean. The small the truncation value, the better the samples will appear at the cost of sample variety.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    SimpleLLM

    SimpleLLM

    950 line, minimal, extensible LLM inference engine built from scratch

    SimpleLLM is a minimal, extensible large language model inference engine implemented in roughly 950 lines of code, built from scratch to serve both as a learning tool and a research platform for novel inference techniques. It provides the core components of an LLM runtime—such as tokenization, batching, and asynchronous execution—without the abstraction overhead of more complex engines, making it easier for developers and researchers to understand and modify. Designed to run efficiently on high-end GPUs like NVIDIA H100 with support for models such as OpenAI/gpt-oss-120b, Simple-LLM implements continuous batching and event-driven inference loops to maximize hardware utilization and throughput. Its straightforward code structure allows anyone experimenting with custom kernels, new batching strategies, or inference optimizations to trace execution from input to output with minimal cognitive overhead.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Spark TTS

    Spark TTS

    Spark-TTS Inference Code

    Spark TTS is an open-source, PyTorch-based text-to-speech inference system that leverages large language models to produce highly natural, intelligible speech from text input. It uses an efficient single-stream architecture where speech tokens are directly reconstructed from the predictions of an LLM, removing the need for external acoustic models or complex vocoders and making the generation pipeline cleaner and faster. The project supports zero-shot voice cloning, meaning it can imitate a new speaker’s voice without dedicated training for that specific voice, and works across languages, including English and Chinese, even in cross-lingual code-switching scenarios. Spark-TTS allows users to control speech characteristics like gender, pitch, and speaking rate to customize synthesized output and support virtual speaker creation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    SparrowRecSys

    SparrowRecSys

    A Deep Learning Recommender System

    SparrowRecSys is an open-source deep learning recommendation system framework designed to demonstrate the architecture and implementation of modern industrial-scale recommender systems. The project integrates multiple machine learning models and data processing pipelines to simulate how real-world recommendation platforms operate. It includes components for offline data processing, feature engineering, model training, real-time data updates, and online recommendation services. SparrowRecSys supports a wide range of state-of-the-art recommendation algorithms, including models for click-through rate prediction and user behavior modeling that are widely used in advertising and content recommendation systems. The system is designed as a modular platform combining technologies such as Spark, TensorFlow, and web server components to represent the full lifecycle of recommendation pipelines.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB