Open Source Python Artificial Intelligence Software - Page 11

Python Artificial Intelligence Software

View 13534 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    DeepDanbooru

    DeepDanbooru

    AI based multi-label girl image classification system

    DeepDanbooru is a deep learning system designed to automatically tag anime-style images using neural networks trained on datasets derived from the Danbooru imageboard. The project focuses on multi-label image classification, where a model predicts multiple descriptive tags that represent visual elements in an image. These tags may include characters, styles, clothing, emotions, or other attributes associated with anime artwork. The system uses convolutional neural networks trained on large datasets of tagged images to learn relationships between visual features and textual labels. Because the Danbooru dataset contains millions of images with extensive annotations, it provides a valuable training resource for machine learning models specializing in illustration analysis. Such datasets have been widely used for tasks including automatic image tagging, anime face detection, and generative modeling research.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    DeepSeek-OCR

    DeepSeek-OCR

    Contexts Optical Compression

    DeepSeek-OCR is an open-source optical character recognition solution built as part of the broader DeepSeek AI vision-language ecosystem. It is designed to extract text from images, PDFs, and scanned documents, and integrates with multimodal capabilities that understand layout, context, and visual elements beyond raw character recognition. The system treats OCR not simply as “read the text” but as “understand what the text is doing in the image”—for example distinguishing captions from body text, interpreting tables, or recognizing handwritten versus printed words. It supports local deployment, enabling organizations concerned about privacy or latency to run the pipeline on-premises rather than send sensitive documents to third-party cloud services. The codebase is written in Python with a focus on modularity: you can swap preprocessing, recognition, and post-processing components as needed for custom workflows.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 3
    DocsGPT

    DocsGPT

    Private AI platform for agents, enterprise search and RAG pipelines

    DocsGPT is an open-source AI platform for deploying private RAG pipelines, AI agents, and enterprise search on your own infrastructure. Connect any data source (PDFs, DOCX, CSV, Excel, HTML, audio, GitHub, databases, URLs) and get accurate, hallucination-free answers with source citations. Choose your LLM: OpenAI, Anthropic, Google Gemini, or local models. Works with Qdrant, MongoDB, and Elasticsearch and more. Deploy via Docker or Kubernetes with full data sovereignty. Build embeddable chat and search widgets, automate multi-step workflows with AI agents, and integrate via Slack, Telegram, Discord, or REST API. Enterprise features include RBAC, 99.9% uptime SLA, and dedicated support. MIT licensed.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 4
    Free-Auto-GPT

    Free-Auto-GPT

    Free AutoGPT enables autonomous AI tasks without paid APIs

    Free Auto GPT is an open source project that delivers a simplified version of an autonomous AI agent capable of completing tasks independently. It allows users to run an AutoGPT-style system without relying on paid OpenAI APIs, making it more accessible for experimentation and personal use. Free Auto GPT can take a goal, break it into smaller steps, and execute actions in a loop to achieve results with minimal human input. Designed for ease of use, the project focuses on removing cost barriers while still demonstrating how autonomous agents function. It is suitable for developers, learners, and hobbyists who want to explore AI-driven automation without subscription requirements. It provides a lightweight implementation that highlights core AutoGPT concepts such as task decomposition, iteration, and independent execution in a simplified environment.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    GLM-4

    GLM-4

    GLM-4 series: Open Multilingual Multimodal Chat LMs

    GLM-4 is a family of open models from ZhipuAI that spans base, chat, and reasoning variants at both 32B and 9B scales, with long-context support and practical local-deployment options. The GLM-4-32B-0414 models are trained on ~15T high-quality data (including substantial synthetic reasoning data), then post-trained with preference alignment, rejection sampling, and reinforcement learning to improve instruction following, coding, function calling, and agent-style behaviors. The GLM-Z1-32B-0414 line adds deeper mathematical, coding, and logical reasoning via extended reinforcement learning and pairwise ranking feedback, while GLM-Z1-Rumination-32B-0414 introduces a “rumination” mode that performs longer, tool-using deep research for complex, open-ended tasks. A lightweight GLM-Z1-9B-0414 brings many of these techniques to a smaller model, targeting strong reasoning under tight resource budgets.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    Image Super-Resolution (ISR)

    Image Super-Resolution (ISR)

    Super-scale your images and run experiments with Residual Dense

    The goal of this project is to upscale and improve the quality of low-resolution images. This project contains Keras implementations of different Residual Dense Networks for Single Image Super-Resolution (ISR) as well as scripts to train these networks using content and adversarial loss components. Docker scripts and Google Colab notebooks are available to carry training and prediction. Also, we provide scripts to facilitate training on the cloud with AWS and Nvidia-docker with only a few commands. When training your own model, start with only PSNR loss (50+ epochs, depending on the dataset) and only then introduce GANS and feature loss. This can be controlled by the loss weights argument. The weights used to produce these images are available directly when creating the model object. ISR is compatible with Python 3.6 and is distributed under the Apache 2.0 license.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 7
    KoboldAI

    KoboldAI

    Your gateway to GPT writing

    This is a browser-based front-end for AI-assisted writing with multiple local & remote AI models. It offers the standard array of tools, including Memory, Author's Note, World Info, Save & Load, adjustable AI settings, formatting options, and the ability to import existing AI Dungeon adventures. You can also turn on Adventure mode and play the game like AI Dungeon Unleashed. Stories can be played like a Novel, a text adventure game or used as a chatbot with an easy toggles to change between the multiple gameplay styles. This makes KoboldAI both a writing assistant, a game and a platform for so much more. The way you play and how good the AI will be depends on the model or service you decide to use. No matter if you want to use the free, fast power of Google Colab, your own high end graphics card, an online service you have an API key for (Like OpenAI or Inferkit) or if you rather just run it slower on your CPU you will be able to find a way to use KoboldAI that works for you.
    Leader badge
    Downloads: 160 This Week
    Last Update:
    See Project
  • 8
    LLaMA 3

    LLaMA 3

    The official Meta Llama 3 GitHub site

    This repository is the former home for Llama 3 model artifacts and getting-started code, covering pre-trained and instruction-tuned variants across multiple parameter sizes. It introduced the public packaging of weights, licenses, and quickstart examples that helped developers fine-tune or run the models locally and on common serving stacks. As the Llama stack evolved, Meta consolidated repositories and marked this one deprecated, pointing users to newer, centralized hubs for models, utilities, and docs. Even as a deprecated repo, it documents the transition path and preserves references that clarify how Llama 3 releases map into the current ecosystem. Practically, it functioned as a bridge between Llama 2 and later Llama releases by standardizing distribution and starter code for inference and fine-tuning. Teams still treat it as historical reference material for version lineage and migration notes.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    MCP for Unity

    MCP for Unity

    AI bridge enabling assistants to control and automate Unity Editor

    Unity MCP is an open source integration that connects AI assistants with the Unity Editor through the Model Context Protocol (MCP). It acts as a bridge that allows language models and AI coding tools to interact directly with a Unity development environment using structured commands and tools. By linking an AI assistant to a running Unity project, the system enables automated operations such as managing project assets, modifying scenes, editing scripts, and performing other development tasks inside the editor. It exposes Unity functionality as callable tools so that AI systems can understand and manipulate game development workflows programmatically. This approach allows developers to control Unity using natural language prompts and automated workflows rather than manual editor interaction. Unity MCP supports various AI assistants and development tools that implement MCP clients, enabling flexible integration with existing AI development environments.
    Downloads: 6 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    MLC LLM

    MLC LLM

    Universal LLM Deployment Engine with ML Compilation

    MLC LLM is a machine learning compiler and deployment framework designed to enable efficient execution of large language models across a wide range of hardware platforms. The project focuses on compiling models into optimized runtimes that can run natively on devices such as GPUs, mobile processors, browsers, and edge hardware. By leveraging machine learning compilation techniques, mlc-llm produces high-performance inference engines that maintain consistent APIs across platforms. The system supports deployment on environments including Linux, macOS, Windows, iOS, Android, and web browsers while utilizing different acceleration technologies such as CUDA, Vulkan, Metal, and WebGPU. It also provides OpenAI-compatible APIs that allow developers to integrate locally deployed models into existing AI applications without major code changes.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 11
    MemPalace

    MemPalace

    The highest-scoring AI memory system ever benchmarked

    MemPalace is an open-source AI memory system designed to solve one of the most persistent limitations of large language models: the loss of context between sessions. Instead of relying on summarization or selective extraction like most memory tools, it takes a radically different approach by storing conversations in their entirety and making them retrievable through structured organization and semantic search. The system is inspired by the classical “memory palace” mnemonic technique, organizing information into hierarchical spaces such as wings, rooms, and halls, which allows AI agents to navigate past knowledge in a more contextual and intuitive way. It operates fully locally using tools like ChromaDB, meaning it requires no API keys, cloud services, or external dependencies once installed. MemPalace emphasizes fidelity over compression, preserving full conversational history to maintain reasoning, nuance, and decision-making context that is typically lost in other systems.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 12
    Microsandbox

    Microsandbox

    Secure local-first microVM sandbox for running untrusted code fast

    Microsandbox is an open source platform designed to securely execute untrusted code in isolated environments using lightweight virtualization techniques. It focuses on combining strong security guarantees with fast startup times by leveraging hardware-level microVM isolation instead of relying solely on traditional containers or full virtual machines. It aims to solve the common tradeoffs between speed, isolation, and control that developers encounter when running untrusted workloads. It provides a local-first and self-hosted approach, allowing users to maintain full ownership of their execution environment without depending on external cloud services. Microsandbox is particularly geared toward AI agent workflows, offering integrations that enable automated systems to safely run generated code and commands. It also supports standard container images, making it compatible with existing development ecosystems and tooling.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 13
    Nexa SDK

    Nexa SDK

    Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML

    Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), and speech-to-text (ASR), and text-to-speech (TTS) capabilities. Additionally, it offers an OpenAI-compatible API server with JSON schema mode for function calling and streaming support, and a user-friendly Streamlit UI. Users can run Nexa SDK in any device with Python environment, and GPU acceleration is supported, including CUDA, Metal, and ROCm. An executable version is also available.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 14
    Open Gauss

    Open Gauss

    Project-scoped Lean workflow orchestrator from Math, Inc.

    Open Gauss is an enterprise-grade open-source relational database management system designed to handle large-scale data processing with high performance, reliability, and security. It is based on the PostgreSQL ecosystem but significantly extends its capabilities through architectural optimizations, AI-driven features, and enterprise-level enhancements. The database organizes data using the relational model, storing structured information in tables composed of rows and columns while supporting standard SQL for querying and management. One of its defining strengths is its optimization for multi-core and distributed environments, allowing it to efficiently process high volumes of concurrent transactions with minimal latency. OpenGauss also incorporates AI-based optimization techniques, such as intelligent query planning, performance prediction, and automated tuning, which help reduce operational complexity and improve efficiency.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 15
    OpenAI Harmony

    OpenAI Harmony

    Renderer for the harmony response format to be used with gpt-oss

    Harmony is a response format developed by OpenAI for use with the gpt-oss model series. It defines a structured way for language models to produce outputs, including regular text, reasoning traces, tool calls, and structured data. By mimicking the OpenAI Responses API, Harmony provides developers with a familiar interface while enabling more advanced capabilities such as multiple output channels, instruction hierarchies, and tool namespaces. The format is essential for ensuring gpt-oss models operate correctly, as they are trained to rely on this structure for generating and organizing their responses. For users accessing gpt-oss through third-party providers like HuggingFace, Ollama, or vLLM, Harmony formatting is handled automatically, but developers building custom inference setups must implement it directly. With its flexible design, Harmony serves as the foundation for creating more interpretable, controlled, and extensible interactions with open-weight language models.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    Piano transcription

    Piano transcription

    Task of transcribing piano recordings into MIDI files

    Piano transcription is an open-source high-resolution piano transcription system by ByteDance that converts raw audio recordings of piano performance into symbolic MIDI files — detecting note onsets, offsets, pitch, velocity, and even pedal usage. The system is implemented in Python (PyTorch) and is capable of accurate transcription of polyphonic piano recordings, even with complex passages and pedal techniques, making it suitable for classical piano music. By using this transcription tool, users can transform live performance audio (or recordings) into editable, machine-readable MIDI — enabling tasks such as analysis, editing, remixing, or generation of piano music. The authors used this system to build a large-scale classical piano MIDI dataset (see next project), but as a standalone tool it enables researchers, musicians, or hobbyists to transcribe their own piano recordings automatically.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 17
    RasaGPT

    RasaGPT

    Headless Rasa chatbot platform with LLM integration and APIs

    RasaGPT is a headless chatbot platform that combines Rasa with modern LLM tooling such as Langchain and LlamaIndex. It serves as a reference implementation and boilerplate for building conversational AI systems with retrieval and context injection. RasaGPT includes a FastAPI backend for creating custom bot endpoints, along with document ingestion and a training pipeline. It simplifies integration challenges between Rasa and LLM libraries, including metadata handling and library conflicts. RasaGPT supports multi-tenant deployments, session management, and custom schemas using pgvector. It also enables Telegram bot integration and remote access via ngrok. Docker support allows easier setup and deployment, particularly on macOS environments. While designed as a working prototype, it provides a practical foundation for developers building LLM-powered chatbot applications with extensible architecture and preconfigured components.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 18
    SAM 3D Objects

    SAM 3D Objects

    Models for object and human mesh reconstruction

    SAM 3D Objects is a foundation model that reconstructs full 3D geometry, texture, and spatial layout of objects and scenes from a single image. Given one RGB image and object masks (for example, from the Segment Anything family), it can generate a textured 3D mesh for each object, including pose and approximate scene layout. The model is specifically designed to be robust in real-world images with clutter, occlusions, small objects, and unusual viewpoints, where many earlier 3D-from-image systems struggle. It supports both single-object and multi-object generation, allowing you to reconstruct entire scenes rather than just isolated items. The repository provides code to run inference, a quickstart demo.py script, and environment setup instructions that connect to hosted checkpoints and configuration files. Outputs are aimed at downstream usability: the reconstructed assets are textured meshes suitable for further editing, rendering, or integration into 3D pipelines and engines.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 19
    Speech-AI-Forge

    Speech-AI-Forge

    Speech-AI-Forge is a project developed around TTS generation model

    Speech-AI-Forge is a full-stack project built around modern text-to-speech generation models, providing both an API server and a Gradio-based web UI for interactive use. At its core, it acts as a hub that wires together multiple speech-related capabilities, including TTS, speech-to-text and LLM-based control flows, behind a consistent interface. The system is designed to be deployed in several ways: you can try it online via hosted demos, spin it up in a one-click Colab environment, run it in Docker containers, or set it up locally with its environment preparation scripts. It is model-agnostic and advertises support for a variety of TTS and speech models such as ChatTTS, CosyVoice, Fish-Speech, FireredTTS and others, as well as Whisper-based ASR, giving you a flexible playground for experimenting with different speech stacks. The project also integrates with general-purpose LLMs (for example GPT- or LLaMA-style models), which can be used to pre-process text, manage conversations.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 20
    TabPFN

    TabPFN

    Foundation Model for Tabular Data

    TabPFN is an open-source machine learning system that introduces a foundation model designed specifically for tabular data analysis. The model is based on transformer architectures and implements a prior-data fitted network that can perform supervised learning tasks such as classification and regression with minimal configuration. Unlike many traditional machine learning workflows that require extensive hyperparameter tuning and training cycles, TabPFN is pre-trained to perform inference directly on tabular datasets. This allows it to generate predictions extremely quickly, often within seconds, while maintaining competitive accuracy on small and medium-sized datasets. The system supports a variety of tabular machine learning tasks and is designed to handle structured datasets commonly found in spreadsheets, databases, and business analytics systems.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 21
    The SpeechBrain Toolkit

    The SpeechBrain Toolkit

    A PyTorch-based Speech Toolkit

    SpeechBrain is an open-source and all-in-one conversational AI toolkit. It is designed to be simple, extremely flexible, and user-friendly. Competitive or state-of-the-art performance is obtained in various domains. SpeechBrain supports state-of-the-art methods for end-to-end speech recognition, including models based on CTC, CTC+attention, transducers, transformers, and neural language models relying on recurrent neural networks and transformers. Speaker recognition is already deployed in a wide variety of realistic applications. SpeechBrain provides different models for speaker recognition, including X-vector, ECAPA-TDNN, PLDA, and contrastive learning. Spectral masking, spectral mapping, and time-domain enhancement are different methods already available within SpeechBrain. Separation methods such as Conv-TasNet, DualPath RNN, and SepFormer are implemented as well. SpeechBrain provides efficient and GPU-friendly speech augmentation pipelines and acoustic features extraction.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 22
    Thinc

    Thinc

    A refreshing functional take on deep learning

    Thinc is a lightweight deep learning library that offers an elegant, type-checked, functional-programming API for composing models, with support for layers defined in other frameworks such as PyTorch, TensorFlow and MXNet. You can use Thinc as an interface layer, a standalone toolkit or a flexible way to develop new models. Previous versions of Thinc have been running quietly in production in thousands of companies, via both spaCy and Prodigy. We wrote the new version to let users compose, configure and deploy custom models built with their favorite framework. Switch between PyTorch, TensorFlow and MXNet models without changing your application, or even create mutant hybrids using zero-copy array interchange. Develop faster and catch bugs sooner with sophisticated type checking. Trying to pass a 1-dimensional array into a model that expects two dimensions? That’s a type error. Your editor can pick it up as the code leaves your fingers.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 23
    deepfakes_faceswap

    deepfakes_faceswap

    Deepfakes Software For All

    Faceswap is the leading free and open source multi-platform deepfakes software. When faceswapping was first developed and published, the technology was groundbreaking, it was a huge step in AI development. It was also completely ignored outside of academia because the code was confusing and fragmentary. It required a thorough understanding of complicated AI techniques and took a lot of effort to figure it out. Until one individual brought it together into a single, cohesive collection.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 24
    higgsfield

    higgsfield

    Fault-tolerant, highly scalable GPU orchestration

    Higgsfield is an open-source, fault-tolerant, highly scalable GPU orchestration, and a machine learning framework designed for training models with billions to trillions of parameters, such as Large Language Models (LLMs).
    Downloads: 6 This Week
    Last Update:
    See Project
  • 25
    spaCy models

    spaCy models

    Models for the spaCy Natural Language Processing (NLP) library

    spaCy is designed to help you do real work, to build real products, or gather real insights. The library respects your time, and tries to avoid wasting it. It's easy to install, and its API is simple and productive. spaCy excels at large-scale information extraction tasks. It's written from the ground up in carefully memory-managed Cython. If your application needs to process entire web dumps, spaCy is the library you want to be using. Since its release in 2015, spaCy has become an industry standard with a huge ecosystem. Choose from a variety of plugins, integrate with your machine learning stack and build custom components and workflows.
    Downloads: 6 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB