Open Source Python Artificial Intelligence Software - Page 32

Python Artificial Intelligence Software

View 13550 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    Riffusion

    Riffusion

    Real-time music generation using stable diffusion techniques AI

    Riffusion (hobby) is a Python-based open source library designed for real-time music and audio generation using stable diffusion techniques. Riffusion (hobby) works by generating and manipulating spectrogram images, which are then converted into playable audio clips, effectively bridging image-based diffusion models with sound synthesis. It implements a diffusion pipeline that supports prompt interpolation, allowing smooth transitions between different musical styles or prompts over time. Riffusion (hobby) serves as the core implementation for audio and image processing, providing essential building blocks for generating music from text prompts. It includes both developer-oriented tools and user-facing components such as a command-line interface and an interactive Streamlit application for experimentation. Additionally, it can run as a Flask server to expose model inference through an API, enabling integration with other applications or services.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    SAM 2

    SAM 2

    The repository provides code for running inference with SAM 2

    SAM2 is a next-generation version of the Segment Anything Model (SAM), designed to improve performance, generalization, and efficiency in promptable image segmentation tasks. It retains the core promptable interface—accepting points, boxes, or masks—but incorporates architectural and training enhancements to produce higher-fidelity masks, better boundary adherence, and robustness to complex scenes. The updated model is optimized for faster inference and lower memory use, enabling real-time interactivity even on larger images or constrained hardware. SAM2 comes with pretrained weights and easy-to-use APIs, enabling developers and researchers to integrate promptable segmentation into annotation tools, vision pipelines, or downstream tasks. The project also includes scripts and notebooks to compare SAM2 against SAM on edge cases, benchmarks showing improvements, and evaluation suites to measure mask quality metrics like IoU and boundary error.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    SDGym

    SDGym

    Benchmarking synthetic data generation methods

    The Synthetic Data Gym (SDGym) is a benchmarking framework for modeling and generating synthetic data. Measure performance and memory usage across different synthetic data modeling techniques – classical statistics, deep learning and more! The SDGym library integrates with the Synthetic Data Vault ecosystem. You can use any of its synthesizers, datasets or metrics for benchmarking. You also customize the process to include your own work. Select any of the publicly available datasets from the SDV project, or input your own data. Choose from any of the SDV synthesizers and baselines. Or write your own custom machine learning model. In addition to performance and memory usage, you can also measure synthetic data quality and privacy through a variety of metrics. Install SDGym using pip or conda. We recommend using a virtual environment to avoid conflicts with other software on your device.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    SWE-agent

    SWE-agent

    SWE-agent takes a GitHub issue and tries to automatically fix it

    SWE-agent turns LMs (e.g. GPT-4) into software engineering agents that can resolve issues in real GitHub repositories. On the SWE-bench, the SWE-agent resolves 12.47% of issues, achieving state-of-the-art performance on the full test set. We accomplish our results by designing simple LM-centric commands and feedback formats to make it easier for the LM to browse the repository, and view, edit, and execute code files. We call this an Agent-Computer Interface (ACI).
    Downloads: 3 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    SWIFT LLM

    SWIFT LLM

    Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs

    SWIFT LLM is a comprehensive framework developed within the ModelScope ecosystem for training, fine-tuning, evaluating, and deploying large language models and multimodal models. The platform provides a full machine learning pipeline that supports tasks ranging from model pre-training to reinforcement learning alignment techniques. It integrates with popular inference engines such as vLLM and LMDeploy to accelerate deployment and runtime performance. The framework also includes support for many modern training strategies, including preference learning methods and parameter-efficient fine-tuning techniques. ms-swift is designed to work with hundreds of language and multimodal models, providing a unified environment for experimentation and production deployment.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    SeaGOAT

    SeaGOAT

    local-first semantic code search engine

    SeaGOAT is an open-source semantic code search engine designed to help developers explore and understand large codebases more efficiently. Instead of relying solely on traditional keyword search, it uses vector embeddings to represent the meaning of code and queries, allowing users to perform semantic searches that find relevant code even when the exact keywords are not present. The tool runs locally on a developer’s machine and processes repositories using a combination of embedding models and conventional search utilities, enabling both semantic and text-based retrieval methods. By combining vector search with tools like ripgrep, SeaGOAT provides a hybrid approach that supports both natural language queries and precise keyword matching in source files. It is built primarily in Python and is intended to work on common operating systems such as Linux, macOS, and Windows.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Search with Lepton

    Search with Lepton

    Lightweight demo to build a conversational AI search engine quickly

    Search with Lepton is an open source demonstration project that shows how to build a conversational search engine using the Lepton AI framework. It combines traditional web search with large language models to provide natural language answers to user queries. It retrieves information from supported search engines and uses that context to generate responses through a retrieval-augmented generation approach. The implementation is intentionally minimal, containing fewer than 500 lines of code while still providing a complete working example of an AI-powered search system. It includes both a backend service written in Python and a web interface that allows users to interact with the search engine in a conversational format. Developers can configure different search providers and language models through environment variables, making it flexible for experimentation and prototyping.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    SeedVR

    SeedVR

    Repo for SeedVR2 & SeedVR

    SeedVR (from the ByteDance-Seed organization) is an open-source research and implementation repository focused on cutting-edge video restoration using diffusion transformer architectures. The project includes both the original SeedVR and its successor SeedVR2 models, which are designed to restore degraded or low-quality video content by learning to reconstruct high-fidelity frames with temporal coherence. These models leverage advanced techniques such as adaptive attention mechanisms and adversarial training to produce visually appealing results in a single inference step, pushing the boundaries of video restoration research. SeedVR’s transformer-based design allows it to handle variable frame resolutions and lengths, and its architecture is optimized to overcome traditional limitations of windowed attention in high-resolution contexts.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    Serena

    Serena

    Agent toolkit providing semantic retrieval and editing capabilities

    Serena is a coding-focused agent toolkit that turns an LLM into a practical software-engineering agent with semantic retrieval and editing over real repositories. It operates as an MCP server (and other integrations), exposing IDE-like tools so agents can locate symbols, reason about code structure, make targeted edits, and validate changes. The toolkit is LLM-agnostic and framework-agnostic, positioning itself as a drop-in capability for different chat UIs, orchestrators, or custom agent stacks. It emphasizes symbol-level understanding rather than naive file-wide diffs, enabling more precise refactors and additions. The repository and ecosystem materials highlight rapid setup, agent interoperability, and examples that show agents iterating on a codebase with guardrails. It’s actively maintained by Oraios, with recent updates, community showcases, and third-party write-ups underscoring interest from the agent tooling community.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Skyvern

    Skyvern

    Automate browser-based workflows with LLMs and Computer Vision

    Skyvern uses a combination of computer vision and AI to understand content on a webpage, making it adaptable to any website. Skyvern takes instructions in natural language, allowing it to execute complex objectives with simple commands. Skyvern is an API-first product. Workflows execute in the cloud, allowing it to run hundreds of workflows at the same time. Skyvern's AI decisions come with built-in explanations, providing clear summaries and justifications for every action. Support for proxies, with support for country, state, or even precise zip-code level targeting. Skyvern understands how to solve CAPTCHAs to complete complicated workflows. Support for authenticating into user accounts, including support for 2FA/TOTP. Extract data from workflows in any schema of your choice including CSV or JSON. Automate procurement pipelines, breeze through government forms, and complete workflows in any language.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Stanza

    Stanza

    Stanford NLP Python library for many human languages

    Stanza is a collection of accurate and efficient tools for the linguistic analysis of many human languages. Starting from raw text to syntactic analysis and entity recognition, Stanza brings state-of-the-art NLP models to languages of your choosing. Stanza is a Python natural language analysis package. It contains tools, which can be used in a pipeline, to convert a string containing human language text into lists of sentences and words, to generate base forms of those words, their parts of speech and morphological features, to give a syntactic structure dependency parse, and to recognize named entities. The toolkit is designed to be parallel among more than 70 languages, using the Universal Dependencies formalism. Stanza is built with highly accurate neural network components that also enable efficient training and evaluation with your own annotated data.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    Streamer-Sales

    Streamer-Sales

    LLM Large Model of Selling Anchor

    Streamer-Sales is an open-source large language model system designed specifically for e-commerce live streaming and automated product promotion. The project focuses on generating persuasive product descriptions and live presentation scripts that mimic the style of professional online sales hosts. By analyzing product characteristics and marketing information, the model can produce engaging explanations that emphasize benefits, features, and emotional appeal to encourage viewers to make purchasing decisions. The system integrates multiple AI technologies including retrieval-augmented generation to incorporate product knowledge, speech synthesis to convert generated scripts into voice output, and digital human generation to create virtual hosts. It also supports automatic speech recognition and agent-based tools that can retrieve additional information such as logistics or product details during live sessions.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Supabase MCP Server

    Supabase MCP Server

    Query MCP enables end-to-end management of Supabase via chat interface

    An open-source MCP server that enables comprehensive management of Supabase projects through natural language interactions, providing capabilities such as SQL execution, schema management, and API integration. ​
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    SwarmUI

    SwarmUI

    Modular AI image and video generation web UI with extensible tools

    SwarmUI is a modular web-based user interface designed for AI-driven image generation, with a strong focus on usability, performance, and extensibility. It serves as a unified environment for working with multiple AI models, including Stable Diffusion and newer image and video generation systems, allowing users to create and manage outputs through a browser interface. SwarmUI is built to accommodate both beginners and advanced users by offering a simple “Generate” interface alongside more advanced workflow tools that expose deeper configuration options. It integrates with underlying systems like node-based workflows, enabling flexible and customizable pipelines for complex generation tasks. SwarmUI also emphasizes scalability, originally inspired by the idea of coordinating multiple GPUs to work together for large batch or grid-based image generation. SwarmUI includes a variety of built-in tools such as image editing, prompt handling, and automation features.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Swarms

    Swarms

    Enterprise multi-agent orchestration framework for scalable AI apps

    Swarms is an enterprise-grade multi-agent orchestration framework designed to help developers build, manage, and scale collaborative AI systems composed of multiple agents. It provides a structured infrastructure for coordinating agents in hierarchical, parallel, or sequential workflows, enabling complex task execution across distributed components. It emphasizes production readiness, offering modular architecture, high availability, and observability features suitable for large-scale deployments. It supports integration with multiple model providers and existing ecosystems, allowing developers to combine different AI tools and frameworks within a unified system. Swarms also includes mechanisms for agent lifecycle management, memory handling, and dynamic composition, making it adaptable to evolving workloads. Additionally, it focuses on developer productivity through APIs, CLI tools, and templates that simplify building and deploying agent-based applications.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    Taipy

    Taipy

    Turns Data and AI algorithms into production-ready web applications

    From simple pilots to production-ready web applications in no time. No more compromise on performance, customization, and scalability. Taipy enhances performance with caching control of graphical events, optimizing rendering by selectively updating graphical components only upon interaction. Effortlessly manage massive datasets with Taipy's built-in decimator for charts, intelligently reducing the number of data points to save time and memory without losing the essence of your data's shape. Struggle with sluggish performance and excessive memory usage, as every data point demands processing. Large datasets become cumbersome, complicating the user experience and data analysis. Scenarios are made easy with Taipy Studio. A powerful VS Code extension that unlocks a convenient graphical editor. Get your methods invoked at a certain time or intervals. Enjoy a variety of predefined themes or build your own.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    TapeAgents

    TapeAgents

    A framework that facilitates all stages of LLM development

    TapeAgents is a framework that facilitates all stages of the Large Language Model (LLM) agent development lifecycle, providing tools for building, testing, and deploying AI agents.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    TaskingAI

    TaskingAI

    Open platform for building, deploying, and managing LLM agents

    TaskingAI is an open source platform designed to simplify the development and deployment of applications powered by large language models. It follows a Backend as a Service approach, allowing developers to separate AI logic from frontend product development while maintaining a structured and scalable workflow. TaskingAI integrates hundreds of language models from multiple providers into a unified system, enabling developers to switch models or combine capabilities without major reconfiguration. It includes a modular architecture that supports components such as assistants, tools, retrieval systems, and conversation management, all accessible through a consistent interface. TaskingAI also provides a built-in user interface for managing projects, testing workflows, and configuring AI agents without needing to rely entirely on code. It supports advanced techniques like retrieval-augmented generation and plugin-based extensions, allowing developers to enhance agent capabilities.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    Tensorforce

    Tensorforce

    A TensorFlow library for applied reinforcement learning

    Tensorforce is an open-source deep reinforcement learning framework built on TensorFlow, emphasizing modularized design and straightforward usability for applied research and practice.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    Tongyi DeepResearch

    Tongyi DeepResearch

    Tongyi Deep Research, the Leading Open-source Deep Research Agent

    DeepResearch (Tongyi DeepResearch) is an open-source “deep research agent” developed by Alibaba’s Tongyi Lab designed for long-horizon, information-seeking tasks. It’s built to act like a research agent: synthesizing, reasoning, retrieving information via the web and documents, and backing its outputs with evidence. The model is about 30.5 billion parameters in size, though at any given token only ~3.3B parameters are active. It uses a mix of synthetic data generation, fine-tuning and reinforcement learning; supports benchmarks like web search, document understanding, question answering, “agentic” tasks; provides inference tools, evaluation scripts, and “web agent” style interfaces. The aim is to enable more autonomous, agentic models that can perform sustained knowledge gathering, reasoning, and synthesis across multiple modalities (web, files, etc.).
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Torch Pruning

    Torch Pruning

    DepGraph: Towards Any Structural Pruning

    Torch-Pruning is an open-source toolkit designed to optimize deep neural networks by performing structural pruning directly within PyTorch models. The library focuses on reducing the size and computational cost of neural networks by removing redundant parameters and channels while maintaining model performance. It introduces a graph-based algorithm called DepGraph that automatically identifies dependencies between layers, allowing parameters to be pruned safely across complex architectures. This dependency analysis makes it possible to prune large networks such as transformers, convolutional networks, and diffusion models without breaking the computational graph. Torch-Pruning physically removes parameters rather than masking them, which results in smaller and faster models during both training and inference. The toolkit supports a wide variety of architectures used in computer vision and large language models, making it a flexible solution for model compression tasks.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    TorchRec

    TorchRec

    Pytorch domain library for recommendation systems

    TorchRec is a PyTorch domain library built to provide common sparsity & parallelism primitives needed for large-scale recommender systems (RecSys). It allows authors to train models with large embedding tables sharded across many GPUs. Parallelism primitives that enable easy authoring of large, performant multi-device/multi-node models using hybrid data-parallelism/model-parallelism. The TorchRec sharder can shard embedding tables with different sharding strategies including data-parallel, table-wise, row-wise, table-wise-row-wise, and column-wise sharding. The TorchRec planner can automatically generate optimized sharding plans for models. Pipelined training overlaps dataloading device transfer (copy to GPU), inter-device communications (input_dist), and computation (forward, backward) for increased performance. Optimized kernels for RecSys powered by FBGEMM. Quantization support for reduced precision training and inference. Common modules for RecSys.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    TradeMaster

    TradeMaster

    TradeMaster is an open-source platform for quantitative trading

    TradeMaster is a first-of-its-kind, best-in-class open-source platform for quantitative trading (QT) empowered by reinforcement learning (RL), which covers the full pipeline for the design, implementation, evaluation and deployment of RL-based algorithms. TradeMaster is composed of 6 key modules: 1) multi-modality market data of different financial assets at multiple granularities; 2) whole data preprocessing pipeline; 3) a series of high-fidelity data-driven market simulators for mainstream QT tasks; 4) efficient implementations of over 13 novel RL-based trading algorithms; 5) systematic evaluation toolkits with 6 axes and 17 measures; 6) different interfaces for interdisciplinary users.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    TypeChat

    TypeChat

    Library for building type-safe natural language interfaces with LLMs

    TypeChat is an open source library developed by Microsoft that simplifies the creation of natural language interfaces by using type definitions to structure interactions with large language models. Traditional natural language interfaces often relied on complex decision trees to interpret user intent and gather required inputs. With the rise of large language models, developers can interpret user requests more easily, but they still face challenges related to output reliability, safety, and structured responses. TypeChat addresses these challenges by replacing traditional prompt engineering with a concept called schema engineering. Instead of writing complex prompts, developers define types that represent the intents supported by their applications. It then uses those type definitions to construct prompts for language models and translate user input into structured data that follows the defined schema.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    UMAP

    UMAP

    Uniform Manifold Approximation and Projection

    Uniform Manifold Approximation and Projection (UMAP) is a dimension reduction technique that can be used for visualization similarly to t-SNE, but also for general non-linear dimension reduction. It is possible to model the manifold with a fuzzy topological structure. The embedding is found by searching for a low-dimensional projection of the data that has the closest possible equivalent fuzzy topological structure. First of all UMAP is fast. It can handle large datasets and high dimensional data without too much difficulty, scaling beyond what most t-SNE packages can manage. This includes very high dimensional sparse datasets. UMAP has successfully been used directly on data with over a million dimensions. Second, UMAP scales well in the embedding dimension—it isn't just for visualization. You can use UMAP as a general-purpose dimension reduction technique as a preliminary step to other machine learning tasks.
    Downloads: 3 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB