Open Source Python Artificial Intelligence Software - Page 53

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Spotlight

    Spotlight

    Deep recommender models using PyTorch

    Spotlight uses PyTorch to build both deep and shallow recommender models. By providing both a slew of building blocks for loss functions (various pointwise and pairwise ranking losses), representations (shallow factorization representations, deep sequence models), and utilities for fetching (or generating) recommendation datasets, it aims to be a tool for rapid exploration and prototyping of new recommender models. Spotlight offers a slew of popular datasets, including Movielens 100K, 1M, 10M, and 20M. It also incorporates utilities for creating synthetic datasets. For example, generate_sequential generates a Markov-chain-derived interaction dataset, where the next item a user chooses is a function of their previous interactions. Recommendations can be seen as a sequence prediction task: given the items a user has interacted with in the past, what will be the next item they will interact with? Spotlight provides a range of models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Stable-Dreamfusion

    Stable-Dreamfusion

    Text-to-3D & Image-to-3D & Mesh Exportation with NeRF + Diffusion

    A pytorch implementation of the text-to-3D model Dreamfusion, powered by the Stable Diffusion text-to-2D model. This project is a work-in-progress and contains lots of differences from the paper. The current generation quality cannot match the results from the original paper, and many prompts still fail badly! Since the Imagen model is not publicly available, we use Stable Diffusion to replace it (implementation from diffusers). Different from Imagen, Stable-Diffusion is a latent diffusion model, which diffuses in a latent space instead of the original image space. Therefore, we need the loss to propagate back from the VAE's encoder part too, which introduces extra time costs in training. We use the multi-resolution grid encoder to implement the NeRF backbone (implementation from torch-ngp), which enables much faster rendering.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    StarVector

    StarVector

    StarVector is a foundation model for SVG generation

    StarVector is a multimodal foundation model designed for generating Scalable Vector Graphics (SVG) from images or textual descriptions. The system treats vector graphics creation as a code generation problem, producing SVG code that can render detailed vector images. Its architecture combines computer vision techniques with language modeling capabilities so it can understand visual inputs and textual prompts simultaneously. The model converts raster images or text instructions into structured vector representations, enabling high-quality vectorization and design generation. This approach allows StarVector to create scalable graphics that maintain visual quality regardless of resolution, which is especially useful for design tools and illustration workflows. Because the model produces SVG code rather than pixel images, the output can be edited programmatically or integrated directly into web and design environments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    StellarGraph

    StellarGraph

    Machine Learning on Graphs

    StellarGraph is a Python library for machine learning on graphs and networks. The StellarGraph library offers state-of-the-art algorithms for graph machine learning, making it easy to discover patterns and answer questions about graph-structured data. It can solve many machine learning tasks. Graph-structured data represent entities as nodes (or vertices) and relationships between them as edges (or links), and can include data associated with either as attributes. For example, a graph can contain people as nodes and friendships between them as links, with data like a person’s age and the date a friendship was established. StellarGraph supports the analysis of many kinds of graphs. StellarGraph is built on TensorFlow 2 and its Keras high-level API, as well as Pandas and NumPy. It is thus user-friendly, modular and extensible. It interoperates smoothly with code that builds on these, such as the standard Keras layers and scikit-learn.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Step-Audio-EditX

    Step-Audio-EditX

    LLM-based Reinforcement Learning audio edit model

    Step-Audio-EditX is an open-source, 3 billion-parameter audio model from StepFun AI designed to make expressive and precise editing of speech and audio as easy as text editing. Rather than treating audio editing as low-level waveform manipulation, this model converts speech into a sequence of discrete “audio tokens” (via a dual-codebook tokenizer) — combining a linguistic token stream and a semantic (prosody/emotion/style) token stream — thereby abstracting audio editing into high-level token operations. This allows users to modify not only what is said (the text) but also how it's said: emotion, tone, speaking style, prosody, accent, even paralinguistic cues. Because the model is trained with a “large-margin learning” objective over many synthesized and natural speech samples, it gains robust control over expressive attributes, and can perform iterative editing: e.g. you could record a line, then ask the model to “make it sadder,” “speak slower,” or “change accent to X.”
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Step1X-3D

    Step1X-3D

    High-Fidelity and Controllable Generation of Textured 3D Assets

    Step1X-3D is an open-source framework for generating high-fidelity textured 3D assets from scratch — both their geometry and surface textures — using modern generative AI techniques. It combines a hybrid architecture: a geometry generation stage using a VAE-DiT model to output a watertight 3D representation (e.g. TSDF surface), and a texture synthesis stage that conditions on geometry and optionally reference input (or prompts) to produce view-consistent textures using a diffusion-based texture module. The result is fully 3D assets — meshes + textures — which can be rendered from any viewpoint, textured consistently, and used in 3D applications. To achieve this, the project includes a massive curated dataset: among more than 5 million candidate 3D assets, it filters and standardizes to produce a high-quality 2 million–asset subset suitable for training.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Story Flicks

    Story Flicks

    Generate high-definition story short videos with one click using AI

    Story Flicks is another open-source project in the AI-assisted video generation / editing space, focused on creating short, story-style videos from script or prompt inputs. It aims to let users generate high-definition short movies or video stories with minimal manual effort, using AI models under the hood to assemble visuals, timing, and possibly narration or subtitles. For creators who want to produce narrative short-form content — whether for social media, storytelling, or prototyping video ideas — story-flicks offers a lightweight, code-backed alternative to complex video editing suites. Because the project is open and modifiable, developers can customize the generation pipeline: adjust story structure, alter rendering parameters, tweak video quality or resolution, or integrate with other AI models (e.g. for audio, voice-over, or image-to-video). It’s especially useful as a starting template or experimentation ground for developers building automated content-creation tools.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    StreamSpeech

    StreamSpeech

    StreamSpeech is a seamless model for offline speech recognition

    StreamSpeech is an “all-in-one” speech model designed to perform offline and simultaneous speech recognition, speech translation, and speech synthesis within a single unified architecture. Developed as part of an ACL 2024 paper, it targets streaming and low-latency scenarios where intermediate results and final translations or synthetic speech must be produced continuously as audio is being received. The model supports eight tasks: offline ASR, speech-to-text translation, speech-to-speech translation, and TTS, as well as their streaming or simultaneous counterparts, all handled by the same underlying system. During simultaneous translation, StreamSpeech can optionally output intermediate ASR transcripts and text translations, giving users or downstream applications real-time visibility into what the system is hearing and how it is translating.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    SuperAGI

    SuperAGI

    A dev-first open source autonomous AI agent framework

    An open-source autonomous AI framework to enable you to develop and deploy useful autonomous agents quickly & reliably. Join a community of developers constantly contributing to make SuperAGI better. Access your agents through a graphical user interface. Interact with agents by giving them input, permissions, etc. Agents typically learn and improve their performance over time with feedback loops. Run multiple agents simultaneously to improve efficiency and productivity. Connect to multiple Vector DBs to enhance your agent’s performance. Each agent is unique, use different models of your choice. Get insights into your agent’s performance and optimize accordingly. Control token usage to manage costs effectively. Enable your agents to learn and adapt by storing their memory. Get notified when agents get stuck in the loop, and provide proactive resolution. Read and store files generated by Agents.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    Sygil WebUI

    Sygil WebUI

    Stable Diffusion web UI

    Sygil WebUI is a browser-based interface for running Stable Diffusion image generation locally or on a server, wrapping common text-to-image and image-to-image workflows into a practical UI. It provides multiple UI modes (including a legacy Gradio interface) and focuses on making iterative prompting, parameter tuning, and post-processing accessible without writing code. The UI exposes core generation controls like resolution, CFG guidance, sampling steps, samplers, seeds, and batch generation so users can reproduce results and refine outputs systematically. It also supports jumping between workflows, such as sending an output directly into Image2Image for variations or into an “Image Lab” style area for enhancement and upscaling. Post-processing and enhancement are a major emphasis: the interface can route images through different upscalers and face-enhancement tools, helping users turn raw generations into cleaner, higher-resolution results.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    TNT

    TNT

    A lightweight library for PyTorch training tools and utilities

    TNT is a lightweight training framework developed by Meta that simplifies the process of building and managing machine learning training loops using PyTorch. The project focuses on providing a flexible yet structured environment for implementing training pipelines without the complexity of large deep learning frameworks. It introduces modular abstractions that allow developers to organize training logic into reusable components such as trainers, evaluators, and callbacks. This design helps separate concerns such as model training, evaluation, logging, and checkpointing, making machine learning experiments easier to manage. The framework is particularly useful for large-scale experiments where maintaining clear training workflows becomes increasingly important. Because it is built on top of PyTorch, the framework integrates naturally with existing deep learning models and datasets.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    Tacotron-2

    Tacotron-2

    DeepMind's Tacotron-2 Tensorflow implementation

    Tacotron-2 is a TensorFlow implementation of DeepMind’s Tacotron-2 end-to-end text-to-speech architecture, which predicts mel spectrograms from raw text and then feeds them to a neural vocoder such as WaveNet. It reproduces the original paper’s hyperparameters exactly via paper_hparams.py, while also offering a tuned hparams.py with extra improvements that often yield better audio quality in practice. The repository is structured as a full training pipeline: dataset preparation, preprocessing into spectrograms, Tacotron training, WaveNet (or Griffin-Lim) vocoder training, and final waveform synthesis. It includes directory layouts and logging directories for multiple datasets such as LJSpeech and M-AILABS en_US/en_UK, making it easier to adapt to new English corpora. Separate log trees track mel-spectrograms, attention plots, evaluation audio, and vocoder outputs, so you can inspect how alignment and audio quality evolve over time.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    TalkingHeads

    TalkingHeads

    A library to communicate with ChatGPT, Claude, Copilot, Gemini

    TalkingHeads is a Python library designed to facilitate communication with various AI chat agents, including ChatGPT, Claude, Copilot, Gemini, HuggingChat, and Pi. It provides a unified interface for interacting with these platforms, simplifying the integration of conversational AI capabilities into applications. TalkingHeads supports browser automation and offers tools to manage sessions, handle prompts, and process responses effectively.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Tencent-Hunyuan-Large

    Tencent-Hunyuan-Large

    Open-source large language model family from Tencent Hunyuan

    Tencent-Hunyuan-Large is the flagship open-source large language model family from Tencent Hunyuan, offering both pre-trained and instruct (fine-tuned) variants. It is designed with long-context capabilities, quantization support, and high performance on benchmarks across general reasoning, mathematics, language understanding, and Chinese / multilingual tasks. It aims to provide competitive capability with efficient deployment and inference. FP8 quantization support to reduce memory usage (~50%) while maintaining precision. High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Tensor2Tensor

    Tensor2Tensor

    Library of deep learning models and datasets

    Deep Learning (DL) has enabled the rapid advancement of many useful technologies, such as machine translation, speech recognition and object detection. In the research community, one can find code open-sourced by the authors to help in replicating their results and further advancing deep learning. However, most of these DL systems use unique setups that require significant engineering effort and may only work for a specific problem or architecture, making it hard to run new experiments and compare the results. Tensor2Tensor, or T2T for short, is a library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research. T2T was developed by researchers and engineers in the Google Brain team and a community of users. It is now deprecated, we keep it running and welcome bug-fixes, but encourage users to use the successor library Trax.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    TensorFlow Quantum

    TensorFlow Quantum

    Open-source Python framework for hybrid quantum-classical ml learning

    TensorFlow Quantum is an open-source software framework designed for building and training hybrid quantum-classical machine learning models within the TensorFlow ecosystem. The framework enables researchers and developers to represent quantum circuits as data and integrate them directly into machine learning workflows. By combining classical deep learning techniques with quantum algorithms, the platform allows experimentation with quantum machine learning methods that may offer advantages for certain computational tasks. TensorFlow Quantum integrates with the Cirq quantum computing framework to define and manipulate quantum circuits, while leveraging TensorFlow’s infrastructure for optimization, automatic differentiation, and large-scale computation. The library also supports high-performance simulation of quantum circuits, enabling researchers to test and evaluate quantum models even without direct access to quantum hardware.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    TensorLy

    TensorLy

    Tensor Learning in Python

    TensorLy is a Python library that aims at making tensor learning simple and accessible. It allows to easily perform tensor decomposition, tensor learning and tensor algebra. Its backend system allows to seamlessly perform computation with NumPy, PyTorch, JAX, TensorFlow, CuPy or Paddle, and run methods at scale on CPU or GPU.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Text-to-LoRA (T2L)

    Text-to-LoRA (T2L)

    Hypernetworks that adapt LLMs for specific benchmark tasks

    Text-to-LoRA is a research project that introduces a method for dynamically adapting large language models using hypernetworks that generate LoRA parameters directly from textual descriptions. Instead of training a new LoRA adapter for every task or dataset, the system can produce task-specific adaptations based solely on a text description of the desired capability. This approach enables models to rapidly internalize new contextual knowledge without performing traditional fine-tuning steps. The project provides a reference implementation of the Doc-to-LoRA method, which allows language models to quickly encode factual information or contextual constraints into lightweight LoRA modules. Developers and researchers can experiment with how textual task descriptions can generate LoRA weights that modify model behavior in real time.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    TextGen

    TextGen

    textgen, Text Generation models

    Implementation of Text Generation models. textgen implements a variety of text generation models, including UDA, GPT2, Seq2Seq, BART, T5, SongNet and other models, out of the box. UDA, non-core word replacement. EDA, simple data augmentation technique: similar words, synonym replacement, random word insertion, deletion, replacement. This project refers to Google's UDA (non-core word replacement) algorithm and EDA algorithm, based on TF-IDF to replace some unimportant words in sentences with synonyms, random word insertion, deletion, replacement, etc. method, generating new text and implementing text augmentation This project realizes the back translation function based on Baidu translation API, first translate Chinese sentences into English, and then translate English into new Chinese. This project implements the training and prediction of Seq2Seq, ConvSeq2Seq, and BART models based on PyTorch, which can be used for text generation tasks such as text translation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    The Hundred-Page Machine Learning Book

    The Hundred-Page Machine Learning Book

    The Python code to reproduce illustrations from Machine Learning Book

    The Hundred-Page Machine Learning Book is the official companion repository for The Hundred-Page Machine Learning Book written by machine learning researcher Andriy Burkov. The repository contains Python code used to generate the figures, visualizations, and illustrative examples presented in the book. Its purpose is to help readers better understand the concepts explained in the text by allowing them to run and experiment with the underlying code themselves. The book itself provides a concise overview of machine learning theory and practice, covering topics such as supervised learning, unsupervised learning, neural networks, and optimization algorithms. The repository complements these explanations by offering practical implementations that demonstrate how various algorithms behave when applied to data. Readers can explore the scripts to reproduce diagrams and observe how mathematical concepts translate into working code.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Toad

    Toad

    Unified terminal AI tool for exploring and editing codebases

    Toad is an open source, terminal-first AI interface designed to unify multiple coding agents into a single workflow. It allows developers to interact with AI models directly inside the command line, making it easier to explore, understand, and modify codebases without leaving the terminal. Built in Python, it focuses on transparency and control by letting users load context intentionally and inspect how the AI processes files. Toad supports structured conversations, enabling navigation through code with clear references instead of opaque outputs. Inspired by notebook-style workflows, it allows reuse of previous interactions and exporting of results. Toad is vendor-agnostic, meaning it can work with different AI agents while maintaining a consistent user experience. It emphasizes developer intent over automation, keeping humans in control of decisions while using AI as a collaborative assistant for coding, refactoring, and analysis.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    TorchMetrics

    TorchMetrics

    Machine learning metrics for distributed, scalable PyTorch application

    TorchMetrics is a collection of 80+ PyTorch metrics implementations and an easy-to-use API to create custom metrics. Your data will always be placed on the same device as your metrics. You can log Metric objects directly in Lightning to reduce even more boilerplate. The module-based metrics contain internal metric states (similar to the parameters of the PyTorch module) that automate accumulation and synchronization across devices! Automatic accumulation over multiple batches. Automatic synchronization between multiple devices. Metric arithmetic. Similar to torch.nn, most metrics have both a module-based and a functional version. The functional versions are simple python functions that as input take torch.tensors and return the corresponding metric as a torch.tensor.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    TorchMetrics AI

    TorchMetrics AI

    Machine learning metrics for distributed, scalable PyTorch application

    TorchMetrics is a collection of 100+ PyTorch metrics implementations and an easy-to-use API to create custom metrics.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Tracking Any Point (TAP)

    Tracking Any Point (TAP)

    DeepMind model for tracking arbitrary points across videos & robotics

    TAPNet is the official Google DeepMind repository for Tracking Any Point (TAP), bundling datasets, models, benchmarks, and demos for precise point tracking in videos. The project includes the TAP-Vid and TAPVid-3D benchmarks, which evaluate long-range tracking of arbitrary points in 2D and 3D across diverse real and synthetic videos. Its flagship models—TAPIR, BootsTAPIR, and the latest TAPNext—use matching plus temporal refinement or next-token style propagation to achieve state-of-the-art accuracy and speed on TAP-Vid. RoboTAP demonstrates how TAPIR-style tracks can drive real-world robot manipulation via efficient imitation, and ships with a dataset of annotated robotics videos. The repo provides JAX and PyTorch checkpoints, Colab demos, and a real-time live demo that runs on a GPU to let you select and track points interactively.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    TreeQuest

    TreeQuest

    A Tree Search Library with Flexible API for LLM Inference-Time Scaling

    TreeQuest, developed by SakanaAI, is a versatile Python library implementing adaptive tree search algorithms—such as AB‑MCTS—for enhancing inference-time performance of large language models (LLMs). It allows developers to define custom state-generation and scoring functions (e.g., via LLMs), and then efficiently explores possible answer trees during runtime. With support for multi-LLM collaboration, checkpointing, and mixed policies, TreeQuest enables smarter, trial‑and‑error question answering by leveraging both breadth (multiple attempts) and depth (iterative refinement) strategies to find better outputs dynamically
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB