Open Source Python Artificial Intelligence Software - Page 96

Python Artificial Intelligence Software

View 13574 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    SlowFast

    SlowFast

    Video understanding codebase from FAIR for reproducing video models

    SlowFast is a video understanding framework that captures both spatial semantics and temporal dynamics efficiently by processing video frames at two different temporal resolutions. The slow pathway encodes semantic context by sampling frames sparsely, while the fast pathway captures motion and fine temporal cues by operating on densely sampled frames with fewer channels. Together, these two pathways complement each other, allowing the network to model both appearance and motion without excessive computational cost. The architecture is modular and supports tasks like action recognition, temporal localization, and video segmentation, performing strongly on benchmarks like Kinetics and AVA. The repository provides training recipes, pretrained models, and distributed pipelines optimized for large-scale video datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Spyse is a software framework for building multi-agent systems. It allows Python developers to build distributed intelligent systems of multiple cooperative agents based on FIPA, OWL, SOA and many others. Spyse is designed for ease-of-use and fun.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Design and develop Recommendation and Adaptive Prediction Engines to address eCommerce opportunities. Build a portfolio of engines by creating and porting algorithms from multiple disciplines to a usable form. Try to solve NetFlix and other challenges.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Snips NLU

    Snips NLU

    Snips Python library to extract meaning from text

    Snips NLU is a Natural Language Understanding python library that allows to parse sentences written in natural language, and extract structured information. It’s the library that powers the NLU engine used in the Snips Console that you can use to create awesome and private-by-design voice assistants. The exact output is a bit richer, the point here is to give a glimpse on what kind of information can be extracted. Behind every chatbot and voice assistant lies a common piece of technology: Natural Language Understanding (NLU). Anytime a user interacts with an AI using natural language, their words need to be translated into a machine-readable description of what they meant. The NLU engine first detects what the intention of the user is (a.k.a. intent), then extracts the parameters (called slots) of the query. The developer can then use this to determine the appropriate action or response.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    Sockeye

    Sockeye

    Sequence-to-sequence framework, focused on Neural Machine Translation

    Sockeye is an open-source sequence-to-sequence framework for Neural Machine Translation built on PyTorch. It implements distributed training and optimized inference for state-of-the-art models, powering Amazon Translate and other MT applications. For a quickstart guide to training a standard NMT model on any size of data, see the WMT 2014 English-German tutorial. If you are interested in collaborating or have any questions, please submit a pull request or issue. You can also send questions to sockeye-dev-at-amazon-dot-com. Developers may be interested in our developer guidelines. Starting with version 3.0.0, Sockeye is also based on PyTorch. We maintain backwards compatibility with MXNet models of version 2.3.x with 3.0.x. If MXNet 2.x is installed, Sockeye can run both with PyTorch or MXNet. All models trained with 2.3.x (using MXNet) can be converted to models running with PyTorch using the converter CLI (sockeye.mx_to_pt).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Sofa is a CUDA-based reasoner
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    A general purpose reinforcement learning engine.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    SparseML

    SparseML

    Libraries for applying sparsification recipes to neural networks

    SparseML is an optimization toolkit for training and deploying deep learning models using sparsification techniques like pruning and quantization to improve efficiency.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    SpeakLogPSU
    SpeakLogPSU can speak chat messages with an individual voice if the NPC or player was configured or with a default one. You will never miss if someone talks to you. Voice cloning can be accomplished with Coqui in less than five minutes without GPU. The result is archived and can be used the next time in game. Some TTS projects already started to add tag support to speak text with emotions or sing it. If a game designer has that in mind with a good chat log she can voiced her game over night. reads the log and sends new chat text to piper. ~/.config/Epic/PSUnreal/Saved/Logs/Pongo_Donjo_chat.txt If a line number is set it can speak all the chat text and waits for new chat text. Python 3.8.10 download: https://www.planeshift.it/Download https://github.com/rhasspy/piper
    Downloads: 0 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10

    Spectral Python

    A python module for hyperspectral image processing

    Spectral Python (SPy) is a python package for reading, viewing, manipulating, and classifying hyperspectral image (HSI) data. SPy includes functions for clustering, dimensionality reduction, supervised classification, and more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Speect
    Speect is a multilingual TTS system. It offers a full text-to-speech system with various API's, as well as an environment for research and development of TTS systems and voices. It is written in ANSI C and uses a plug-in mechanism for extensions. Speect also includes an extensive set of Python bindings for quick implementation of new ideas, these bindings are derived from SWIG interface files and can easily be extended for other languages supported by SWIG. Speect is free and open source software. As a collection it is distributed under a MIT license.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Spektral

    Spektral

    Graph Neural Networks with Keras and Tensorflow 2

    Spektral is a Python library for graph deep learning, based on the Keras API and TensorFlow 2. The main goal of this project is to provide a simple but flexible framework for creating graph neural networks (GNNs). You can use Spektral for classifying the users of a social network, predicting molecular properties, generating new graphs with GANs, clustering nodes, predicting links, and any other task where data is described by graphs. Spektral implements some of the most popular layers for graph deep learning. Spektral also includes lots of utilities for representing, manipulating, and transforming graphs in your graph deep learning projects. Spektral is compatible with Python 3.6 and above, and is tested on the latest versions of Ubuntu, MacOS, and Windows. Other Linux distros should work as well. The 1.0 release of Spektral is an important milestone for the library and brings many new features and improvements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    SpikingJelly

    SpikingJelly

    SpikingJelly is an open-source deep learning framework

    SpikingJelly is an open-source deep learning framework for spiking neural networks that is primarily built on top of PyTorch and aimed at neuromorphic computing research. The project provides the components needed to build, train, and evaluate neural models that communicate through discrete spikes rather than the continuous activations used in conventional artificial neural networks. This makes it especially relevant for researchers interested in biologically inspired computing, event-driven processing, and energy-efficient AI systems. The framework includes neuron models, surrogate gradient training methods, encoding strategies, network components, and utilities for simulation and experimentation, allowing users to develop a wide variety of spiking architectures. It also supports integration with familiar PyTorch workflows, which lowers the barrier for machine learning practitioners who want to explore spiking approaches without abandoning mainstream tooling.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Spinning Up in Deep RL

    Spinning Up in Deep RL

    Educational resource to help anyone learn deep reinforcement learning

    Welcome to Spinning Up in Deep RL! This is an educational resource produced by OpenAI that makes it easier to learn about deep reinforcement learning (deep RL). For the unfamiliar, reinforcement learning (RL) is a machine learning approach for teaching agents how to solve tasks by trial and error. Deep RL refers to the combination of RL with deep learning. At OpenAI, we believe that deep learning generally, and deep reinforcement learning specifically, will play central roles in the development of powerful AI technology. To ensure that AI is safe, we have to come up with safety strategies and algorithms that are compatible with this paradigm. As a result, we encourage everyone who asks this question to study these fields. However, while there are many resources to help people quickly ramp up on deep learning, deep reinforcement learning is more challenging to break into.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Spring PyAI is a wrapper for the SkirmishAI Interface of the SpringRTS Engine. More Infrmation about SpringRTS can be found at http://www.springrts.com
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Stable Baselines

    Stable Baselines

    A fork of OpenAI Baselines, implementations of reinforcement learning

    Stable Baselines is a set of improved implementations of reinforcement learning algorithms based on OpenAI Baselines. You can read a detailed presentation of Stable Baselines in the Medium article. These algorithms will make it easier for the research community and industry to replicate, refine, and identify new ideas, and will create good baselines to build projects on top of. We expect these tools will be used as a base around which new ideas can be added, and as a tool for comparing a new approach against existing ones. We also hope that the simplicity of these tools will allow beginners to experiment with a more advanced toolset, without being buried in implementation details.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Stable Diffusion WebUI Forge

    Stable Diffusion WebUI Forge

    Stable Diffusion WebUI Forge is a platform on top of Stable Diffusion

    Stable Diffusion WebUI Forge is a performance- and feature-oriented fork of the popular AUTOMATIC1111 interface that experiments with new backends, memory optimizations, and UX improvements. It targets heavy users and researchers who push large models, control nets, and high-resolution pipelines where default settings can become bottlenecks. The fork typically introduces toggles for scheduler behavior, attention implementations, caching, and precision modes to reach better speed or quality on given hardware. It also focuses on stability during long sessions, aiming to reduce out-of-memory failures and provide clearer diagnostics when they occur. The UI surfaces advanced options in a way that remains recognizable to WebUI users, so migration costs are low while gaining experimental features. In practice, Forge serves as a proving ground for ideas that may later influence upstream tools, giving power users early access to cutting-edge techniques.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Stable Diffusion in Docker

    Stable Diffusion in Docker

    Run the Stable Diffusion releases in a Docker container

    Run the Stable Diffusion releases in a Docker container with txt2img, img2img, depth2img, pix2pix, upscale4x, and inpaint. Run the Stable Diffusion releases on Huggingface in a GPU-accelerated Docker container. By default, the pipeline uses the full model and weights which requires a CUDA capable GPU with 8GB+ of VRAM. It should take a few seconds to create one image. On less powerful GPUs you may need to modify some of the options; see the Examples section for more details. If you lack a suitable GPU you can set the options --device cpu and --onnx instead. Since it uses the model, you will need to create a user access token in your Huggingface account. Save the user access token in a file called token.txt and make sure it is available when building the container. Create an image from an existing image and a text prompt. Modify an existing image with its depth map and a text prompt.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    StarGAN

    StarGAN

    Official PyTorch Implementation

    StarGAN is an implementation of the Star Generative Adversarial Network, a model designed for multi-domain image-to-image translation using a single unified GAN architecture. Unlike earlier GAN approaches that required separate models for each domain pair, StarGAN enables flexible attribute transfer across multiple domains within one network, significantly improving efficiency and scalability. The repository includes full training and inference pipelines for tasks such as facial attribute manipulation and style transfer. It demonstrates adversarial training strategies, domain classification losses, and generator-discriminator coordination required for stable multi-domain translation. Researchers and practitioners often use the project as a reference when studying conditional GANs and advanced image synthesis techniques. Overall, the repository provides a clean and practical baseline for experimenting with multi-domain generative modeling in PyTorch.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    State of Open Source AI

    State of Open Source AI

    Clarity in the current fast-paced mess of Open Source innovation

    This repository is the source for a book (or large written work) titled “The State of Open Source AI”. The goal of the project is to bring clarity to the rapidly evolving open-source AI ecosystem by documenting trends, models, tools, standards, deployment practices, and challenges. It acts as both a snapshot and a guide: readers can see what’s “hot now” in open AI infrastructure, what open licensing or governance issues are emerging, how deployment options compare, and what gaps remain. Because the AI domain moves quickly, part of the aim is to make the content maintainable and updateable by the community. The structure includes chapters or sections about model formats, evaluation benchmarks, hardware/backends, MLOps systems, alignment and safety issues, and open datasets. The repository contains the text (in Markdown or similar), configuration for build or publishing (static site or e-book), and contributor guidelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Step-Audio

    Step-Audio

    Open-source framework for intelligent speech interaction

    Step-Audio is a unified, open-source framework aimed at building intelligent speech systems that combine both comprehension and generation: it integrates large language models (LLMs) with speech input/output to handle not only semantic understanding but also rich vocal characteristics like tone, style, dialect, emotion, and prosody. The design moves beyond traditional separate-component pipelines (ASR → text model → TTS), instead offering a multimodal model that ingests speech or audio and produces speech accordingly, enabling natural dialogue, voice cloning, and expressive speech synthesis. Through its architecture, Step-Audio supports multilingual interaction, dialects, emotional tones (joy, sadness, etc.), and even more creative speech styles (like rap or singing), while allowing dynamic control over speech characteristics. It also provides a “generative data engine,” which can produce synthetic speech data (cloning voices, varying style) to support TTS training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Step-Audio 2

    Step-Audio 2

    Multi-modal large language model designed for audio understanding

    Step-Audio2 is an advanced, end-to-end multimodal large language model designed for high-fidelity audio understanding and natural speech conversation: unlike many pipelines that separate speech recognition, processing, and synthesis, Step-Audio2 processes raw audio, reasons about semantic and paralinguistic content (like emotion, speaker characteristics, non-verbal cues), and can generate contextually appropriate responses — including potentially generating or transforming audio output. It integrates a latent-space audio encoder, discrete acoustic tokens, and reinforcement-learning–based training (CoT + RL) to enhance its ability to capture and reproduce voice styles, intonations, and subtle vocal cues. Moreover, Step-Audio2 supports tool-calling and retrieval-augmented generation (RAG), allowing it to access external knowledge sources or audio/text databases, thus reducing hallucinations and improving coherence in complex dialogues.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Step1X-3D

    Step1X-3D

    High-Fidelity and Controllable Generation of Textured 3D Assets

    Step1X-3D is an open-source framework for generating high-fidelity textured 3D assets from scratch — both their geometry and surface textures — using modern generative AI techniques. It combines a hybrid architecture: a geometry generation stage using a VAE-DiT model to output a watertight 3D representation (e.g. TSDF surface), and a texture synthesis stage that conditions on geometry and optionally reference input (or prompts) to produce view-consistent textures using a diffusion-based texture module. The result is fully 3D assets — meshes + textures — which can be rendered from any viewpoint, textured consistently, and used in 3D applications. To achieve this, the project includes a massive curated dataset: among more than 5 million candidate 3D assets, it filters and standardizes to produce a high-quality 2 million–asset subset suitable for training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Step1X-Edit

    Step1X-Edit

    A SOTA open-source image editing model

    Step1X-Edit is a state-of-the-art open-source image editing model/framework that uses a multimodal large language model (LLM) together with a diffusion-based image decoder to let users edit images simply via natural-language instructions plus a reference image. You supply an existing image and a textual command — e.g. “add a ruby pendant on the girl’s neck” or “make the background a sunset over mountains” — and the model interprets the instruction, computes a latent embedding combining the image content and user intent, then decodes a new image implementing the edit. The model targets general-purpose editing: from object addition/removal, style changes, recoloring, retouching, background replacement, to complex transformations like changing lighting, mood, or art style. The authors trained it on a large curated dataset and benchmarked it on a newly introduced evaluation suite, showing that Step1X-Edit significantly outperforms previous open-source baselines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    The purpose of this program is to prove that given a finite number of Monkeys and a finite amount of time one monkey will be able to make more money off the stock market than any human being. Checkout our web site for an uncorrupted download.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB