Open Source Python Artificial Intelligence Software - Page 92

Python Artificial Intelligence Software

View 13574 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 1
    QUCAS-PDP Astronomical Image Framework

    QUCAS-PDP Astronomical Image Framework

    After 4/15/26 this project will be archived as 9 pipelines are conso

    A collection of four interconnected open‑source projects that explore the quantum nature of the universe – from the early cosmos to extreme astrophysical environments. 📸 Live Demo The application is deployed on Streamlit Cloud: Live App test now: https://huggingface.co/spaces/QCAUS/QCAUS QCI AstroEntangle Refiner – FDM soliton physics & image processing Magnetar QED Explorer – Magnetar fields, dark photons & vacuum QED Primordial Photon–DarkPhoton Entanglement – Von Neumann evolution in an expanding universe QCIS (Quantum Cosmology Integration Suite) – Quantum‑corrected cosmological perturbations 🔭 Overview These four projects form a complete computational framework for quantum‑inspired astrophysics. Together they enable: Image analysis of galaxy clusters (Abell, Bullet, etc.) using Fuzzy Dark Matter (FDM) soliton overlays. Simulation of magnetar magnetospheres with strong‑field QED, dark photon production
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Qodo Cover

    Qodo Cover

    AI tool that generates tests to improve code coverage quickly

    Qodo Cover is an open source developer tool designed to automate the creation of unit tests using generative AI, helping teams improve code coverage with minimal manual effort. It operates as a command-line interface and can also be integrated into continuous integration workflows, making it adaptable to different development environments. It analyzes an existing codebase, identifies gaps in test coverage, and generates new tests that target uncovered or weakly tested areas. It follows an iterative workflow where generated tests are executed, validated, and refined to ensure they contribute meaningful coverage improvements. Internally, Qodo Cover uses a modular architecture that includes components for prompt generation, AI interaction, coverage analysis, and test validation. It supports scanning entire repositories to automatically detect test files, gather relevant context, and extend test suites accordingly.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    QuantResearch

    QuantResearch

    Quantitative analysis, strategies and backtests

    QuantResearch is a large educational repository dedicated to quantitative finance, algorithmic trading, and financial machine learning research. The project contains numerous notebooks and research materials demonstrating quantitative analysis techniques used in financial markets. These include implementations of factor models, statistical arbitrage strategies, portfolio optimization methods, and reinforcement learning approaches to trading. The repository also explores financial modeling topics such as vector autoregression, Gaussian mixture models, and option pricing techniques. Many notebooks demonstrate backtesting pipelines that allow users to evaluate trading strategies using historical market data. The project integrates machine learning methods with traditional quantitative finance models, illustrating how statistical techniques can be applied to asset management and trading.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Quote2Image

    Quote2Image

    A Python library for turning text quotes into graphical images

    A Python library for turning text quotes into graphical images. Generate an image using RGB background and foreground. The package comes with a built-in GenerateColors function that generates a fg and bg color with the correct amount of luminosity and returns them in tuples. Generate an image using a custom background image. The package comes with a builtin GenerateColors function that generates a fg and bg color with the correct amount of luminosity and returns them in tuples. We can generate an image using a custom background image using the ImgObject that gives us alot of flexibility on how we want our background Image to be. You are allowed to use, modify, and distribute the module. You are allowed to distribute modified versions of the module, as long as you follow the terms of the license.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Add Two Lines of Code. Get Full APM. Icon
    Add Two Lines of Code. Get Full APM.

    AppSignal installs in minutes and auto-configures dashboards, alerts, and error tracking.

    Works out of the box for Rails, Django, Express, Phoenix, and more. Monitoring exceptions and performance in no time.
    Start Free
  • 5
    Qwen-Audio

    Qwen-Audio

    Chat & pretrained large audio language model proposed by Alibaba Cloud

    Qwen-Audio is a large audio-language model developed by Alibaba Cloud, built to accept various types of audio input (speech, natural sounds, music, singing) along with text input, and output text. There is also an instruction-tuned version called Qwen-Audio-Chat which supports conversational interaction (multi-round), audio + text input, creative tasks and reasoning over audio. It uses multi-task training over many different audio tasks (30+), and achieves strong multi-benchmarks performance without task-specific fine‐tuning. It includes features such as flexible multi-run chat, audio understanding/reasoning, music appreciation, and also tool usage (e.g. voice editing).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Qwen2-Audio

    Qwen2-Audio

    Repo of Qwen2-Audio chat & pretrained large audio language model

    Qwen2-Audio is a large audio-language model by Alibaba Cloud, part of the Qwen series. It is trained to accept various audio signal inputs (including speech, sounds, etc.) and perform both voice chat and audio analysis, producing textual responses. It supports two major modes: Voice Chat (interactive voice only input) and Audio Analysis (audio + text instructions), with both base and instruction-tuned models. It is evaluated on many benchmarks (speech recognition, translation, sound classification, emotion, etc.), and offers pretrained models (e.g. 7B) released via ModelScope and Hugging Face. Code & examples provided with Hugging Face transformers, and usage via AutoProcessor, model classes etc. High performance on many standard benchmarks: ASR, speech-emotion recognition, vocal sound classification, speech translation etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Qwen2.5-Math

    Qwen2.5-Math

    A series of math-specific large language models of our Qwen2 series

    Qwen2.5-Math is a series of mathematics-specialized large language models in the Qwen2 family, released by Alibaba’s QwenLM. It includes base models (1.5B / 7B / 72B parameters), instruction-tuned versions, and a reward model (RM) to improve alignment. Unlike its predecessor Qwen2-Math, Qwen2.5-Math supports both Chain-of-Thought (CoT) reasoning and Tool-Integrated Reasoning (TIR) for solving math problems, and works in both Chinese and English. It is optimized for solving mathematical benchmarks and exams; the 72B-Instruct model achieves state-of-the-art results among open source models on many English and Chinese math tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Qwen2.5-Omni

    Qwen2.5-Omni

    Capable of understanding text, audio, vision, video

    Qwen2.5-Omni is an end-to-end multimodal flagship model in the Qwen series by Alibaba Cloud, designed to process multiple modalities (text, images, audio, video) and generate responses both as text and natural speech in streaming real-time. It supports “Thinker-Talker” architecture, and introduces innovations for aligning modalities over time (for example synchronizing video/audio), robust speech generation, and low-VRAM/quantized versions to make usage more accessible. It holds state-of-the-art performance in many multimodal benchmarks, particularly spoken language understanding, audio reasoning, image/video understanding, etc. Very strong benchmark performance across modalities (audio understanding, speech recognition, image/video reasoning) and often outperforming or matching single-modality models at a similar scale. Real-time streaming responses, including natural speech synthesis (text-to-speech) and chunked inputs for low latency interaction.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Qwen3 Embedding

    Qwen3 Embedding

    Designed for text embedding and ranking tasks

    Qwen3-Embedding is a model series from the Qwen family designed specifically for text embedding and ranking tasks. It builds upon the Qwen3 base/dense models and offers several sizes (0.6B, 4B, 8B parameters), for both embedding and reranking, with high multilingual capability, long‐context understanding, and reasoning. It achieves state-of-the-art performance on benchmarks like MTEB (Multilingual Text Embedding Benchmark) and supports instruction-aware embedding (i.e. embedding task instructions along with queries) and flexible embedding/vector dimension definitions. It is meant for tasks such as text retrieval, classification, clustering, bitext mining, and code retrieval.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    Qwen3-Omni

    Qwen3-Omni

    Qwen3-omni is a natively end-to-end, omni-modal LLM

    Qwen3-Omni is a natively end-to-end multilingual omni-modal foundation model that processes text, images, audio, and video and delivers real-time streaming responses in text and natural speech. It uses a Thinker-Talker architecture with a Mixture-of-Experts (MoE) design, early text-first pretraining, and mixed multimodal training to support strong performance across all modalities without sacrificing text or image quality. The model supports 119 text languages, 19 speech input languages, and 10 speech output languages. It achieves state-of-the-art results: across 36 audio and audio-visual benchmarks, it hits open-source SOTA on 32 and overall SOTA on 22, outperforming or matching strong closed-source models such as Gemini-2.5 Pro and GPT-4o. To reduce latency, especially in audio/video streaming, Talker predicts discrete speech codecs via a multi-codebook scheme and replaces heavier diffusion approaches.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Qwen3-VL-Embedding

    Qwen3-VL-Embedding

    Multimodal embedding and reranking models built on Qwen3-VL

    Qwen3-VL-Embedding (with its companion Qwen3-VL-Reranker) is a state-of-the-art multimodal embedding and reranking model suite built on the open-sourced Qwen3-VL foundation, developed to handle diverse inputs including text, images, screenshots, and videos. The core embedding model maps such inputs into semantically rich vectors in a unified representation space, enabling similarity search, clustering, and cross-modal retrieval. The reranking model then precisely scores relevance between a given query and candidate documents, enhancing retrieval accuracy in complex multimodal tasks. Together, they support advanced information retrieval workflows such as image-text search, visual question answering (VQA), and video-text matching, while providing out-of-the-box support for more than 30 languages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    R-KV

    R-KV

    Redundancy-aware KV Cache Compression for Reasoning Models

    R-KV is an open-source research project that focuses on improving the efficiency of large language model inference through key-value cache compression techniques. Modern transformer models rely heavily on KV caches during autoregressive decoding, which store intermediate attention states to accelerate generation. However, these caches can consume large amounts of memory, especially in reasoning-oriented models with long context windows. R-KV introduces a method for compressing the KV cache during decoding, allowing models to maintain reasoning performance while reducing memory consumption and computational overhead. The approach focuses on identifying which attention heads and cache components are most important for maintaining reasoning quality, allowing less critical information to be compressed or discarded. This results in more efficient inference without significantly degrading model performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    R1-V

    R1-V

    Witness the aha moment of VLM with less than $3

    R1-V is an initiative aimed at enhancing the generalization capabilities of Vision-Language Models (VLMs) through Reinforcement Learning in Visual Reasoning (RLVR). The project focuses on building a comprehensive framework that emphasizes algorithm enhancement, efficiency optimization, and task diversity to achieve general vision-language intelligence and visual/GUI agents. The team's long-term goal is to contribute impactful open-source research in this domain.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Robotic Manipulator Development and Simulation Environment in Python and Blender. IMPORTANT: Development moved to github. http://github.com/ajnsit/r2d3
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    RA.Aid

    RA.Aid

    Develop software autonomously

    RA.Aid is an AI-powered assistant designed to enhance the efficiency of software development workflows. It integrates seamlessly with various development environments, providing intelligent code suggestions, automated documentation generation, and real-time error detection. By leveraging advanced machine learning models, RA.Aid aims to reduce development time and improve code quality.​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    RAG-Retrieval

    RAG-Retrieval

    Unify Efficient Fine-tuning of RAG Retrieval, including Embedding

    RAG-Retrieval is an open-source framework for building and training retrieval systems used in retrieval-augmented generation pipelines. Retrieval-augmented generation combines large language models with external knowledge retrieval to improve factual accuracy and domain-specific reasoning. This repository provides end-to-end infrastructure for training retrieval models, performing inference, and distilling embedding models for improved performance. It includes implementations of modern embedding architectures designed to map documents and queries into vector spaces for efficient similarity search. The framework also supports reranking models that refine retrieved results using large language models or lightweight transformer architectures. Additional training techniques such as preference-based supervised fine-tuning and embedding distillation are included to improve retrieval quality.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    RAGs

    RAGs

    Build ChatGPT over your data, all with natural language

    RAGs is an open-source application designed to simplify the creation of retrieval-augmented generation pipelines through an interactive interface. Built with Streamlit and powered by the LlamaIndex ecosystem, the tool allows users to construct AI assistants that answer questions using their own data sources. Instead of requiring extensive programming knowledge, the application allows users to configure and build a RAG system using natural language instructions. The system automatically generates pipeline configurations that control how documents are retrieved, processed, and summarized before being used by a language model to generate responses. Users can also inspect and adjust parameters such as the number of retrieved documents, summarization strategies, and query settings through a configuration interface. Once the pipeline is created, the system enables conversational queries over the connected data sources, effectively creating a personalized knowledge assistant.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    RAPTOR

    RAPTOR

    The official implementation of RAPTOR

    RAPTOR is a retrieval architecture designed to improve retrieval-augmented generation systems by organizing documents into hierarchical structures that enable more effective context retrieval. Traditional RAG systems typically retrieve small text chunks independently, which can limit a model’s ability to understand broader document context. RAPTOR addresses this limitation by recursively embedding, clustering, and summarizing documents to create a tree-structured hierarchy of information. Each level of the tree represents summaries at different levels of abstraction, allowing retrieval to operate at both detailed and high-level conceptual layers. During inference, the system can navigate this hierarchical representation to retrieve information that best matches the user’s query while preserving broader contextual understanding. This approach improves question-answering performance on complex tasks that require reasoning across long documents or multiple sources.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19

    RDRPOSTagger

    A Rule-based Part-of-Speech and Morphological Tagging Toolkit

    RDRPOSTagger is a robust, easy-to-use and language-independent rule-based toolkit for Part-of-Speech (POS) and morphological tagging. RDRPOSTagger obtains fast performance in both learning and tagging process. RDRPOSTagger also achieves a very competitive accuracy in comparison to the state-of-the-art results. RDRPOSTagger now supports pre-trained POS and morphological tagging models for Bulgarian, Czech, Dutch, English, French, German, Hindi, Italian, Portuguese, Spanish, Swedish, Thai and Vietnamese. Additionally, RDRPOSTagger supports the pre-trained Universal POS tagging models for 40 languages. See the full usage of RDRPOSTagger at: http://rdrpostagger.sourceforge.net/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    RL Baselines Zoo

    RL Baselines Zoo

    A collection of 100+ pre-trained RL agents using Stable Baselines

    RL Baselines Zoo is a comprehensive training framework and collection of pre-trained RL agents using Stable-Baselines3. It offers tools for training, tuning, and evaluating RL algorithms across many standard environments, including MuJoCo, Atari, and robotics simulations. Designed for reproducible RL research and benchmarking, it includes scripts, hyperparameter presets, and best practices for training robust agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    RL with PyTorch

    RL with PyTorch

    Clean, Robust, and Unified PyTorch implementation

    RL with PyTorch is a research-oriented repository that provides implementations of deep reinforcement learning algorithms using the PyTorch framework. The project focuses on helping developers and researchers understand reinforcement learning methods by providing clean and reproducible implementations of well-known algorithms. It includes code for popular deep reinforcement learning techniques such as Deep Q-Networks, policy gradient methods, actor-critic architectures, and other modern RL approaches. The repository is structured so that users can easily experiment with different algorithms and training environments. Many examples demonstrate how agents learn to interact with simulated environments through trial and error using reinforcement learning principles. The codebase emphasizes clarity and modular design so that researchers can extend the implementations or use them for experimentation and benchmarking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    RL-Glue has moved to google code. Please see: http://code.google.com/p/rl-glue/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    RLCard

    RLCard

    Reinforcement Learning / AI Bots in Card (Poker) Games

    RLCard is a toolkit for reinforcement learning research on card games. It includes several popular card games and focuses on learning algorithms for imperfect information games like poker and blackjack.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    RLHF-Reward-Modeling

    RLHF-Reward-Modeling

    Recipes to train reward model for RLHF

    RLHF-Reward-Modeling is an open-source research framework focused on training reward models used in reinforcement learning from human feedback for large language models. In RLHF pipelines, reward models are responsible for evaluating generated responses and assigning scores that guide the model toward outputs that better match human preferences. The repository provides training recipes and implementations for building reward and preference models using modern machine learning frameworks. It supports multiple optimization strategies commonly used in alignment pipelines, including reinforcement learning with PPO, iterative supervised fine-tuning using rejection sampling, and direct preference optimization methods. The project also includes evaluation results showing that the trained reward models can achieve competitive performance compared with other open-source alignment systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    RNNLIB is a recurrent neural network library for sequence learning problems. Applicable to most types of spatiotemporal data, it has proven particularly effective for speech and handwriting recognition. full installation and usage instructions given at http://sourceforge.net/p/rnnl/wiki/Home/
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB