Open Source Python Artificial Intelligence Software - Page 83

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    This project provides a framework for testing and comparing different machine learning algorithms (particularly reinforcement learning methods) in different scenarios. Its intended area of application is in research and education.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Malware Analysis Network in Taiwan <Man in Taiwan, MiT> Welcome to contact us (TonTon@TWMAN.ORG) if you are interested in collaborating with us. This project is open source and distributed under the GNU General Public License version 3. Please feel free to add to or modify this source and propose changes or new converters. Developer & Copyrighted by : TonTon Hsien-De Huang Prompter: Jazz Yao-Tsung Wang, Figaro Chen-Ho Yang | Logo Desinger:Temaki Guo Community on Google+:http://X.TWMAN.ORG/Community/ SourceForge: https://sourceforge.net/projects/twmanplus/files/ FaceBook: https://www.facebook.com/TWMAN.PLUS
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    ManimML

    ManimML

    ManimML is a project focused on providing animations

    ManimML is a project focused on providing animations and visualizations of common machine-learning concepts with the Manim Community Library. Please check out our paper. We want this project to be a compilation of primitive visualizations that can be easily combined to create videos about complex machine-learning concepts. Additionally, we want to provide a set of abstractions that allow users to focus on explanations instead of software engineering.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Market Reporter

    Market Reporter

    Automatic Generation of Brief Summaries of Time-Series Data

    Market Reporter automatically generates short comments that describe time series data of stock prices, FX rates, etc. This is an implementation of Murakami et al. This tool stores data to Amazon S3. Ask the manager to give you AmazonS3FullAccess and issue a credential file. For details, please read AWS Identity and Access Management. Install Docker and Docker Compose. Edit envs/docker-compose.yaml according to your environment. Then, launch containers by docker-compose. We recommend to use pipenv to make a Python environment for this project. Suppose you have a database named master on your local machine. Prediction submodule generates a single comment of a financial instrument at specified time by loading a trained model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 5

    MarketSim

    A python based auction market simulator for agricultural trade

    The market assumes an environment in which farmers sell their produce through brokers and traders locate produce to buy through brokers. The major aim of the simulator is to experiment with various reputation mechanisms to manage bottlenecks and to model various adversarial scenarios. The market is aimed to simulate agricultural trade in developing countries. It is written in python and mysql database on Linux.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Mask2Former

    Mask2Former

    Code release for "Masked-attention Mask Transformer

    Mask2Former is a unified segmentation architecture that handles semantic, instance, and panoptic segmentation with one model and one training recipe. Its core idea is to cast segmentation as mask classification: a transformer decoder predicts a set of mask queries, each with an associated class score, eliminating the need for task-specific heads. A pixel decoder fuses multi-scale features and feeds masked attention in the transformer so each query focuses computation on its current spatial support. This leads to accurate masks with sharp boundaries and strong small-object performance while remaining efficient on high-resolution inputs. The project provides extensive configurations and pretrained models across popular benchmarks like COCO, ADE20K, and Cityscapes. Built on top of Detectron2, it includes training scripts, inference tools, and visualization utilities that make experimentation straightforward.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    A Maya plug-in for importing Massive simulations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    MatMul-Free LM

    MatMul-Free LM

    Implementation for MatMul-free LM

    MatMul-Free LM is an experimental implementation of a large language model architecture designed to eliminate traditional matrix multiplication operations used in transformer networks. Since matrix multiplication is one of the most computationally expensive components of modern language models, the project explores alternative computational strategies that reduce hardware requirements while maintaining comparable performance. The architecture relies on quantization-aware training and lightweight operations to replace conventional dense matrix multiplications with more efficient alternatives. These optimizations can significantly reduce memory consumption and potentially improve computational efficiency during both training and inference. The repository provides implementations of models at several parameter scales and includes tools for experimenting with the architecture using modern machine learning frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    MatchZoo

    MatchZoo

    Facilitating the design, comparison and sharing of deep text models

    The goal of MatchZoo is to provide a high-quality codebase for deep text matching research, such as document retrieval, question answering, conversational response ranking, and paraphrase identification. With the unified data processing pipeline, simplified model configuration and automatic hyper-parameters tunning features equipped, MatchZoo is flexible and easy to use. Preprocess your input data in three lines of code, keep track parameters to be passed into the model. Make use of MatchZoo customized loss functions and evaluation metrics. Initialize the model, fine-tune the hyper-parameters. Generate pair-wise training data on-the-fly, evaluate model performance using customized callbacks on validation data. MatchZoo is dependent on Keras and Tensorflow.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Matrix

    Matrix

    Multi-Agent daTa geneRation Infra and eXperimentation framework

    Matrix is a distributed, large-scale engine for multi-agent synthetic data generation and experiments: it provides the infrastructure to run thousands of “agentic” workflows concurrently (e.g. multiple LLMs interacting, reasoning, generating content, data-processing pipelines) by leveraging distributed computing (like Ray + cluster management). The idea is to treat data generation as a “data-to-data” transformation: each input item defines a task, and the runtime orchestrates asynchronous, peer-to-peer agent workflows, avoiding global synchronization bottlenecks. That design makes Matrix particularly well-suited for large-batch inference, model benchmarking, data curation, augmentation, or generation — whether for language, code, dialogue, or multimodal tasks. It supports both open-source LLMs and proprietary models (via integration with model backends), and works with containerized or sandboxed environments for safe tool execution or external code runs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MaxText

    MaxText

    A simple, performant and scalable Jax LLM

    MaxText is a high-performance, highly scalable open-source framework designed to train and fine-tune large language models using the JAX ecosystem. The project acts as both a reference implementation and a practical training library that demonstrates best practices for building and scaling transformer-based language models on modern accelerator hardware. It is optimized to run efficiently on Google Cloud TPUs and GPUs, enabling researchers and engineers to train models ranging from small experiments to extremely large distributed workloads. The framework focuses on simplicity while still supporting advanced techniques such as model sharding, distributed computation, and high-throughput training pipelines. MaxText includes ready-to-use configurations and reproducible training examples that help developers understand how to deploy large-scale AI workloads with modern machine learning infrastructure.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Mctx

    Mctx

    Monte Carlo tree search in JAX

    mctx is a Monte Carlo Tree Search (MCTS) library developed by Google DeepMind for reinforcement learning research. It enables efficient and flexible implementation of MCTS algorithms, including those used in AlphaZero and MuZero.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    MedgeClaw

    MedgeClaw

    Open-source AI research assistant for biomedicine

    MedgeClaw is a specialized AI-powered research assistant tailored for biomedical and scientific workflows, built on top of OpenClaw and Claude Code architectures. It integrates a large library of domain-specific skills, enabling it to perform complex analyses in areas such as genomics, drug discovery, and clinical research. The system connects conversational interfaces with computational environments, allowing users to initiate research tasks through messaging platforms while the backend executes analyses using tools like R and Python. It includes a real-time dashboard that displays progress, generated code, and outputs, providing transparency throughout the research process. MedgeClaw also supports reproducibility by generating structured reports and maintaining consistent environments through containerization. Its architecture combines conversational AI, automated pipelines, and scientific tooling into a unified workflow.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Medusa

    Medusa

    Framework for Accelerating LLM Generation with Multiple Decoding Heads

    Medusa is a framework aimed at accelerating the generation capabilities of Large Language Models (LLMs) by employing multiple decoding heads. This approach allows for parallel processing during text generation, significantly enhancing throughput and reducing response times. Medusa is designed to be simple to implement and integrates with existing LLM infrastructures, making it a practical solution for scaling LLM applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    Megatron is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Mellum-4b-base

    Mellum-4b-base

    JetBrains’ 4B parameter code model for completions

    Mellum-4b-base is JetBrains’ first open-source large language model designed and optimized for code-related tasks. Built with 4 billion parameters and a LLaMA-style architecture, it was trained on over 4.2 trillion tokens across multiple programming languages, including datasets such as The Stack, StarCoder, and CommitPack. With a context window of 8,192 tokens, it excels at code completion, fill-in-the-middle tasks, and intelligent code suggestions for professional developer tools and IDEs. The model is efficient for both cloud inference with vLLM and local deployment using llama.cpp or Ollama, thanks to its bf16 precision and AMP training. While the base model is not fine-tuned for downstream tasks, it is designed to be easily adapted through supervised fine-tuning (SFT) or reinforcement learning (RL). Benchmarks on RepoBench, SAFIM, and HumanEval demonstrate its competitive performance, with specialized fine-tuned versions for Python already showing strong improvements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    MemOS

    MemOS

    AI memory OS for LLM and Agent systems

    MemOS is an experimental operating system and runtime built around the concept of memory-centric computing, where memory objects are first-class citizens and program execution is organized around efficient, persistent memory access rather than traditional process and file system boundaries. The project explores rethinking system abstractions by tightly coupling computation with memory objects so that programs can operate on large datasets without expensive serialization or context switching. It aims to support advanced workflows like persistent in-memory data structures, crash-resilient state handling, and seamless sharing of data across tasks without copying. By abandoning some of the historical assumptions of Unix-style operating systems, MemOS attempts to unlock new performance and scalability tradeoffs for applications that need high throughput and low latency on memory-intensive workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Mental is a neural network engine implemented in pure Python. Design goals are easy-to-read source code, completeness, and ease of use. Presently implements feed-forward, back-propagation networks. Intended primarily for instructional use.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Mentat

    Mentat

    Mentat - The AI Coding Assistant

    Mentat is the AI tool that assists you with any coding task, right from your command line. Unlike Copilot, Mentat coordinates edits across multiple locations and files. And unlike ChatGPT, Mentat already has the context of your project, no copy and pasting is required. Run Mentat from within your project directory. Mentat uses Git, so if your project doesn't already have Git set up, run git init. List the files you would like Mentat to read and edit as arguments. Mentat will add each of them to context, so be careful not to exceed the GPT-4 token context limit.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Mesh R-CNN

    Mesh R-CNN

    code for Mesh R-CNN, ICCV 2019

    Mesh R-CNN is a 3D reconstruction and object understanding framework developed by Facebook Research that extends Mask R-CNN into the 3D domain. Built on top of Detectron2 and PyTorch3D, Mesh R-CNN enables end-to-end 3D mesh prediction directly from single RGB images. The model learns to detect, segment, and reconstruct detailed 3D mesh representations of objects in natural images, bridging the gap between 2D perception and 3D understanding. Unlike voxel-based or point-based approaches, Mesh R-CNN uses a differentiable mesh representation, allowing it to efficiently refine surface geometry while maintaining high spatial detail. The system combines 2D detection from Mask R-CNN with 3D reasoning modules that output full mesh reconstructions aligned with the input image. It has been evaluated on datasets such as Pix3D, where it demonstrates state-of-the-art performance in reconstructing real-world object geometry.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    MeshCNN in PyTorch

    MeshCNN in PyTorch

    Convolutional Neural Network for 3D meshes in PyTorch

    MeshCNN is a deep learning framework designed specifically for processing 3D triangular mesh data using convolutional neural networks. Unlike traditional CNNs that operate on images or voxel grids, MeshCNN performs convolution operations directly on the edges of mesh structures. This design allows the model to capture geometric relationships between mesh elements while preserving the underlying topology of 3D shapes. The framework introduces specialized layers such as edge-based convolution, mesh pooling, and mesh unpooling operations that enable hierarchical feature learning on mesh surfaces. These capabilities make the architecture well suited for tasks such as 3D object classification, segmentation, and geometric analysis. The project provides training pipelines, dataset preparation tools, and visualization utilities to support experiments with mesh-based neural networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Meshwork Analysis terminal

    Meshwork Analysis terminal

    Quantitative analytical Finance Terminal

    MWA Finance Terminal allows users to research and obtain data about various financial instruments and allows them to perform Statistical analysis with Various ML Technologies. It's divided into Modules and Options within them.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Meta Agents Research Environments (ARE)

    Meta Agents Research Environments (ARE)

    Meta Agents Research Environments is a comprehensive platform

    Meta Agents Research Environments (ARE) is a simulation and benchmarking platform. It is designed to evaluate AI agents in dynamic, evolving, multi-step tasks. Unlike static benchmarks, ARE supports environments where agents must adapt to changes over time and reason over sequences of actions. It interacts with applications and faces uncertainty. The included Gaia2 benchmark offers 800 scenarios across multiple “universes”. It can test reasoning, memory, tool use, and adaptability. Integration with simulated applications/agent APIs (email, file system, etc.). Support for multiple AI model backends/providers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Meta-World

    Meta-World

    Collections of robotics environments

    Meta-World is an open-source benchmark suite of robotic manipulation environments focused on multi-task and meta reinforcement learning. It provides a large collection of continuous-control tasks, such as reaching, pushing, opening doors, and manipulating objects with a simulated robot arm. The library defines standardized benchmarks like MT1, MT10, and MT50 for multi-task learning, where a single policy is trained across different numbers of tasks. It also offers meta-learning benchmarks (ML1, ML10, ML45) that evaluate few-shot adaptation to new goals or entirely new tasks. The environments adhere to the Gymnasium API, which makes them easy to plug into existing RL pipelines, and they support both synchronous and asynchronous vectorized execution for running many environments in parallel. Installation is done via pip, with official support for Python versions 3.8 through 3.11 on Linux and macOS, and the project is licensed under MIT to encourage broad academic and industry use.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    MetaCLIP

    MetaCLIP

    ICLR2024 Spotlight: curation/training code, metadata, distribution

    MetaCLIP is a research codebase that extends the CLIP framework into a meta-learning / continual learning regime, aiming to adapt CLIP-style models to new tasks or domains efficiently. The goal is to preserve CLIP’s strong zero-shot transfer capability while enabling fast adaptation to domain shifts or novel class sets with minimal data and without catastrophic forgetting. The repository provides training logic, adaptation strategies (e.g. prompt tuning, adapter modules), and evaluation across base and target domains to measure how well the model retains its general knowledge while specializing as needed. It includes utilities to fine-tune vision-language embeddings, compute prompt or adapter updates, and benchmark across transfer and retention metrics. MetaCLIP is especially suited for real-world settings where a model must continuously incorporate new visual categories or domains over time.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB