Open Source Python Artificial Intelligence Software - Page 81

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    LongBench

    LongBench

    LongBench v2 and LongBench (ACL 25'&24')

    LongBench is a comprehensive benchmark designed to evaluate the ability of large language models to understand and reason over very long textual contexts. Traditional language model benchmarks typically evaluate tasks involving relatively short inputs, which does not reflect many real-world applications such as analyzing large documents or entire code repositories. LongBench addresses this gap by providing datasets that require models to process and reason over long sequences of text across multiple tasks. The benchmark includes multiple categories such as single-document question answering, multi-document reasoning, summarization, long dialogue understanding, and code analysis. It supports bilingual evaluation in English and Chinese to assess multilingual capabilities across extended contexts. Newer versions of the benchmark introduce extremely long context windows ranging from thousands to millions of tokens, enabling researchers to test the limits of modern long-context models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    LongWriter

    LongWriter

    Unleashing 10,000+ Word Generation from Long Context LLMs

    LongWriter is an open-source framework and set of large language models designed to enable ultra-long text generation that can exceed 10,000 words while maintaining coherence and structure. Traditional large language models can process large inputs but often struggle to generate long outputs due to limitations in training data and alignment strategies. LongWriter addresses this challenge by introducing a specialized dataset and training approach that encourages models to produce longer responses. The system uses an agent-based pipeline called AgentWrite that decomposes large writing tasks into smaller subtasks, allowing the model to produce long documents section by section. Researchers also created the LongWriter-6k dataset containing thousands of examples with outputs ranging from a few thousand to tens of thousands of words.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3

    Lumi-HSP

    This is an AI language model that can predict Heart failure or stroke

    Using thsi AI model, you can predict the chances of heart stroke and heart failure. HIGLIGHTS : 1. Accuracy of this model is 95% 2. This model uses the powerful Machine Learning algorithm "GradientBoosting" for predicting the outcomes. 3. An easy to use model and accessible to everyone.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    MAE (Masked Autoencoders)

    MAE (Masked Autoencoders)

    PyTorch implementation of MAE

    MAE (Masked Autoencoders) is a self-supervised learning framework for visual representation learning using masked image modeling. It trains a Vision Transformer (ViT) by randomly masking a high percentage of image patches (typically 75%) and reconstructing the missing content from the remaining visible patches. This forces the model to learn semantic structure and global context without supervision. The encoder processes only the visible patches, while a lightweight decoder reconstructs the full image—making pretraining computationally efficient. After pretraining, the encoder serves as a powerful backbone for downstream tasks like image classification, segmentation, and detection, achieving top performance with minimal fine-tuning. The repository provides pretrained models, fine-tuning scripts, evaluation protocols, and visualization tools for reconstruction quality and learned features.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 5
    MAI-UI

    MAI-UI

    Real-World Centric Foundation GUI Agents

    MAI-UI is a cutting-edge open-source project that implements a family of foundation GUI (Graphical User Interface) agent models capable of interpreting natural language and performing real-world GUI navigation and control tasks across mobile and desktop environments. Developed by Tongyi-MAI (Alibaba’s research initiative), the MAI-UI models are multimodal agents trained to understand user instructions and corresponding screenshots, grounding those instructions to on-screen elements and generating sequences of GUI actions such as taps, swipes, text input, and system commands. Unlike traditional UI frameworks, MAI-UI emphasizes realistic deployment by supporting agent–user interaction (clarifying ambiguous instructions), integration with external tool APIs using MCP calls, and a device–cloud collaboration mechanism that dynamically routes computation to on-device or cloud models based on task state and privacy constraints.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    MARS5

    MARS5

    MARS5 speech model (TTS) from CAMB.AI

    MARS5-TTS is CAMB.AI’s open-source English speech model designed for high-quality text-to-speech and voice emulation. It uses a two-stage architecture that combines an autoregressive (AR) model with a non-autoregressive (NAR) model, giving it both expressiveness and speed. The model is built to handle prosodically challenging content such as sports commentary, anime dialogue, and other high-energy or highly varied speech patterns with realistic rhythm and intonation. To control speaker identity, MARS5 uses a short reference audio clip, typically between 2 and 12 seconds, from which it learns the voice characteristics. It supports two main inference modes: shallow clone, which is faster and only needs the reference audio, and deep clone, which additionally uses the transcript of the reference audio to increase similarity and naturalness at the cost of more computation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    MCP Neo4j

    MCP Neo4j

    Model Context Protocol with Neo4j

    An implementation of the Model Context Protocol with Neo4j, enabling natural language interactions with Neo4j databases and facilitating operations such as schema retrieval and Cypher query execution. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    MCP Shell Server

    MCP Shell Server

    Shell command execution server implementing the Model Context Protocol

    A secure shell command execution server implementing the Model Context Protocol (MCP), allowing remote execution of whitelisted shell commands with support for standard input. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    MCP Snowflake Server

    MCP Snowflake Server

    A Model Context Protocol (MCP) server implementation

    An MCP server implementation that facilitates database interactions with Snowflake, allowing execution of SQL queries and presentation of data insights as resources. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    MCP UI

    MCP UI

    SDK for building interactive UI components over MCP for AI tools

    mcp-ui is a software development kit designed to bring interactive user interface capabilities to applications built on the Model Context Protocol (MCP). It enables developers to create rich, dynamic UI components that can be delivered from an MCP server and rendered seamlessly by a compatible client. Instead of returning only text responses, tools can provide structured UI resources such as HTML or remote-rendered components, allowing more engaging and functional interactions. mcp-ui introduces a standardized approach where tools and their associated interfaces are linked through metadata, enabling clients to automatically discover and display the correct UI. It includes both client-side and server-side SDKs, making it possible to define UI elements on the backend and handle user interactions on the frontend. It supports multiple programming environments, including TypeScript, Python, and Ruby, broadening its accessibility for developers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    MCP ZoomEye

    MCP ZoomEye

    A Model Context Protocol server that provides network asset info

    The ZoomEye MCP Server is a Model Context Protocol server that provides network asset information based on query conditions, allowing Large Language Models to obtain data by querying ZoomEye using dorks and other search parameters. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    META is a decision making software which aims are to track computer attackers, computer attacks and to help investigators finding useful elements.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    MGIE

    MGIE

    Guiding Instruction-based Image Editing via Multimodal Large Language

    MGIE—Guiding Instruction-based Image Editing—demonstrates how a multimodal LLM can parse natural-language editing instructions and then drive image transformations accordingly. The project focuses on making edits explainable and controllable: the model interprets text guidance, reasons over image content, and outputs edits aligned with user intent. It’s positioned as an ICLR 2024 Spotlight work, with code and references that show how to connect language planning to concrete image operations. This bridges a gap between free-form prompts and precise edits by letting users describe “what” and “where” in everyday language. The repo includes instructions, examples, and links that situate MGIE within Apple’s broader line of multimodal research. For practitioners, MGIE provides a blueprint for text-to-edit systems that are more semantically grounded than naive prompt-only pipelines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    MING

    MING

    A large-scale model of medical consultation in Chinese

    MING is an open-source medical large language model designed for intelligent medical consultation and question answering in Chinese. The project focuses on building a healthcare-focused conversational system capable of responding to medical questions, analyzing case descriptions, and guiding diagnostic reasoning. It is trained using medical instruction tuning so that the model can understand patient symptoms and respond with structured explanations and clinical suggestions. One of its primary goals is to simulate a multi-round medical consultation process, allowing the system to ask follow-up questions before offering diagnostic recommendations. This interactive capability makes it suitable for conversational health applications, patient triage scenarios, and educational demonstrations. The model is built on transformer-based architectures using frameworks such as PyTorch and integrates with Hugging Face tooling for training and inference workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    A High-Order Multi-Variate Approximation Scheme for Arbitrary Data Sets, C implementation of the method described in http://web.mit.edu/qiqi/www/paper/interpolation.pdf, with Python and Fortran interfaces.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    ML Ferret

    ML Ferret

    Refer and Ground Anything Anywhere at Any Granularity

    Ferret is Apple’s end-to-end multimodal large language model designed specifically for flexible referring and grounding: it can understand references of any granularity (boxes, points, free-form regions) and then ground open-vocabulary descriptions back onto the image. The core idea is a hybrid region representation that mixes discrete coordinates with continuous visual features, so the model can fluidly handle “any-form” referring while maintaining precise spatial localization. The repo presents the vision-language pipeline, model assets, and paper resources that show how Ferret answers questions, follows instructions, and returns grounded outputs rather than just text. In practice, this enables tasks like “find that small red icon next to the chart and describe it” where both the linguistic reference and the visual region are ambiguous without fine spatial reasoning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    ML Retreat

    ML Retreat

    Machine Learning Journal for Intermediate to Advanced Topics

    ML Retreat is an open-source learning repository that serves as a structured journal documenting advanced topics in machine learning and artificial intelligence. The project compiles detailed notes, technical explanations, and curated resources that guide readers through complex concepts across modern AI research. Rather than functioning as a traditional tutorial series, the repository is organized as a learning journey that progressively explores increasingly advanced subjects. Topics include large language models, graph neural networks, mechanistic interpretability, transformer architectures, and emerging research areas such as quantum machine learning. The repository includes references to influential research papers, lectures, and educational content from well-known machine learning educators.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    ML workspace

    ML workspace

    All-in-one web-based IDE specialized for machine learning

    All-in-one web-based development environment for machine learning. The ML workspace is an all-in-one web-based IDE specialized for machine learning and data science. It is simple to deploy and gets you started within minutes to productively built ML solutions on your own machines. This workspace is the ultimate tool for developers preloaded with a variety of popular data science libraries (e.g., Tensorflow, PyTorch, Keras, Sklearn) and dev tools (e.g., Jupyter, VS Code, Tensorboard) perfectly configured, optimized, and integrated. Usable as remote kernel (Jupyter) or remote machine (VS Code) via SSH. Easy to deploy on Mac, Linux, and Windows via Docker. Jupyter, JupyterLab, and Visual Studio Code web-based IDEs.By default, the workspace container has no resource constraints and can use as much of a given resource as the host’s kernel scheduler allows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ML-NLP

    ML-NLP

    This project is a common knowledge point and code implementation

    ML-NLP is a large open-source repository that collects theoretical knowledge, practical explanations, and code examples related to machine learning, deep learning, and natural language processing. The project is designed primarily as a learning resource for algorithm engineers and students preparing for technical interviews in machine learning or NLP roles. It compiles important concepts that frequently appear in machine learning discussions, including neural network architectures, training methods, and common algorithmic techniques. The repository also includes example implementations and explanatory materials that help readers understand the mechanics behind machine learning and NLP algorithms. In addition to technical explanations, the project organizes content into topic areas such as deep learning fundamentals, natural language processing techniques, and algorithm engineering practices.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MLBox

    MLBox

    MLBox is a powerful Automated Machine Learning python library

    MLBox is a powerful Automated Machine Learning python library. Fast reading and distributed data preprocessing/cleaning/formatting. Highly robust feature selection and leak detection. Accurate hyper-parameter optimization in high-dimensional space. State-of-the-art predictive models for classification and regression (Deep Learning, Stacking, LightGBM,...) Prediction with model interpretation. MLBox has been developed and used by many active community members. Your help is very valuable to make it better for everyone.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    MLOps Course

    MLOps Course

    Learn how to design, develop, deploy and iterate on ML apps

    The MLOps Course by Goku Mohandas is an open-source curriculum that teaches how to combine machine learning with solid software engineering to build production-grade ML applications. It is structured around the full lifecycle: data pipelines, modeling, experiment tracking, deployment, testing, monitoring, and iteration. The repository itself contains configuration, code examples, and links to accompanying lessons hosted on the Made With ML site, which provide detailed narrative explanations and diagrams. Instead of focusing only on model training, the course emphasizes best practices like modular code design, CI/CD, containerization, reproducibility, and responsible ML (including monitoring and feedback loops). This makes it particularly valuable for engineers transitioning from “notebooks and prototypes” to real systems that must be robust, maintainable, and observable in production.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    MLPerf

    MLPerf

    Reference implementations of MLPerf™ training benchmarks

    This is a repository of reference implementations for the MLPerf training benchmarks. These implementations are valid as starting points for benchmark implementations but are not fully optimized and are not intended to be used for "real" performance measurements of software frameworks or hardware. Benchmarking the performance of training ML models on a wide variety of use cases, software, and hardware drives AI performance across the tech industry. The MLPerf Training working group draws on expertise in AI and the technology that powers AI from across the industry to design and create industry-standard benchmarks. Together, we create the reference implementations, rules, policies, and procedures to benchmark a wide variety of AI workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    MMAction2

    MMAction2

    OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark

    OpenMMLab's next generation video understanding toolbox and benchmark. MMAction2 is an open-source toolbox for video understanding based on PyTorch. It is a part of the OpenMMLab project. Modular design: We decompose a video understanding framework into different components. One can easily construct a customized video understanding framework by combining different modules. Support four major video understanding tasks: MMAction2 implements various algorithms for multiple video understanding tasks, including action recognition, action localization, Spatio-temporal action detection, and skeleton-based action detection. We support 27 different algorithms and 20 different datasets for the four major tasks. We provide detailed documentation and API reference, as well as unit tests. We support Multigrid on Kinetics400, achieve 76.07% Top-1 accuracy and accelerate training speed.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB