Open Source Python Artificial Intelligence Software - Page 50

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    LSTMs for Human Activity Recognition

    LSTMs for Human Activity Recognition

    Human Activity Recognition example using TensorFlow on smartphone

    LSTM-Human-Activity-Recognition is a machine learning project that demonstrates how recurrent neural networks can be used to recognize human activities from sensor data. The repository implements a deep learning model based on Long Short-Term Memory (LSTM) networks to classify physical activities using time-series data collected from wearable sensors. The project uses the well-known Human Activity Recognition dataset derived from smartphone accelerometer and gyroscope signals. Through the use of sequential neural network architectures, the system learns patterns in motion data that correspond to activities such as walking, sitting, standing, or climbing stairs. The repository includes data preprocessing scripts, neural network architecture definitions, and training pipelines that allow researchers to reproduce and modify the experiments. It serves as an educational example of how deep learning models can process temporal sensor signals for pattern recognition tasks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    LangChain Extract

    LangChain Extract

    Did you say you like data?

    LangChain Extract is an open-source reference application designed to demonstrate how large language models can be used to extract structured data from unstructured text and document files. The project implements a lightweight web service that allows developers to define extraction schemas and apply them to various sources such as plain text, HTML, or PDF documents. Built using FastAPI and the LangChain framework, the application exposes a REST API that can process documents and return structured outputs that match user-defined JSON schemas. Developers can create reusable “extractors” that define what type of information should be pulled from a document, along with example prompts that improve extraction quality through in-context learning.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    LangChain MCP

    LangChain MCP

    Model Context Protocol tool support for LangChain

    langchain-mcp provides Model Context Protocol (MCP) tool support for LangChain, a framework for developing applications powered by language models. It allows developers to create an MCPToolkit with a client.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    LatentSync

    LatentSync

    Taming Stable Diffusion for Lip Sync

    LatentSync is an open-source framework from ByteDance that produces high-quality lip-synchronization for video by using an audio-conditioned latent diffusion model, bypassing traditional intermediate motion representations. In effect, given a source video (with masked or reference frames) and an audio track, LatentSync directly generates frames whose lip motions and expressions align with the audio, producing convincing talking-head or animated lip-sync output. The system leverages a U-Net diffusion backbone, with cross-attention of audio embeddings (via an audio encoder) and reference video frames to guide generation, and applies a set of loss functions (temporal, perceptual, sync-net based) to enforce lip-sync accuracy, visual fidelity, and temporal consistency. Over versions, LatentSync has improved temporal stability and lowered resource requirements — making inference more practical (e.g. 8 GB VRAM for earlier versions, somewhat higher for latest models).
    Downloads: 1 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    LeWorldModel

    LeWorldModel

    Official code base for LeWorldModel: Stable End-to-End Joint-Embedding

    LeWorldModel is a minimalist tiling window manager designed for the X11 windowing system, focusing on simplicity, performance, and efficient use of screen space. It provides automatic window tiling behavior, organizing application windows into structured layouts without requiring manual resizing or positioning. The project emphasizes a lightweight design, minimizing resource usage while maintaining responsiveness and stability. It is highly configurable through source code or configuration files, allowing users to tailor behavior, keybindings, and layouts to their preferences. le-wm is intended for users who prefer keyboard-driven workflows and a distraction-free desktop environment. Its architecture avoids unnecessary complexity, making it easy to understand, modify, and extend.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Lightly

    Lightly

    A python library for self-supervised learning on images

    A python library for self-supervised learning on images. We, at Lightly, are passionate engineers who want to make deep learning more efficient. That's why - together with our community - we want to popularize the use of self-supervised methods to understand and curate raw image data. Our solution can be applied before any data annotation step and the learned representations can be used to visualize and analyze datasets. This allows selecting the best core set of samples for model training through advanced filtering. We provide PyTorch, PyTorch Lightning and PyTorch Lightning distributed examples for each of the models to kickstart your project. Lightly requires Python 3.6+ but we recommend using Python 3.7+. We recommend installing Lightly in a Linux or OSX environment. With lightly, you can use the latest self-supervised learning methods in a modular way using the full power of PyTorch. Experiment with different backbones, models, and loss functions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Lightweight' GAN

    Lightweight' GAN

    Implementation of 'lightweight' GAN, proposed in ICLR 2021

    Implementation of 'lightweight' GAN proposed in ICLR 2021, in Pytorch. The main contribution of the paper is a skip-layer excitation in the generator, paired with autoencoding self-supervised learning in the discriminator. Quoting the one-line summary "converge on single gpu with few hours' training, on 1024 resolution sub-hundred images". Augmentation is essential for Lightweight GAN to work effectively in a low data setting. You can test and see how your images will be augmented before they pass into a neural network (if you use augmentation). The general recommendation is to use suitable augs for your data and as many as possible, then after some time of training disable the most destructive (for image) augs. You can turn on automatic mixed precision with one flag --amp. You should expect it to be 33% faster and save up to 40% memory. Aim is an open-source experiment tracker that logs your training runs, and enables a beautiful UI to compare them.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    Ling

    Ling

    Ling is a MoE LLM provided and open-sourced by InclusionAI

    Ling is a Mixture-of-Experts (MoE) large language model (LLM) provided and open-sourced by inclusionAI. The project offers different sizes (Ling-lite, Ling-plus) and emphasizes flexibility and efficiency: being able to scale, adapt expert activation, and perform across a range of natural language/reasoning tasks. Example scripts, inference pipelines, and documentation. The codebase includes inference, examples, models, documentation, and model download infrastructure. As more developers and researchers engage with the platform, we can expect rapid advancements and improvements, leading to even more sophisticated applications. Model inference and API code (e.g. integration with Transformers). This collaborative approach accelerates development and ensures that the models remain at the forefront of technology, addressing emerging challenges in various fields.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    LitterBox

    LitterBox

    A secure sandbox environment for malware developers and red teamers

    LitterBox is a controlled malware-analysis and payload-testing sandbox aimed at red teams who need to validate evasions and behaviors before deployment. It provides an isolated environment to exercise payloads against modern detection stacks, verify signatures and heuristics, and observe runtime characteristics without leaking binaries to third-party vendors. The README frames typical use cases: testing evasion, validating detections, analyzing behavior, and keeping sensitive tooling in-house. Repo metadata and author pages highlight an active security-tools ecosystem around the maintainer, with CI and pull-request activity suggesting ongoing development. The project positions itself as a safe proving ground to reduce surprises in the field while minimizing operational risk. For teams exploring MCP integrations, notes mention pairing with LLM agents for assisted analysis.
    Downloads: 1 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    LiveKit Agents

    LiveKit Agents

    Framework for building realtime multimodal voice AI agents apps

    LiveKit Agents is an open source framework designed for building realtime AI agents that can participate as programmable entities within communication sessions. It enables developers to create conversational and multimodal agents capable of processing voice, audio, and other inputs in realtime environments. These agents can join LiveKit rooms as participants and interact with users or systems through speech, text, and other modalities. LiveKit Agents provides libraries and tooling that allow developers to combine speech-to-text, large language models, and text-to-speech services to build interactive AI experiences. It is designed to run server-side and can integrate with various AI model providers and realtime APIs to support different application requirements. LiveKit Agents also includes tools for scheduling and managing agent tasks, making it easier to connect users to automated assistants in live communication scenarios.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    LongCat-Image

    LongCat-Image

    Foundation model for image generation

    LongCat-Image is an open-source foundation model for image generation and editing created by the LongCat team at Meituan, designed to deliver high-quality visual outputs while remaining efficient and accessible for developers and researchers. Rather than relying on massive parameter counts typical of many cutting-edge models, LongCat-Image achieves strong photorealism, stable structure, and accurate bilingual (Chinese and English) text rendering with a more compact ~6-billion parameter architecture, making it competitive with much larger alternatives despite its relatively lean design. The model excels at both text-to-image generation and instruction-guided image editing, offering users versatile capabilities for creative and practical tasks—whether generating art, mockups, or adjusting existing visuals with fine control.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    MCP Bridge

    MCP Bridge

    A middleware to provide an openAI compatible endpoint

    MCP-Bridge serves as a middleware that connects the OpenAI API with MCP tools, allowing developers to utilize MCP functionalities through the OpenAI API interface. It provides endpoints compatible with OpenAI, facilitating seamless integration and enabling the use of MCP tools without requiring explicit MCP support in clients. ​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    MCP Server OpenDAL

    MCP Server OpenDAL

    Model Context Protocol Server for Apache OpenDAL™

    Model Context Protocol Server for Apache OpenDAL™ is an MCP server implementation that provides access to various storage services via Apache OpenDAL. It enables seamless interactions with multiple storage backends through a unified interface. ​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    MCP UI

    MCP UI

    SDK for building interactive UI components over MCP for AI tools

    mcp-ui is a software development kit designed to bring interactive user interface capabilities to applications built on the Model Context Protocol (MCP). It enables developers to create rich, dynamic UI components that can be delivered from an MCP server and rendered seamlessly by a compatible client. Instead of returning only text responses, tools can provide structured UI resources such as HTML or remote-rendered components, allowing more engaging and functional interactions. mcp-ui introduces a standardized approach where tools and their associated interfaces are linked through metadata, enabling clients to automatically discover and display the correct UI. It includes both client-side and server-side SDKs, making it possible to define UI elements on the backend and handle user interactions on the frontend. It supports multiple programming environments, including TypeScript, Python, and Ruby, broadening its accessibility for developers.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    MLOps Course

    MLOps Course

    Learn how to design, develop, deploy and iterate on ML apps

    The MLOps Course by Goku Mohandas is an open-source curriculum that teaches how to combine machine learning with solid software engineering to build production-grade ML applications. It is structured around the full lifecycle: data pipelines, modeling, experiment tracking, deployment, testing, monitoring, and iteration. The repository itself contains configuration, code examples, and links to accompanying lessons hosted on the Made With ML site, which provide detailed narrative explanations and diagrams. Instead of focusing only on model training, the course emphasizes best practices like modular code design, CI/CD, containerization, reproducibility, and responsible ML (including monitoring and feedback loops). This makes it particularly valuable for engineers transitioning from “notebooks and prototypes” to real systems that must be robust, maintainable, and observable in production.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Machine Learning Financial Laboratory

    Machine Learning Financial Laboratory

    MlFinLab helps portfolio managers and traders

    MlFinLab is a comprehensive Python library designed to support the development of machine learning strategies in quantitative finance and algorithmic trading. The project provides a large collection of tools that implement techniques from academic research on financial machine learning. It covers the full lifecycle of developing data-driven trading strategies, including data preprocessing, feature engineering, labeling techniques, model training, and performance evaluation. Many of the algorithms implemented in the library are based on concepts introduced in advanced quantitative finance literature and peer-reviewed research. The library also includes tools for constructing specialized financial data structures, generating predictive features, and evaluating trading strategies through backtesting. Its architecture emphasizes reproducibility, robust testing, and well-documented code so that researchers and practitioners can reliably experiment with financial machine learning models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Machine Learning PyTorch Scikit-Learn

    Machine Learning PyTorch Scikit-Learn

    Code Repository for Machine Learning with PyTorch and Scikit-Learn

    Initially, this project started as the 4th edition of Python Machine Learning. However, after putting so much passion and hard work into the changes and new topics, we thought it deserved a new title. So, what’s new? There are many contents and additions, including the switch from TensorFlow to PyTorch, new chapters on graph neural networks and transformers, a new section on gradient boosting, and many more that I will detail in a separate blog post. For those who are interested in knowing what this book covers in general, I’d describe it as a comprehensive resource on the fundamental concepts of machine learning and deep learning. The first half of the book introduces readers to machine learning using scikit-learn, the defacto approach for working with tabular datasets. Then, the second half of this book focuses on deep learning, including applications to natural language processing and computer vision.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Made With ML

    Made With ML

    Learn how to develop, deploy and iterate on production-grade ML

    Made-With-ML is an open-source educational repository and course designed to teach developers how to build production-grade machine learning systems using modern MLOps practices. The project focuses on bridging the gap between experimental machine learning notebooks and real-world software systems that can be deployed, monitored, and maintained at scale. It provides structured lessons and practical code examples that demonstrate how to design machine learning workflows, manage datasets, train models, evaluate performance, and deploy inference services. The repository organizes these concepts into modular Python scripts that follow software engineering best practices such as testing, configuration management, logging, and version control. Through a combination of tutorials, notebooks, and production-ready scripts, the project demonstrates how machine learning applications should be developed as maintainable systems rather than isolated experiments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    MagicTime

    MagicTime

    Time-lapse Video Generation Models as Metamorphic Simulators

    This repository is the official implementation of MagicTime, a metamorphic video generation pipeline based on the given prompts. The main idea is to enhance the capacity of video generation models to accurately depict the real world through our proposed methods and dataset. Compared to general videos, metamorphic videos contain physical knowledge, long persistence, and strong variation, making them difficult to generate.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    MaiBot

    MaiBot

    Maimaibot, a (more focused) multi-platform intelligent agent

    MaiBot is an open-source conversational AI agent designed to participate in group chats and behave like a socially aware digital persona. The project focuses on creating a more human-like interactive experience by combining large language models with behavioral planning and contextual awareness. Instead of functioning as a traditional command-driven chatbot, the system attempts to simulate natural social participation within group conversations. It can generate responses that imitate human speech patterns, learn slang or expressions from chat participants, and adapt its conversational style based on previous interactions. The architecture includes a memory system that stores conversation history and contextual information, allowing the bot to recall previous events and maintain continuity in discussions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    MaskFormer

    MaskFormer

    Per-Pixel Classification is Not All You Need for Semantic Segmentation

    MaskFormer is a unified framework for image segmentation developed by Facebook Research, designed to bridge the gap between semantic, instance, and panoptic segmentation within a single architecture. Unlike traditional segmentation pipelines that treat these tasks separately, MaskFormer reformulates segmentation as a mask classification problem, enabling a consistent and efficient approach across multiple segmentation domains. Built on top of Detectron2, it supports a wide range of datasets including ADE20K, Cityscapes, COCO-Stuff, and Mapillary Vistas, and provides pretrained baselines for each. The model achieves strong performance and scalability while simplifying training and evaluation workflows. Its successor, Mask2Former, extends the same meta-architecture to achieve state-of-the-art results across all major segmentation benchmarks. MaskFormer’s modular design, dataset integration, and compatibility with existing Detectron2 models make it an essential research tool.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    MedGemma

    MedGemma

    Collection of Gemma 3 variants that are trained for performance

    MedGemma is a collection of specialized open-source AI models created by Google as part of its Health AI Developer Foundations initiative, built on the Gemma 3 family of transformer models and trained for medical text and image comprehension tasks that help accelerate the development of healthcare-focused AI applications. It includes multiple variants such as a 4 billion-parameter multimodal model that can process both medical images and text and a 27 billion-parameter text-only (and multimodal) model that offers deeper clinical reasoning and understanding at higher capacity, making it suitable for complex tasks like medical question answering, summarization of clinical notes, or generating reports from radiology images. The multimodal versions pair a SigLIP-based image encoder pre-trained on diverse de-identified medical imaging data.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    MegaParse

    MegaParse

    File Parser optimised for LLM Ingestion with no loss

    MegaParse is a file parser optimized for Large Language Model (LLM) ingestion, ensuring no loss of information. It efficiently parses various document formats, such as PDFs, DOCX, and PPTX, converting them into formats ideal for processing by LLMs. This tool is essential for applications that require accurate and comprehensive data extraction from diverse document types.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    MegaTTS 3

    MegaTTS 3

    Official PyTorch Implementation

    MegaTTS3 is an open-source text-to-speech (TTS) and voice-cloning system from ByteDance that aims to deliver high-quality, expressive speech synthesis, including zero-shot voice cloning of previously unseen speakers. Its backbone is a lightweight diffusion-transformer (on the order of ~0.45 B parameters), which enables efficient inference while still producing high-fidelity audio. Given a reference audio sample (and corresponding latent representation), MegaTTS3 can generate speech in the style and voice timbre of that speaker — useful for personalized TTS, voice-overs, dubbing, or multi-speaker applications. The system supports both Chinese and English (with code-switching), making it versatile across languages, and offers controls for accent strength, voice similarity, intelligibility vs. similarity tradeoffs, and other speech parameters to fine-tune output.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    MeloTTS

    MeloTTS

    High-quality multi-lingual text-to-speech library by MyShell.ai

    MeloTTS is an open-source text-to-speech (TTS) system that generates natural-sounding speech from text input. It utilizes advanced machine-learning models to produce high-quality audio outputs.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB