Open Source Python Artificial Intelligence Software - Page 34

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Elasticsearch MCP Server

    Elasticsearch MCP Server

    A Model Context Protocol (MCP) server implementation

    This MCP server implementation provides interaction capabilities with Elasticsearch and OpenSearch, enabling functionalities such as document searching, index analysis, and cluster management through a set of tools. ​
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Elia

    Elia

    Terminal-based LLM chat tool with multi-model and local support

    Elia is an open source terminal-based interface designed for interacting with large language models in a fast and efficient way. It runs entirely in the command line, offering a keyboard-driven experience that reduces the need for switching between apps. Users can chat with both proprietary models like ChatGPT and Claude, as well as local models such as Llama 3, Mistral, and Gemma. Elia stores conversations in a local SQLite database, making it easy to revisit past interactions. It supports flexible usage with inline and full-screen chat modes, along with simple configuration through a single file. Installation is straightforward via pipx, and users can customize themes, system prompts, and model settings. Elia is built for developers and power users who prefer a streamlined, terminal-first workflow for working with AI models.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    EvoAgentX

    EvoAgentX

    Self-evolving AI agent framework for automated workflows

    EvoAgentX is an open source framework for building, evaluating, and continuously improving LLM-based agents and multi-agent workflows. It moves beyond static pipelines by introducing a self-evolving system where agents are automatically generated, tested, and optimised through iterative feedback. Developers can define goals in natural language, while the framework handles workflow creation, execution, and refinement. Its modular architecture supports layered components for agents, workflows, evaluation, and evolution, enabling flexible experimentation and scaling. EvoAgentX integrates optimisation algorithms to refine prompts, tool usage, and workflow structures over time. This allows agents to adapt dynamically instead of relying on fixed logic. It is designed for researchers and developers who want to automate complex agent systems and improve performance through continuous learning cycles, reducing manual orchestration and enabling more efficient development.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    FLAML

    FLAML

    A fast library for AutoML and tuning

    FLAML is a lightweight Python library that finds accurate machine learning models automatically, efficiently and economically. It frees users from selecting learners and hyperparameters for each learner. For common machine learning tasks like classification and regression, it quickly finds quality models for user-provided data with low computational resources. It supports both classical machine learning models and deep neural networks. It is easy to customize or extend. Users can find their desired customizability from a smooth range: minimal customization (computational resource budget), medium customization (e.g., scikit-style learner, search space, and metric), or full customization (arbitrary training and evaluation code). It supports fast automatic tuning, capable of handling complex constraints/guidance/early stopping. FLAML is powered by a new, cost-effective hyperparameter optimization and learner selection method invented by Microsoft Research.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 5
    Fairseq

    Fairseq

    Facebook AI Research Sequence-to-Sequence Toolkit written in Python

    Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks. We provide reference implementations of various sequence modeling papers. Recent work by Microsoft and Google has shown that data parallel training can be made significantly more efficient by sharding the model parameters and optimizer state across data parallel workers. These ideas are encapsulated in the new FullyShardedDataParallel (FSDP) wrapper provided by fairscale. Fairseq can be extended through user-supplied plug-ins. Models define the neural network architecture and encapsulate all of the learnable parameters. Criterions compute the loss function given the model outputs and targets. Tasks store dictionaries and provide helpers for loading/iterating over Datasets, initializing the Model/Criterion and calculating the loss.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    FastMCP

    FastMCP

    The fast, Pythonic way to build Model Context Protocol servers

    FastMCP is a fast, Pythonic framework for building servers and clients using the Model Context Protocol (MCP). It abstracts away protocol complexity like serialization, validation, and error handling, letting developers focus entirely on their business logic. With simple decorators, you can expose Python functions as tools, resources, or prompts that AI agents can safely and efficiently use. FastMCP introduces clear abstractions—components, providers, and transforms—that make it easy to control what agents see and how they interact with your system. The framework is opinionated by design, ensuring best practices and protocol compliance are the default rather than an extra burden. Actively maintained and widely adopted, FastMCP powers a majority of MCP servers and has become the de facto standard for production-ready MCP applications.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Fedhf

    Fedhf

    A Flexible Federated Learning Simulator

    FedHF is a Python-based simulator for flexible, heterogeneous, and asynchronous federated learning research. It provides configurable resource models, supports asynchronous protocols, and accelerates experimentation.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    FullTClash

    FullTClash

    General proxy performance testing tool based on Clash using Telegram

    Back end part useClash project(It can also be called nowmihomo)The relevant code is used as the outing agent. The front end part uses Telegram API as the interactive interface, which needs to be used in conjunction with Telegram, that is, a Telegram robot (bot), FullTClash bot is a Telegram robot (hereinafter referred to as bot) carrying its test tasks.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    GLM-4.5V

    GLM-4.5V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.5V is the preceding iteration in the GLM-V series that laid much of the groundwork for general multimodal reasoning and vision-language understanding. It embodies the design philosophy of mixing visual and textual modalities into a unified model capable of general-purpose reasoning, content understanding, and generation, while already supporting a wide variety of tasks: from image captioning and visual question answering to content recognition, GUI-based agents, video understanding, and long-document interpretation. GLM-4.5V emerged from a training framework that leverages scalable reinforcement learning (with curriculum sampling) to boost performance across tasks ranging from STEM problem solving to long-context reasoning, giving it broad applicability beyond narrow benchmarks. When it was released, it achieved state-of-the-art results on a large collection of public multimodal benchmarks for open-source models.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 10
    GLM-4.6V

    GLM-4.6V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.6V represents the latest generation of the GLM-V family and marks a major step forward in multimodal AI by combining advanced vision-language understanding with native “tool-call” capabilities, long-context reasoning, and strong generalization across domains. Unlike many vision-language models that treat images and text separately or require intermediate conversions, GLM-4.6V allows inputs such as images, screenshots or document pages directly as part of its reasoning pipeline — and can output or act via tools seamlessly, bridging perception and execution. Its architecture supports a very large context window (on the order of 128K tokens during training), which lets it handle complex multimodal inputs like long documents, multi-page reports, or video transcripts, while maintaining coherence across extended content. In benchmarks and internal evaluations, GLM-4.6V achieves state-of-the-art (SoTA) performance among models of comparable parameter scale on multimodal reasoning.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    GPT Neo

    GPT Neo

    An implementation of model parallel GPT-2 and GPT-3-style models

    An implementation of model & data parallel GPT3-like models using the mesh-tensorflow library. If you're just here to play with our pre-trained models, we strongly recommend you try out the HuggingFace Transformer integration. Training and inference is officially supported on TPU and should work on GPU as well. This repository will be (mostly) archived as we move focus to our GPU-specific repo, GPT-NeoX. NB, while neo can technically run a training step at 200B+ parameters, it is very inefficient at those scales. This, as well as the fact that many GPUs became available to us, among other things, prompted us to move development over to GPT-NeoX. All evaluations were done using our evaluation harness. Some results for GPT-2 and GPT-3 are inconsistent with the values reported in the respective papers. We are currently looking into why, and would greatly appreciate feedback and further testing of our eval harness.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    GPTme

    GPTme

    Your agent in your terminal, equipped with local tools

    GPTMe is a personal AI chatbot designed for self-reflection, journaling, and productivity, using GPT models to generate personalized insights and responses.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Get Physics Done (GPD)

    Get Physics Done (GPD)

    The first open-source agentic AI physicist

    Get Physics Done (GPD) is an open-source project designed to accelerate scientific research in physics by leveraging modern computational tools and automation techniques. It aims to simplify the process of performing simulations, calculations, and experimental analysis by providing structured workflows that integrate computational physics methods with reproducible research practices. The project focuses on reducing the friction involved in setting up experiments, running simulations, and analyzing results, allowing researchers to focus more on scientific insight rather than infrastructure. It emphasizes automation and reproducibility, ensuring that experiments can be easily replicated and extended by other researchers. The framework is adaptable to different areas of physics, making it suitable for both theoretical and applied research scenarios.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    GitDiagram

    GitDiagram

    AI tool that converts GitHub repositories into interactive diagrams

    GitDiagram is an open source web application designed to help developers quickly understand the structure and architecture of GitHub repositories by automatically generating interactive diagrams. It analyzes repository metadata such as the file tree and project documentation to build a visual representation of how different components of a project relate to one another. It uses an AI-powered pipeline to interpret repository structure and transform that information into system design diagrams rendered with Mermaid visualization. These diagrams provide a high-level overview of a codebase, making it easier for developers to explore unfamiliar projects or understand large and complex repositories. Users can interact with the generated diagrams by clicking components to navigate directly to related files or directories within the repository. GitDiagram combines a modern web frontend with a backend service that processes repository data and generates diagrams dynamically.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Gym

    Gym

    Toolkit for developing and comparing reinforcement learning algorithms

    Gym by OpenAI is a toolkit for developing and comparing reinforcement learning algorithms. It supports teaching agents, everything from walking to playing games like Pong or Pinball. Open source interface to reinforce learning tasks. The gym library provides an easy-to-use suite of reinforcement learning tasks. Gym provides the environment, you provide the algorithm. You can write your agent using your existing numerical computation library, such as TensorFlow or Theano. It makes no assumptions about the structure of your agent, and is compatible with any numerical computation library, such as TensorFlow or Theano. The gym library is a collection of test problems — environments — that you can use to work out your reinforcement learning algorithms. These environments have a shared interface, allowing you to write general algorithms.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    Hiera

    Hiera

    A fast, powerful, and simple hierarchical vision transformer

    Hiera is a hierarchical vision transformer designed to be fast, simple, and strong across image and video recognition tasks. The core idea is to use straightforward hierarchical attention with a minimal set of architectural “bells and whistles,” achieving competitive or superior accuracy while being markedly faster at inference and often faster to train. The repository provides installation options (from source or Torch Hub), a model zoo with pre-trained checkpoints, and code for evaluation and fine-tuning on standard benchmarks. Documentation emphasizes that model weights may have separate licensing and that the code targets practical experimentation for both research and downstream tasks. Community discussions cover topics like dataset pretrains, integration in other frameworks, and comparisons with related implementations. Security and contribution guidelines follow Meta’s open-source practices, and activity shows ongoing interest and usage across the community.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    High-Level Training Utilities Pytorch

    High-Level Training Utilities Pytorch

    High-level training, data augmentation, and utilities for Pytorch

    Contains significant improvements, bug fixes, and additional support. Get it from the releases, or pull the master branch. This package provides a few things. A high-level module for Keras-like training with callbacks, constraints, and regularizers. Comprehensive data augmentation, transforms, sampling, and loading. Utility tensor and variable functions so you don't need numpy as often. Have any feature requests? Submit an issue! I'll make it happen. Specifically, any data augmentation, data loading, or sampling functions. ModuleTrainer. The ModuleTrainer class provides a high-level training interface that abstracts away the training loop while providing callbacks, constraints, initializers, regularizers, and more. You also have access to the standard evaluation and prediction functions. Torchsample provides a wide range of callbacks, generally mimicking the interface found in Keras.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    Hindsight

    Hindsight

    Hindsight: Agent Memory That Learns

    Hindsight is an advanced, open-source memory system for AI agents designed to enable long-term learning, reasoning, and consistency across interactions by treating memory as a first-class component of intelligence rather than a simple retrieval layer. It addresses one of the core limitations of modern AI agents, which is their inability to retain and meaningfully use past experiences over time, by introducing a structured, biomimetic memory architecture inspired by how human memory works. Instead of relying solely on vector similarity or basic retrieval techniques, Hindsight organizes information into distinct categories such as facts, experiences, beliefs, and observations, allowing agents to differentiate between raw data and inferred knowledge. The system operates through three core mechanisms—retain, recall, and reflect—which respectively handle storing information, retrieving relevant context, and generating new insights based on accumulated experience.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    HivisionIDPhoto

    HivisionIDPhoto

    HivisionIDPhotos: a lightweight and efficient AI ID photos tools

    HivisionIDPhotos is an open-source AI project designed to automatically generate professional ID photographs from ordinary portrait images. The system uses computer vision and machine learning models to detect faces, segment the subject from the background, and produce standardized identification photos suitable for official documents. It is designed as a lightweight tool that can perform inference offline and run efficiently on CPUs without requiring powerful GPUs. The software analyzes portrait images, performs background removal, aligns the face according to ID photo standards, and produces images in various official size formats. It also allows the generation of layout sheets such as six-inch photo arrangements for printing multiple ID photos on a single page. The project focuses on building a practical pipeline for automated ID photo production using AI-based segmentation and image processing techniques.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    HolyClaude

    HolyClaude

    AI coding workstation: Claude Code + web UI + 5 AI CLIs + headless

    HolyClaude is a developer-focused toolkit designed to enhance and extend the capabilities of Claude Code environments by providing structured prompts, utilities, and workflow enhancements for AI-assisted coding. The project centers around improving how developers interact with AI agents, enabling more efficient code generation, debugging, and task execution through optimized prompt engineering. It includes predefined templates and interaction patterns that guide the AI toward producing more accurate and context-aware responses. HolyClaude emphasizes productivity by reducing friction in iterative development cycles, allowing users to refine outputs quickly without repeatedly crafting instructions from scratch. The toolkit is modular in nature, enabling developers to adapt it to different coding scenarios or integrate it into their existing workflows. It also reflects broader trends in AI-assisted development, where prompt design becomes a critical factor in output quality.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Homemade Machine Learning

    Homemade Machine Learning

    Python examples of popular machine learning algorithms

    homemade-machine-learning is a repository by Oleksii Trekhleb containing Python implementations of classic machine-learning algorithms done “from scratch”, meaning you don’t rely heavily on high-level libraries but instead write the logic yourself to deepen understanding. Each algorithm is accompanied by mathematical explanations, visualizations (often via Jupyter notebooks), and interactive demos so you can tweak parameters, data, and observe outcomes in real time. The purpose is pedagogical: you’ll see linear regression, logistic regression, k-means clustering, neural nets, decision trees, etc., built in Python using fundamentals like NumPy and Matplotlib, not hidden behind API calls. It is well suited for learners who want to move beyond library usage to understand how algorithms operate internally—how cost functions, gradients, updates and predictions work.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    HunyuanVideo

    HunyuanVideo

    HunyuanVideo: A Systematic Framework For Large Video Generation Model

    HunyuanVideo is a cutting-edge framework designed for large-scale video generation, leveraging advanced AI techniques to synthesize videos from various inputs. It is implemented in PyTorch, providing pre-trained model weights and inference code for efficient deployment. The framework aims to push the boundaries of video generation quality, incorporating multiple innovative approaches to improve the realism and coherence of the generated content. Release of FP8 model weights to reduce GPU memory usage / improve efficiency. Parallel inference code to speed up sampling, utilities and tests included.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    Image GPT

    Image GPT

    Large-scale autoregressive pixel model for image generation by OpenAI

    Image-GPT is the official research code and models from OpenAI’s paper Generative Pretraining from Pixels. The project adapts GPT-2 to the image domain, showing that the same transformer architecture can model sequences of pixels without altering its fundamental structure. It provides scripts to download pretrained checkpoints of different model sizes (small, medium, large) trained on large-scale datasets and includes utilities for handling color quantization with a 9-bit palette. Researchers can use the code to sample new images, evaluate generative loss on datasets like ImageNet or CIFAR-10, and explore the impact of scaling on performance. While the repository is archived and provided as-is, it remains a valuable starting point for experimenting with autoregressive transformers applied directly to raw pixel data. By demonstrating GPT’s flexibility across modalities, Image-GPT influenced subsequent multimodal generative research.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    Image classification models for Keras

    Image classification models for Keras

    Keras code and weights files for popular deep learning models

    All architectures are compatible with both TensorFlow and Theano, and upon instantiation the models will be built according to the image dimension ordering set in your Keras configuration file at ~/.keras/keras.json. For instance, if you have set image_dim_ordering=tf, then any model loaded from this repository will get built according to the TensorFlow dimension ordering convention, "Width-Height-Depth". Pre-trained weights can be automatically loaded upon instantiation (weights='imagenet' argument in model constructor for all image models, weights='msd' for the music tagging model). Weights are automatically downloaded if necessary, and cached locally in ~/.keras/models/. This repository contains code for the following Keras models, VGG16, VGG19, ResNet50, Inception v3, and CRNN for music tagging.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    Infinity

    Infinity

    Low-latency REST API for serving text-embeddings

    Infinity is a high-throughput, low-latency REST API for serving vector embeddings, supporting all sentence-transformer models and frameworks. Infinity is developed under MIT License. Infinity powers inference behind Gradient.ai and other Embedding API providers.
    Downloads: 3 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB