Open Source Python Artificial Intelligence Software - Page 75

Python Artificial Intelligence Software

View 13574 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    GELab-Zero

    GELab-Zero

    GUI Exploration Lab. One of the best GUI agent solutions

    GELab-Zero is an open-source “GUI Agent” framework aiming to automate interactions with graphical user interfaces (GUIs), combining both the agent model and all supporting infrastructure — including inference, input orchestration, and GUI automation logic — in a plug-and-play package that runs locally, without cloud dependencies. The idea is to let developers or users harness an AI agent that can simulate clicking, typing, reading UI elements, and interacting with apps in a human-like way via the GUI, which can enable tasks like automated testing, scriptable workflows, or even autonomous usage of GUI-based applications. Because GELab-Zero is fully open-source and doesn’t require external services, it offers privacy and control: everything runs locally under your control. The project provides a lightweight base model (4B parameters in its public release) that can run on modest hardware (depending on quantization), making it more accessible than many large-scale AI solutions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    GLM-130B

    GLM-130B

    GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)

    GLM-130B is an open bilingual (English and Chinese) dense language model with 130 billion parameters, released by the Tsinghua KEG Lab and collaborators as part of the General Language Model (GLM) series. It is designed for large-scale inference and supports both left-to-right generation and blank filling, making it versatile across NLP tasks. Trained on over 400 billion tokens (200B English, 200B Chinese), it achieves performance surpassing GPT-3 175B, OPT-175B, and BLOOM-176B on multiple benchmarks, while also showing significant improvements on Chinese datasets compared to other large models. The model supports efficient inference via INT8 and INT4 quantization, reducing hardware requirements from 8× A100 GPUs to as little as a single server with 4× RTX 3090s. Built on the SwissArmyTransformer (SAT) framework and compatible with DeepSpeed and FasterTransformer, it supports high-speed inference (up to 2.5× faster) and reproducible evaluation across 30+ benchmark tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    GPT Academic

    GPT Academic

    Research-oriented chatbot framework

    GPT Academic is a research-oriented chatbot framework designed to integrate large language models (LLMs) into academic workflows. It provides tools for structured document processing, citation management, and enhanced interaction with research papers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    GPT All Star

    GPT All Star

    AI-powered code generation tool for scratch development of web apps

    AI-powered code generation tool for scratch development of web applications with a team collaboration of autonomous AI agents. This is a research project, and its primary value is to explore the possibility of autonomous AI agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    GPT PILOT

    GPT PILOT

    The first real AI developer

    GPT PILOT is an open-source AI developer assistant designed to build full applications by collaborating with a human developer throughout the software lifecycle. Unlike simple autocomplete tools, it aims to function as a true AI engineer that can generate features, set up environments, debug code, and request feedback when necessary. The system works by asking clarifying questions, producing product requirements, and then implementing the application step by step while the user supervises. It powers the Pythagora VS Code extension and relies on coordinated AI agents that mimic roles in a real development workflow. GPT Pilot is intended to automate the majority of routine coding work while leaving strategic decisions and final review to the human developer. Overall, the project represents an ambitious attempt to move from AI coding assistance toward semi-autonomous software development.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    GPT-2 FR

    GPT-2 FR

    GPT-2 French demo | Démo française de GPT-2

    OpenAI GPT-2 model trained on four different datasets in French. Books in French, French film scripts, reports of parliamentary debates, Tweet by Emmanuel Macron, allowing to generate text. Tensorflow and gpt-2-simple are required in order to fine-tune GPT-2. Create an environment then install the two packages pip install tensorflow==1.14 gpt-2-simple. A script and a notebook are available in the src folder to fine-tune GPT-2 on your own datasets. The output of each workout, i.e. the folder checkpoint/run1, is to be put ingpt2-model/model1 model2 model3 etc. You can run the script deploy_cloudrun.shto deploy all your different models (into gpt2-model) at once. However, you must have already initialized the gcloud CLI tool (Cloud SDK).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    GPT-2 Output Dataset

    GPT-2 Output Dataset

    Dataset of GPT-2 outputs for research in detection, biases, and more

    The GPT-2 Output Dataset is a large collection of model-generated text, released by OpenAI alongside the GPT-2 research paper to study the behaviors and limitations of large language models. It contains 250,000 samples of GPT-2 outputs, generated with different sampling strategies such as top-k truncation, to highlight the diversity and quality of model completions. The dataset also includes corresponding human-written text for comparison, enabling researchers to explore methods for distinguishing machine-generated content from human-authored text. The repository provides scripts and metadata for working with the dataset, with the goal of supporting research in areas like detection, evaluation of text coherence, and analysis of generative models. While no active development is expected, the dataset remains a useful benchmark for tasks involving text classification, style analysis, and generative model evaluation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    GPT-Code UI

    GPT-Code UI

    An open source implementation of OpenAI's ChatGPT Code interpreter

    An open source implementation of OpenAI's ChatGPT Code interpreter. Simply ask the OpenAI model to do something and it will generate & execute the code for you. You can put a .env in the working directory to load the OPENAI_API_KEY environment variable. For Azure OpenAI Services, there are also other configurable variables like deployment name. See .env.azure-example for more information. Note that model selection on the UI is currently not supported for Azure OpenAI Services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    GPT-NeoX

    GPT-NeoX

    Implementation of model parallel autoregressive transformers on GPUs

    This repository records EleutherAI's library for training large-scale language models on GPUs. Our current framework is based on NVIDIA's Megatron Language Model and has been augmented with techniques from DeepSpeed as well as some novel optimizations. We aim to make this repo a centralized and accessible place to gather techniques for training large-scale autoregressive language models, and accelerate research into large-scale training. For those looking for a TPU-centric codebase, we recommend Mesh Transformer JAX. If you are not looking to train models with billions of parameters from scratch, this is likely the wrong library to use. For generic inference needs, we recommend you use the Hugging Face transformers library instead which supports GPT-NeoX models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    GPTCache

    GPTCache

    Semantic cache for LLMs. Fully integrated with LangChain

    ChatGPT and various large language models (LLMs) boast incredible versatility, enabling the development of a wide range of applications. However, as your application grows in popularity and encounters higher traffic levels, the expenses related to LLM API calls can become substantial. Additionally, LLM services might exhibit slow response times, especially when dealing with a significant number of requests. To tackle this challenge, we have created GPTCache, a project dedicated to building a semantic cache for storing LLM responses. This project is undergoing swift development, and as such, the API may be subject to change at any time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    GPU Puzzles

    GPU Puzzles

    Solve puzzles. Learn CUDA

    GPU Puzzles is an educational project designed to teach GPU programming concepts through interactive coding exercises and puzzles. Instead of presenting traditional lecture-style explanations, the project immerses learners directly in hands-on programming tasks that demonstrate how GPU computation works. The exercises are implemented using Python with the Numba CUDA interface, which allows Python code to compile into GPU kernels that run on CUDA-enabled hardware. By solving progressively more complex puzzles, learners gain a practical understanding of how parallel algorithms operate on graphics processing units. The project emphasizes experimentation and problem solving, encouraging learners to discover GPU programming techniques through trial and exploration. It can be run in cloud environments such as Google Colab, making it easy for beginners to start experimenting without configuring local GPU hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    GPflow

    GPflow

    Gaussian processes in TensorFlow

    GPflow is a package for building Gaussian process models in Python. It implements modern Gaussian process inference for composable kernels and likelihoods. GPflow builds on TensorFlow 2.4+ and TensorFlow Probability for running computations, which allows fast execution on GPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Galileo is a library for developing custom distributed genetic algorithms developed in Python. It provides a robust set of objects that can be used directly or as the basis of derived objects. Its modularity makes it easy to extend the functionality. The
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14

    Game of Turmites

    Conway's Game of Life and Turmites Combined!

    This really isn't a game. It's all very randomly generated, and there is no way for any user input. I'll consider putting some in later. I had been wanting to make the Game of Life for some time as well as make some kind of genetic algorithm based code. So, here is what I came up with. While this may just seem like simplify a graphical display of what boredom looks like... well, it really doesn't go much past that point. If you Don't know what Conway's game of life is: It's the Black (Or white, I may have changed them) cells that follow a simple set of instructions based on the state of its adjacent cells. Turmites are types of little Turing machines, following their own set of instructions. They will always move forward, but depending on the color of cell they are on and their current internal state, they will change directions. Requires Pygame.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Gamera is a framework for the creation of structured document analysis applications by domain experts. It combines a programming library with GUI tools for the training and interactive development of recognition systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    pygpr is a collection of algorithms that can be used to perform Gaussian process regression and global optimization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Gaze at the landscape is a wallpaper switcher that uses on-line sources of pictures to provide a delightful desktop environment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Gemma in PyTorch

    Gemma in PyTorch

    The official PyTorch implementation of Google's Gemma models

    gemma_pytorch provides the official PyTorch reference for running and fine-tuning Google’s Gemma family of open models. It includes model definitions, configuration files, and loading utilities for multiple parameter scales, enabling quick evaluation and downstream adaptation. The repository demonstrates text generation pipelines, tokenizer setup, quantization paths, and adapters for low-rank or parameter-efficient fine-tuning. Example notebooks walk through instruction tuning and evaluation so teams can benchmark and iterate rapidly. The code is organized to be legible and hackable, exposing attention blocks, positional encodings, and head configurations. With standard PyTorch abstractions, it integrates easily into existing training loops, loggers, and evaluation harnesses.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    GeneralAI

    GeneralAI

    Large-scale Self-supervised Pre-training Across Tasks, Languages, etc.

    Fundamental research to develop new architectures for foundation models and AI, focusing on modeling generality and capability, as well as training stability and efficiency.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Generative Models

    Generative Models

    Collection of generative models, e.g. GAN, VAE in Pytorch

    This project is a comprehensive open-source collection of implementations of various generative machine learning models designed to help researchers and developers experiment with deep generative techniques. The repository contains practical implementations of well-known architectures such as Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), Restricted Boltzmann Machines, and Helmholtz Machines, implemented primarily using modern deep learning frameworks like PyTorch and TensorFlow. These models are widely used in artificial intelligence to generate new data that resembles the training data, such as images, text, or other structured outputs. The repository serves as an educational and experimental environment where users can study how generative models work internally and replicate results from academic research papers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Genoss GPT

    Genoss GPT

    One API for all LLMs either Private or Public

    One line replacement for openAI ChatGPT & Embeddings powered by OSS models. Genoss is a pioneering open-source initiative that aims to offer a seamless alternative to OpenAI models such as GPT 3.5 & 4, using open-source models like GPT4ALL.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    GiantMIDI-Piano

    GiantMIDI-Piano

    Classical piano MIDI dataset

    GiantMIDI-Piano is a large-scale symbolic classical piano music dataset built by applying the piano_transcription system on a vast collection of piano performance recordings. The dataset contains thousands of piano works, spanning a large number of composers and styles, with each piece transcribed into high-precision MIDI files capturing note events, pedal usage, velocities, etc. It provides a resource for music information retrieval (MIR), symbolic music modeling, composer classification, music generation, analysis of classical piano repertoire, and data-driven research in musicology or AI-based composition. Because the dataset is machine-generated via an automated transcription pipeline, it offers consistency, scale, and accessibility that would be difficult to achieve manually — enabling researchers to work with large corpora of piano music without copyright restrictions on symbolic data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Gluon CV Toolkit

    Gluon CV Toolkit

    Gluon CV Toolkit

    GluonCV provides implementations of state-of-the-art (SOTA) deep learning algorithms in computer vision. It aims to help engineers, researchers, and students quickly prototype products, validate new ideas and learn computer vision. It features training scripts that reproduce SOTA results reported in latest papers, a large set of pre-trained models, carefully designed APIs and easy-to-understand implementations and community support. From fundamental image classification, object detection, semantic segmentation and pose estimation, to instance segmentation and video action recognition. The model zoo is the one-stop shopping center for many models you are expecting. GluonCV embraces a flexible development pattern while is super easy to optimize and deploy without retaining a heavyweight deep learning framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    GluonNLP

    GluonNLP

    NLP made easy

    GluonNLP is a toolkit that helps you solve NLP problems. It provides easy-to-use tools that helps you load the text data, process the text data, and train models. To facilitate both the engineers and researchers, we provide command-line-toolkits for downloading and processing the NLP datasets. Gluon NLP makes it easy to evaluate and train word embeddings. Here are examples to evaluate the pre-trained embeddings included in the Gluon NLP toolkit as well as example scripts for training embeddings on custom datasets. Fasttext models trained with the library of Facebook research are exported both in text and a binary format. Unlike the text format, the binary format preserves information about subword units and consequently supports the computation of word vectors for words unknown during training (and not included in the text format). Besides training new fastText embeddings with Gluon NLP it is also possible to load the binary format into a Block provided by the Gluon NLP toolkit.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    GluonTS

    GluonTS

    Probabilistic time series modeling in Python

    GluonTS is a Python package for probabilistic time series modeling, focusing on deep learning based models. GluonTS requires Python 3.6 or newer, and the easiest way to install it is via pip. We train a DeepAR-model and make predictions using the simple "airpassengers" dataset. The dataset consists of a single time-series, containing monthly international passengers between the years 1949 and 1960, a total of 144 values (12 years * 12 months). We split the dataset into train and test parts, by removing the last three years (36 months) from the train data. Thus, we will train a model on just the first nine years of data. Python has the notion of extras – dependencies that can be optionally installed to unlock certain features of a package. We make extensive use of optional dependencies in GluonTS to keep the amount of required dependencies minimal. To still allow users to opt-in to certain features, we expose many extra dependencies.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB