Open Source Python Artificial Intelligence Software - Page 68

Python Artificial Intelligence Software

View 13566 business solutions

Browse free open source Python Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 1
    ConvNeXt V2

    ConvNeXt V2

    Code release for ConvNeXt V2 model

    ConvNeXt V2 is an evolution of the ConvNeXt architecture that co-designs convolutional networks alongside self-supervised learning. The V2 version introduces a fully convolutional masked autoencoder (FCMAE) framework where parts of the image are masked and the network reconstructs the missing content, marrying convolutional inductive bias with powerful pretraining. A key innovation is a new Global Response Normalization (GRN) layer added to the ConvNeXt backbone, which enhances feature competition across channels. The result is a convnet that competes strongly with transformer architectures on recognition benchmarks while being efficient and hardware-friendly. The repository provides official PyTorch implementations for multiple model sizes (Atto, Femto, Pico, up through Huge), conversion from JAX weights, code for pretraining/fine-tuning, and pretrained checkpoints. It supports both self-supervised pretraining and supervised fine-tuning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Conversational Health Agents (CHA)

    Conversational Health Agents (CHA)

    A Personalized LLM-powered Agent Frameworks

    CHA, or Conversational Health Agents, is an open-source framework designed to build intelligent healthcare assistants powered by large language models and external data sources. The system enables developers to create personalized AI agents that can interact with users through natural language while performing multi-step reasoning and task execution. It integrates orchestration capabilities that allow the agent to gather information from APIs, knowledge bases, and external services in order to generate more accurate and context-aware responses. The framework supports modular components such as planning, tool execution, and multimodal input processing, which makes it suitable for complex healthcare applications. It also includes a web-based interface for interacting with the agent, making it accessible for testing and deployment in real-world scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Convolution arithmetic

    Convolution arithmetic

    A technical report on convolution arithmetic in deep learning

    A technical report on convolution arithmetic in the context of deep learning. The code and the images of this tutorial are free to use as regulated by the licence and subject to proper attribution. The animations will be output to the gif directory. Individual animation steps will be output in PDF format to the pdf directory and in PNG format to the png directory. We introduce a guide to help deep learning practitioners understand and manipulate convolutional neural network architectures. The guide clarifies the relationship between various properties (input shape, kernel shape, zero padding, strides and output shape) of convolutional, pooling and transposed convolutional layers, as well as the relationship between convolutional and transposed convolutional layers. Relationships are derived for various cases, and are illustrated in order to make them intuitive.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    A lightweight and fast implementation of Conway\'s Game of Life and related cellular automata. It includes a pattern viewer running X Windows and a Python module intended to help in designing complex patterns.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Copter! is a SimCopter clone written in C++. Copter! will feature much more than the original SimCopter, and will be very moddable through it's modular design. Mods will be released which allow the player to be a policeman, or other characters.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Redundancy due to cut-paste operations in text creates bias in machine learning for NLP. This module takes a directory and produces a subset of the files in that directory (in a list) with an upper bound on similarity between two files.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Couler

    Couler

    Unified Interface for Constructing and Managing Workflows

    Couler is a system designed for unified machine learning workflow optimization in the cloud. Couler endeavors to provide a unified interface for constructing and optimizing workflows across various workflow engines, such as Argo Workflows, Tekton Pipelines, and Apache Airflow. Couler enhances workflow efficiency through features like Autonomous Workflow Construction, Automatic Artifact Caching Mechanisms, Big Workflow Auto Parallelism Optimization, and Automatic Hyperparameters Tuning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Courses (Anthropic)

    Courses (Anthropic)

    Anthropic's educational courses

    Anthropic’s courses repository is a growing collection of self-paced learning materials that teach practical AI skills using Claude and the Anthropic API. It’s organized as a sequence of hands-on courses—starting with API fundamentals and prompt engineering—so learners build capability step by step rather than in isolation. Each course mixes short readings with runnable notebooks and exercises, guiding you through concepts like model parameters, streaming, multimodal prompts, structured outputs, and evaluation. Assignments emphasize realistic tasks such as building small utilities, testing prompts against edge cases, and measuring quality so you learn to ship things that work. The materials are written for developers but remain friendly to newcomers, with clear setup instructions and minimal boilerplate. Because the repo is live and maintained, lessons are updated as the SDK and models evolve, and issues are used to track fixes, clarifications, and new modules.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Cradle framework

    Cradle framework

    The Cradle framework is a first attempt at General Computer Control

    Cradle is an open-source framework designed to enable AI agents to perform complex computer tasks by interacting with software environments in a way similar to human users. The system introduces the concept of General Computer Control, where AI agents receive screenshots as input and perform actions through simulated keyboard and mouse operations. This approach allows agents to interact with any software interface without relying on specialized APIs or predefined automation scripts. The framework integrates reasoning, planning, and memory modules that help the agent understand its environment and execute long sequences of actions. Cradle agents are capable of performing tasks across a wide variety of environments, including computer applications and video games, demonstrating the generality of the approach. The architecture includes modules that allow agents to observe their environment, reflect on past actions, plan future steps, and accumulate useful skills for later tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    CrypTen

    CrypTen

    A framework for Privacy Preserving Machine Learning

    CrypTen is a research framework developed by Facebook Research for privacy-preserving machine learning built directly on top of PyTorch. It provides a secure and intuitive environment for performing computations on encrypted data using Secure Multiparty Computation (SMPC). Designed to make secure computation accessible to machine learning practitioners, CrypTen introduces a CrypTensor object that behaves like a regular PyTorch tensor, allowing users to seamlessly apply automatic differentiation and neural network operations. Its design mirrors PyTorch’s modular and library-based structure, enabling flexible experimentation, debugging, and model development. The framework supports both encryption and decryption of tensors and operations such as addition and multiplication over encrypted values. Although not yet production-ready, CrypTen focuses on advancing real-world secure ML applications, such as training and inference over private datasets, without exposing sensitive data.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    CutLER

    CutLER

    Code release for Cut and Learn for Unsupervised Object Detection

    CutLER is an approach for unsupervised object detection and instance segmentation that trains detectors without human-annotated labels, and the repo also includes VideoCutLER for unsupervised video instance segmentation. The method follows a “Cut-and-LEaRn” recipe: bootstrap object proposals, refine them iteratively, and train detection/segmentation heads to discover objects across diverse datasets. The codebase provides training and inference scripts, model configs, and references to benchmarking results that report large gains over prior unsupervised baselines. It’s intended for researchers exploring self-supervised and unsupervised recognition, offering a practical path to scale beyond costly labeled corpora. The README links papers and gives a high-level overview of components and expected outputs, with pointers to demos and assets. The repository is actively starred and structured as a typical research release with license, contribution guidelines, and security policy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Cybersecurity AI

    Cybersecurity AI

    Cybersecurity AI (CAI), the framework for AI Security

    CAI (Cybersecurity AI) is a lightweight open-source framework intended to help security practitioners build and deploy AI-assisted automation for defensive and offensive security workflows. The project frames itself as a practical foundation for “AI security,” focusing on turning security tasks into agentic workflows that can be composed, executed, and iterated on by practitioners. Rather than being a single-purpose tool, CAI is positioned as a framework that supports building multiple security automations and integrating them into existing processes. It is designed for real-world usability, aiming to reduce friction for teams experimenting with AI agents in security operations, assessment, and response contexts. The framework emphasizes extensibility so users can connect models, tools, and supporting components depending on their environment and constraints.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    D2L.ai

    D2L.ai

    Interactive deep learning book with multi-framework code

    Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 300 universities from 55 countries including Stanford, MIT, Harvard, and Cambridge. This open-source book represents our attempt to make deep learning approachable, teaching you the concepts, the context, and the code. The entire book is drafted in Jupyter notebooks, seamlessly integrating exposition figures, math, and interactive examples with self-contained code. Offers sufficient technical depth to provide a starting point on the path to actually becoming an applied machine learning scientist.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    D4RL

    D4RL

    Collection of reference environments, offline reinforcement learning

    D4RL (Datasets for Deep Data-Driven Reinforcement Learning) is a benchmark suite focused on offline reinforcement learning — i.e., learning policies from fixed datasets rather than via online interaction with the environment. It contains standardized environments, tasks and datasets (observations, actions, rewards, terminals) aimed at enabling reproducible research in offline RL. Researchers can load a dataset for a given task (e.g., maze navigation, manipulation) and apply their algorithm without the need to collect fresh transitions, which accelerates experimentation and comparison. The API is based on Gymnasium (via gym.make) and each environment also exposes a method get_dataset() that returns the offline data to learn from. The repository emphasizes open science, reproducibility, and benchmarking at scale, making it easier to compare algorithms on equal footing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    DALL-E in Pytorch

    DALL-E in Pytorch

    Implementation / replication of DALL-E, OpenAI's Text to Image

    Implementation / replication of DALL-E (paper), OpenAI's Text to Image Transformer, in Pytorch. It will also contain CLIP for ranking the generations. Kobiso, a research engineer from Naver, has trained on the CUB200 dataset here, using full and deepspeed sparse attention. You can also skip the training of the VAE altogether, using the pretrained model released by OpenAI! The wrapper class should take care of downloading and caching the model for you auto-magically. You can also use the pretrained VAE offered by the authors of Taming Transformers! Currently only the VAE with a codebook size of 1024 is offered, with the hope that it may train a little faster than OpenAI's, which has a size of 8192. In contrast to OpenAI's VAE, it also has an extra layer of downsampling, so the image sequence length is 256 instead of 1024 (this will lead to a 16 reduction in training costs, when you do the math).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    DATAGEN

    DATAGEN

    AI-driven multi-agent research assistant automating hypothesis

    DATAGEN is an AI-driven multi-agent research and data analysis platform designed to automate complex analytical workflows. The system coordinates multiple specialized AI agents that collaborate to perform tasks such as hypothesis generation, data collection, analysis, visualization, and report creation. Instead of requiring users to manually orchestrate each stage of a research process, the platform allows these agents to coordinate automatically and handle the workflow end-to-end. The project integrates several modern AI frameworks including LangChain, LangGraph, and large language models to manage reasoning and data processing tasks. Through this architecture, the system can combine structured data analysis with natural language reasoning to generate insights and research outputs. The platform is designed for researchers, analysts, and developers who want to accelerate data exploration and automate parts of the research lifecycle.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    DC-TTS

    DC-TTS

    TensorFlow Implementation of DC-TTS: yet another text-to-speech model

    DC-TTS is a TensorFlow implementation of the DC-TTS architecture, a fully convolutional text-to-speech system designed to be efficiently trainable while producing natural speech. It follows the “Efficiently Trainable Text-to-Speech System Based on Deep Convolutional Networks with Guided Attention” paper, but the author adapts and extends the design to make it practical for real experiments. The model is split into two networks: Text2Mel, which maps text to mel-spectrograms, and SSRN (spectrogram super-resolution network), which converts low-resolution mel-spectrograms into high-resolution magnitude spectrograms suitable for waveform synthesis. Training scripts, data loaders, and hyperparameter configurations are provided to reproduce results on several datasets, including LJ Speech for English, a Korean single-speaker dataset, and audiobook data from Nick Offerman and Kate Winslet.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    DCGAN in TensorLayerX

    DCGAN in TensorLayerX

    The Simplest DCGAN Implementation

    This is an implementation of Deep Convolutional Generative Adversarial Networks. First, download the aligned face images from google or baidu to a data folder. Please place dataset 'img_align_celeba.zip' under 'data/celebA/' by default.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    DCVGAN

    DCVGAN

    DCVGAN: Depth Conditional Video Generation, ICIP 2019.

    This paper proposes a new GAN architecture for video generation with depth videos and color videos. The proposed model explicitly uses the information of depth in a video sequence as additional information for a GAN-based video generation scheme to make the model understands scene dynamics more accurately. The model uses pairs of color video and depth video for training and generates a video using the two steps. Generate the depth video to model the scene dynamics based on the geometrical information. To add appropriate color to the geometrical information of the scene, the domain translation from depth to color is performed for each image. This model has three networks in the generator. In addition, the model has two discriminators.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    DETR

    DETR

    End-to-end object detection with transformers

    PyTorch training code and pretrained models for DETR (DEtection TRansformer). We replace the full complex hand-crafted object detection pipeline with a Transformer, and match Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch. What it is. Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It consists of a set-based global loss, which forces unique predictions via bipartite matching, and a Transformer encoder-decoder architecture. Given a fixed small set of learned object queries, DETR reasons about the relations of the objects and the global image context to directly output the final set of predictions in parallel. Due to this parallel nature, DETR is very fast and efficient.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    DIG

    DIG

    A library for graph deep learning research

    The key difference with current graph deep learning libraries, such as PyTorch Geometric (PyG) and Deep Graph Library (DGL), is that, while PyG and DGL support basic graph deep learning operations, DIG provides a unified testbed for higher level, research-oriented graph deep learning tasks, such as graph generation, self-supervised learning, explainability, 3D graphs, and graph out-of-distribution. If you are working or plan to work on research in graph deep learning, DIG enables you to develop your own methods within our extensible framework, and compare with current baseline methods using common datasets and evaluation metrics without extra efforts. It includes unified implementations of data interfaces, common algorithms, and evaluation metrics for several advanced tasks. Our goal is to enable researchers to easily implement and benchmark algorithms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    DIGITS

    DIGITS

    Deep Learning GPU training system

    The NVIDIA Deep Learning GPU Training System (DIGITS) puts the power of deep learning into the hands of engineers and data scientists. DIGITS can be used to rapidly train the highly accurate deep neural network (DNNs) for image classification, segmentation and object detection tasks. DIGITS simplifies common deep learning tasks such as managing data, designing and training neural networks on multi-GPU systems, monitoring performance in real-time with advanced visualizations, and selecting the best performing model from the results browser for deployment. DIGITS is completely interactive so that data scientists can focus on designing and training networks rather than programming and debugging. DIGITS is available as a free download to the members of the NVIDIA Developer Program. DIGITS is available on NVIDIA GPU Cloud (NGC) as an optimized container for on-demand usage. Sign-up for an NGC account and get started with DIGITS in minutes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    DINOv2

    DINOv2

    PyTorch code and models for the DINOv2 self-supervised learning

    DINOv2 is a self-supervised vision learning framework that produces strong, general-purpose image representations without using human labels. It builds on the DINO idea of student–teacher distillation and adapts it to modern Vision Transformer backbones with a carefully tuned recipe for data augmentation, optimization, and multi-crop training. The core promise is that a single pretrained backbone can transfer well to many downstream tasks—from linear probing on classification to retrieval, detection, and segmentation—often requiring little or no fine-tuning. The repository includes code for training, evaluating, and feature extraction, with utilities to run k-NN or linear evaluation baselines to assess representation quality. Pretrained checkpoints cover multiple model sizes so practitioners can trade accuracy for speed and memory depending on their deployment constraints.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    DLRM

    DLRM

    An implementation of a deep learning recommendation model (DLRM)

    DLRM (Deep Learning Recommendation Model) is Meta’s open-source reference implementation for large-scale recommendation systems built to handle extremely high-dimensional sparse features and embedding tables. The architecture combines dense (MLP) and sparse (embedding) branches, then interacts features via dot product or feature interactions before passing through further dense layers to predict click-through, ranking scores, or conversion probabilities. The implementation is optimized for performance at scale, supporting multi-GPU and multi-node execution, quantization, embedding partitioning, and pipelined I/O to feed huge embeddings efficiently. It includes data loaders for standard benchmarks (like Criteo), training scripts, evaluation tools, and capabilities like mixed precision, gradient compression, and memory fusion to maximize throughput.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    DPM-Solver

    DPM-Solver

    Fast ODE Solver for Diffusion Probabilistic Model Sampling

    DPM-Solver is a machine learning research implementation focused on accelerating the sampling process in diffusion probabilistic models used for generative AI tasks. Diffusion models are powerful generative systems capable of producing high-quality images and other data, but traditional sampling methods often require hundreds or thousands of computational steps. The project introduces a specialized numerical solver designed to approximate the diffusion process using a small number of high-order integration steps. By reformulating the sampling problem as the solution of a diffusion-related ordinary differential equation, the solver can produce high-quality samples much more efficiently. This approach significantly reduces the computational cost required to generate images while maintaining strong generation quality.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB