Open Source Python Software - Page 51

Python Software

Python Clear Filters

Browse free open source Python Software and projects below. Use the toggles on the left to filter open source Python Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    AWS SAM CLI

    AWS SAM CLI

    CLI tool to build, test, debug, and deploy Serverless applications

    The AWS Serverless Application Model (SAM) CLI is an open-source CLI tool that helps you develop serverless applications containing Lambda functions, Step Functions, API Gateway, EventBridge, SQS, SNS and more. The AWS Serverless Application Model (SAM) is an open-source framework for building serverless applications. It provides shorthand syntax to express functions, APIs, databases, and event source mappings. With just a few lines per resource, you can define the application you want and model it using YAML. During deployment, SAM transforms and expands the SAM syntax into AWS CloudFormation syntax, enabling you to build serverless applications faster. To get started with building SAM-based applications, use the SAM CLI. SAM CLI provides a Lambda-like execution environment that lets you locally build, test, debug, and deploy AWS serverless applications.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 2
    AWS X-Ray SDK for Python

    AWS X-Ray SDK for Python

    AWS X-Ray SDK for the Python programming language

    AWS X-Ray SDK for the Python programming language. The AWS X-Ray SDK for Python is compatible with Python 2.7, 3.4, 3.5, 3.6, 3.7, 3.8, and 3.9. X-Ray Python SDK will by default generate no-op trace and entity id for unsampled requests and secure random trace and entity id for sampled requests. If customer wants to enable generating secure random trace and entity id for all the (sampled/unsampled) requests (this is applicable for trace id injection into logs use case) then they should set the AWS_XRAY_NOOP_ID environment variable as False. Oftentimes, it may be useful to be able to disable X-Ray for specific use cases, whether to stop X-Ray from sending traces at any moment or to test code functionality that originally depended on X-Ray instrumented packages to begin segments prior to the code call. For example, if your application relied on an XRayMiddleware to instrument incoming web requests, and you have a method that begins subsegments based on the segment generated.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    AWorld

    AWorld

    Build, evaluate and train General Multi-Agent Assistance with ease

    AWorld (Agent World) is an agent runtime/framework. It supports building, evaluating, and training self-improving intelligent agents and multi-agent systems (MAS). It is designed to provide infrastructure for agent orchestration, iterative learning, and environment interaction at scale. Scalable training across environments and distributed setups. Support for multi-agent collaboration/orchestration (MAS). The system is intended to help agents evolve via experience. It provides features to help and coordinate across multiple agents. It can also scale their training across environments.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 4
    Advanced Alchemy

    Advanced Alchemy

    A carefully crafted, thoroughly tested, optimized companion library

    advanced-alchemy is an opinionated ORM toolkit built on SQLAlchemy and designed for integration with the Litestar web framework. It simplifies common ORM patterns such as CRUD, pagination, and async support while providing extensibility and best practices out of the box. It’s designed to accelerate backend development for modern Python web apps.
    Downloads: 5 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    Agent Behavior Monitoring

    Agent Behavior Monitoring

    The open source post-building layer for agents

    Agent Behavior Monitoring is an open-source framework designed to monitor, evaluate, and improve the behavior of AI agents operating in real or simulated environments. The system focuses on agent behavior monitoring by collecting interaction data and analyzing how agents perform across different scenarios and tasks. Developers can use the framework to observe agent actions in both online production environments and offline evaluation settings, making it useful for debugging and performance analysis. Judgeval transforms agent interaction trajectories into structured evaluation datasets that can be used for reinforcement learning, supervised fine-tuning, or other forms of post-training improvement. The framework includes tools that analyze agent behavior patterns and group interaction trajectories by behavior type or topic, allowing researchers to detect weaknesses or unexpected behaviors.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 6
    AgentOps

    AgentOps

    Python SDK for agent monitoring, LLM cost tracking, benchmarking, etc.

    Industry-leading developer platform to test and debug AI agents. We built the tools so you don't have to. Visually track events such as LLM calls, tools, and multi-agent interactions. Rewind and replay agent runs with point-in-time precision. Keep a full data trail of logs, errors, and prompt injection attacks from prototype to production. Native integrations with the top agent frameworks.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    Alerta

    Alerta

    Alerta monitoring system

    Email was not designed to be used as an alert console. It is not a scalable solution when it comes to monitoring and alert visualization. A minimal installation of Alerta can be deployed quickly and easily as monitoring requirements and confidence grow. There are integrations available with Prometheus, Riemann, Nagios, Zabbix, netdata, Sensu, Pingdom and Cloudwatch. Integrating bespoke systems is easy using the API or command-line tool. Alerts are submitted in JSON format to an HTTP API. Alerts can be queried from the command line or viewed in a slick web console optimized for desktop, tablet, and mobile. User logins can be added using Google, GitHub or GitLab OAuth and programmatic access is managed using API keys.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    AlphaFold 3

    AlphaFold 3

    AlphaFold 3 inference pipeline

    AlphaFold 3, developed by Google DeepMind, is an advanced deep learning system for predicting biomolecular structures and interactions with exceptional accuracy. This repository provides the complete inference pipeline for running AlphaFold 3, though access to the model parameters is restricted and must be obtained directly from Google under specific terms of use. The system is designed for scientific research applications in structural biology, biochemistry, and bioinformatics, enabling accurate modeling of proteins, ligands, and covalent modifications. Users can perform local predictions via Docker containers, integrating AlphaFold 3’s inference process with provided JSON input configurations. The software includes flexible options for running both data preprocessing and GPU-accelerated inference, allowing users to adapt to available computational resources.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9
    Ansible Role: prometheus

    Ansible Role: prometheus

    Deploy Prometheus monitoring system

    Ansible-Prometheus is an Ansible role for automating the deployment and configuration of Prometheus monitoring systems.
    Downloads: 5 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    Arize Phoenix

    Arize Phoenix

    Uncover insights, surface problems, monitor, and fine tune your LLM

    Phoenix provides ML insights at lightning speed with zero-config observability for model drift, performance, and data quality. Phoenix is an Open Source ML Observability library designed for the Notebook. The toolset is designed to ingest model inference data for LLMs, CV, NLP and tabular datasets. It allows Data Scientists to quickly visualize their model data, monitor performance, track down issues & insights, and easily export to improve. Deep Learning Models (CV, LLM, and Generative) are an amazing technology that will power many of future ML use cases. A large set of these technologies are being deployed into businesses (the real world) in what we consider a production setting.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    ArkID

    ArkID

    Enterprise IDaaS/IAM platform system

    Rich plug-in, quickly builds an exclusive IDaaS/IAM platform. Easy integration into all your applications. Unified identity, certification, and authority management system. Extendable bottom application architecture based on Plug-in interpolation. You can flexibly and quickly add new functions to the main program without changing the main program. Achieve centralized and safe storage of corporate organizational structure and identity information of massive personnel. Establish a correspondence in multiple dimensions and securely integrate enterprise identity data sources. To achieve efficient and unified management of enterprise personnel, organizational structure, and application of information on a platform.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    Astron Agent

    Astron Agent

    Enterprise platform for building and orchestrating AI agent workflows

    Astron Agent is an enterprise-grade platform designed for building and managing intelligent AI agent workflows in production environments. It provides a development environment that combines workflow orchestration, model management, and integration with various AI tools and services. Astron Agent enables organizations to design complex agent-driven processes that coordinate models, automation tools, and enterprise systems. It also integrates robotic process automation capabilities so agents can execute tasks across digital systems instead of only generating responses. Astron Agent supports scalable and high-availability deployments, allowing teams to run reliable AI agent infrastructure in distributed environments. It includes collaboration features that allow teams to develop, manage, and operate AI applications together. With its extensible architecture and enterprise-focused design, it aims to help organizations build production-ready intelligent agent solutions.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    AutoGPT

    AutoGPT

    Powerful tool that lets you create and run intelligent agents

    AutoGPT is an experimental open-source application showcasing the capabilities of the GPT-4 language model. This program, driven by GPT-4, chains together LLM "thoughts", to autonomously achieve whatever goal you set. As one of the first examples of GPT-4 running fully autonomously, AutoGPT pushes the boundaries of what is possible with AI.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    AutoGPTQ

    AutoGPTQ

    An easy-to-use LLMs quantization package with user-friendly apis

    AutoGPTQ is an implementation of GPTQ (Quantized GPT) that optimizes large language models (LLMs) for faster inference by reducing their computational footprint while maintaining accuracy.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 15
    Bittensor

    Bittensor

    Internet-scale Neural Networks

    Bittensor is a decentralized machine learning protocol that allows AI models to collaborate, learn, and earn tokens within a global network. It introduces a blockchain-based economy for neural networks, where participants are incentivized to contribute valuable knowledge and compute power. Bittensor combines peer-to-peer learning with on-chain rewards, creating a self-governing, scalable AI system that evolves without centralized control. It is a novel approach to aligning incentives in AI development, empowering open contributions while preserving model ownership and decentralization.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 16
    Bootstrap Your Own Latent (BYOL)

    Bootstrap Your Own Latent (BYOL)

    Usable Implementation of "Bootstrap Your Own Latent" self-supervised

    Practical implementation of an astoundingly simple method for self-supervised learning that achieves a new state-of-the-art (surpassing SimCLR) without contrastive learning and having to designate negative pairs. This repository offers a module that one can easily wrap any image-based neural network (residual network, discriminator, policy network) to immediately start benefitting from unlabelled image data. There is now new evidence that batch normalization is key to making this technique work well. A new paper has successfully replaced batch norm with group norm + weight standardization, refuting that batch statistics are needed for BYOL to work. Simply plugin your neural network, specifying (1) the image dimensions as well as (2) the name (or index) of the hidden layer, whose output is used as the latent representation used for self-supervised training.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 17
    Brownie

    Brownie

    A Python-based development and testing framework for smart contracts

    Brownie is a Python-based development and testing framework for smart contracts targeting the Ethereum Virtual Machine. Powerful debugging tools, including python-style tracebacks and custom error strings. The recommended way to install Brownie is via pipx. pipx installs Brownie into a virtual environment and makes it available directly from the command-line. Once installed, you will never have to activate a virtual environment prior to using Brownie. Brownie documentation is hosted at Read the Docs. Use tox to run the complete suite against the full set of build targets, or pytest to run tests against a specific version of Python. If you are using pytest you must include the -p no:pytest-brownie flag to prevent it from loading the Brownie plugin.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    Browser Use MCP Server

    Browser Use MCP Server

    Browse the web, directly from Cursor etc.

    A browser automation server implementing the Model Context Protocol, designed to allow AI assistants to browse the web directly from applications like Cursor. It supports natural language commands for web navigation and interaction. ​
    Downloads: 5 This Week
    Last Update:
    See Project
  • 19
    CTGAN

    CTGAN

    Conditional GAN for generating synthetic tabular data

    CTGAN is a collection of Deep Learning based synthetic data generators for single table data, which are able to learn from real data and generate synthetic data with high fidelity. If you're just getting started with synthetic data, we recommend installing the SDV library which provides user-friendly APIs for accessing CTGAN. The SDV library provides wrappers for preprocessing your data as well as additional usability features like constraints. When using the CTGAN library directly, you may need to manually preprocess your data into the correct format, for example, continuous data must be represented as floats. Discrete data must be represented as ints or strings. The data should not contain any missing values.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 20
    CUDA Python

    CUDA Python

    Performance meets Productivity

    CUDA Python is a unified Python interface for accessing and working with the NVIDIA CUDA platform, enabling developers to build GPU-accelerated applications entirely in Python. It acts as a metapackage composed of multiple submodules that provide both high-level and low-level access to CUDA functionality, including runtime APIs, driver APIs, and JIT compilation tools. The project is designed to simplify GPU programming by offering Pythonic abstractions while still exposing the full power of CUDA for advanced users. It integrates tightly with the broader Python GPU ecosystem, including Numba for kernel compilation and CCCL for parallel primitives, allowing developers to write performant code without leaving Python. The toolkit also includes utilities for profiling, memory management, distributed computing, and numerical operations, making it suitable for scientific computing, AI, and data processing workloads.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 21
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing large models to run within the constraints of mobile hardware. It supports a wide range of AI tasks including text generation, speech-to-text, vision processing, and retrieval-augmented workflows through a unified API interface. A notable feature of Cactus is its hybrid execution model, which can dynamically route tasks between on-device processing and cloud services when additional compute is required.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    ChatterBot

    ChatterBot

    Machine learning, conversational dialog engine for creating chat bots

    ChatterBot is a Python library that makes it easy to generate automated responses to a user’s input. ChatterBot uses a selection of machine learning algorithms to produce different types of responses. This makes it easy for developers to create chat bots and automate conversations with users. For more details about the ideas and concepts behind ChatterBot see the process flow diagram. The language independent design of ChatterBot allows it to be trained to speak any language. Additionally, the machine-learning nature of ChatterBot allows an agent instance to improve it’s own knowledge of possible responses as it interacts with humans and other sources of informative data. An untrained instance of ChatterBot starts off with no knowledge of how to communicate. Each time a user enters a statement, the library saves the text that they entered and the text that the statement was in response to. As ChatterBot receives more input the number of responses that it can reply increase.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Classical Language Toolkit (CLTK)

    Classical Language Toolkit (CLTK)

    The Classical Language Toolkit

    The Classical Language Toolkit (CLTK) is a Python library offering natural language processing support for classical languages, including Latin, Greek, and others.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 24
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    CodeContests, developed by Google DeepMind, is a large-scale competitive programming dataset designed for training and evaluating machine learning models on code generation and problem solving. This dataset played a central role in the development of AlphaCode, DeepMind’s model for solving programming problems at a human-competitive level, as published in Science. CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder, Codeforces, CodeChef, Aizu, and HackerEarth. Each problem includes structured metadata, problem descriptions, paired input/output test cases, and multiple correct and incorrect solutions in various programming languages. The dataset is distributed in Riegeli format using Protocol Buffers, with separate training, validation, and test splits for reproducible machine learning experiments.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 25
    CodeLlama

    CodeLlama

    Inference code for CodeLlama models

    Code Llama is a family of Llama-based code models optimized for programming tasks such as code generation, completion, and repair, with variants specialized for base coding, Python, and instruction following. The repo documents the sizes and capabilities (e.g., 7B, 13B, 34B) and highlights features like infilling and large input context to support real IDE workflows. It targets both general software synthesis and language-specific productivity, offering strong performance among open models at release time. Typical usage includes prompt-driven generation, function or class completion, and zero-shot adherence to natural-language instructions about code changes. The ecosystem provides multiple distributions (e.g., HF format) so developers can integrate with standard toolchains and serving stacks. As part of the broader Llama effort, Code Llama complements instruction-tuned chat models by focusing on code-centric tasks and editor integrations.
    Downloads: 5 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB