Open Source Python Software - Page 89

Python Software

Python Clear Filters

Browse free open source Python Software and projects below. Use the toggles on the left to filter open source Python Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Bot Framework SDK for Python

    Bot Framework SDK for Python

    Build and connect intelligent bots that interact naturally

    This repository contains code for the Python version of the Microsoft Bot Framework SDK, which is part of the Microsoft Bot Framework - a comprehensive framework for building enterprise-grade conversational AI experiences. This SDK enables developers to model conversation and build sophisticated bot applications using Python. SDKs for JavaScript and .NET are also available. The Microsoft Bot Framework provides what you need to build and connect intelligent bots that interact naturally wherever your users are talking, from text/sms to Skype, Slack, Office 365 mail and other popular services.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Chainer

    Chainer

    A flexible deep learning framework

    Chainer is a Python-based deep learning framework. It provides automatic differentiation APIs based on dynamic computational graphs as well as high-level APIs for neural networks.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    ChatGLM-6B

    ChatGLM-6B

    ChatGLM-6B: An Open Bilingual Dialogue Language Model

    ChatGLM-6B is an open bilingual (Chinese + English) conversational language model based on the GLM architecture, with approximately 6.2 billion parameters. The project provides inference code, demos (command line, web, API), quantization support for lower memory deployment, and tools for finetuning (e.g., via P-Tuning v2). It is optimized for dialogue and question answering with a balance between performance and deployability in consumer hardware settings. Support for quantized inference (INT4, INT8) to reduce GPU memory requirements. Automatic mode switching between precision/memory tradeoffs (full/quantized).
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    ChatTTS webUI & API

    ChatTTS webUI & API

    A simple native web interface that uses ChatTTS to synthesize text

    ChatTTS-ui is a local web interface and API wrapper around the ChatTTS speech synthesis system, designed to make advanced TTS models easy to use from a browser. It runs a small backend server (Python + Torch + ffmpeg) and exposes a simple webpage where you can type text, adjust parameters, and generate audio. The project supports Chinese, English, and mixed text with digits and control symbols, making it suitable for bilingual content and numerically heavy text like announcements or prompts. From version 0.96 onward, ffmpeg installation is required for deployment, and previous CSV/PT voice tables are no longer valid, so users instead work with updated “voice value” parameters. For convenience, there is a prepackaged Windows build: you download a release archive, extract it, and double-click app.exe to start the web UI, which opens on localhost:9966.
    Downloads: 3 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    Code2Prompt

    Code2Prompt

    Convert codebases into structured prompts optimized for LLM analysis

    code2prompt is an open source command line tool designed to convert an entire codebase into a structured prompt that can be easily used with large language models. It analyzes a project directory, gathers relevant source files, and formats them into a single prompt that includes the source tree and code content. This approach helps developers quickly provide full project context to AI models without manually copying files or assembling prompts. code2prompt is built in Rust and focuses on performance, enabling fast traversal of large repositories while maintaining low resource usage. It also respects common project conventions such as .gitignore, ensuring that unnecessary files are automatically excluded from the generated prompt. The generated output can be saved to a file, printed to standard output, or copied to the clipboard for immediate use. In addition to the core command line interface, the project also includes a library, Python bindings, and an MCP server.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    Colab-MCP

    Colab-MCP

    An MCP server for interacting with Google Colab

    Colab-MCP is an open-source Model Context Protocol server developed by Google that enables AI agents to directly interact with and control Google Colab environments programmatically, transforming Colab into a fully automated, agent-accessible workspace. Instead of relying on manual notebook usage, the system allows MCP-compatible agents to execute code, manage files, install dependencies, and orchestrate entire development workflows within Colab’s cloud infrastructure. This approach bridges the gap between local AI agents and remote high-performance compute environments, allowing users to offload heavy workloads such as machine learning training, data analysis, and dependency-heavy tasks to Colab’s GPU and TPU resources. By exposing Colab as an MCP server, the tool enables seamless integration with a wide range of AI assistants and agent frameworks, creating a standardized interface for tool use and execution.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    ComfyUI-Copilot

    ComfyUI-Copilot

    AI assistant for ComfyUI workflow generation, debugging, and tuning

    ComfyUI-Copilot is an AI-powered assistant designed to extend the capabilities of ComfyUI by simplifying and automating complex workflow development tasks. It functions as a custom node integrated directly into the ComfyUI environment, allowing users to interact with workflows through natural language and intelligent suggestions. ComfyUI-Copilot focuses on reducing the complexity of building node-based pipelines for generative AI tasks such as image generation, making it more accessible to both beginners and experienced users. It supports the entire workflow lifecycle, including generation, debugging, rewriting, and parameter optimization, helping users iterate more efficiently. ComfyUI-Copilot leverages large language model capabilities to analyze user intent, recommend nodes, and suggest models that match specific requirements. It also provides automated error detection and repair suggestions, improving reliability during development.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Cube Studio

    Cube Studio

    Cube Studio open source cloud native one-stop machine learning

    Cube Studio is an open-source, cloud-native end-to-end machine learning and AI platform designed to support the full lifecycle of AI development — from data preparation and interactive notebook coding to distributed training, model tuning, and deployment in production-ready environments. It provides a unified interface where teams can manage data sources, track datasets, and build pipelines using drag-and-drop workflow orchestration, making it accessible for both engineers and data scientists working at scale. The platform supports distributed training across multiple machines and GPUs, integrates tools for automated hyperparameter search and logging, and can serve models via inference services that include virtualized GPU support for efficient utilization.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    DeepClaude

    DeepClaude

    Unleash Next-Level AI

    DeepClaude is an open-source AI orchestration system that combines multiple state-of-the-art language models into a unified pipeline to achieve higher performance across tasks such as coding, reasoning, and content generation. It is built around the concept of model collaboration, where one model specializes in reasoning while another focuses on output refinement, resulting in more accurate and efficient responses. The system commonly pairs models such as DeepSeek R1 with Claude or Gemini, leveraging their complementary strengths to produce results that outperform individual models in benchmarks and real-world usage scenarios. DeepClaude is designed with compatibility in mind, supporting OpenAI-style APIs and allowing integration with various third-party model providers and routing services. It includes a graphical configuration interface and Docker-based deployment options, making it accessible to both developers and non-technical users who want to run advanced AI systems locally.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    Depth Pro

    Depth Pro

    Sharp Monocular Metric Depth in Less Than a Second

    Depth Pro is a foundation model for zero-shot metric monocular depth estimation, producing sharp, high-frequency depth maps with absolute scale from a single image. Unlike many prior approaches, it does not require camera intrinsics or extra metadata, yet still outputs metric depth suitable for downstream 3D tasks. Apple highlights both accuracy and speed: the model can synthesize a ~2.25-megapixel depth map in around 0.3 seconds on a standard GPU, enabling near real-time applications. The repo and research page emphasize boundary fidelity and crisp geometry, addressing a common weakness in monocular depth where edges can blur. Community integrations (e.g., inference wrappers and UI nodes) have sprung up around the model, reflecting practical interest in video, AR, and generative pipelines. As a general-purpose monocular depth backbone, Depth Pro slots into 3D reconstruction, relighting, and scene understanding workflows that benefit from metric predictions.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Diffusers

    Diffusers

    State-of-the-art diffusion models for image and audio generation

    Diffusers is the go-to library for state-of-the-art pretrained diffusion models for generating images, audio, and even 3D structures of molecules. Whether you're looking for a simple inference solution or training your own diffusion models, Diffusers is a modular toolbox that supports both. Our library is designed with a focus on usability over performance, simple over easy, and customizability over abstractions. State-of-the-art diffusion pipelines that can be run in inference with just a few lines of code. Interchangeable noise schedulers for different diffusion speeds and output quality. Pretrained models that can be used as building blocks, and combined with schedulers, for creating your own end-to-end diffusion systems. We recommend installing Diffusers in a virtual environment from PyPi or Conda. For more details about installing PyTorch and Flax, please refer to their official documentation.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    Django Rules

    Django Rules

    Awesome Django authorization, without the database

    rules is a tiny but powerful app providing object-level permissions to Django, without requiring a database. At its core, it is a generic framework for building rule-based systems, similar to decision trees. It can also be used as a standalone library in other contexts and frameworks. Versatile. Decorate callables to build complex graphs of predicates. Predicates can be any type of callable -- simple functions, lambdas, methods, callable class objects, partial functions, decorated functions, anything really. A good Django citizen. Seamless integration with Django views, templates and the Admin for testing for object-level permissions. Efficient and smart. No need to mess around with a database to figure out whether John really wrote that book. Simple. Dive in the code. You'll need 10 minutes to figure out how it works. Powerful. rules comes complete with advanced features, such as invocation context and storage for arbitrary data, skipping evaluation of predicates.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    EQGRP

    EQGRP

    Decrypted content of eqgrp-auction-file.tar.xz

    EQGRP is a public release of the so-called Equation Group hacking tools, originally leaked online in 2017. The repository serves as an archive and reference for security researchers, documenting the exploit frameworks, implants, and utilities that were allegedly used by a highly sophisticated threat actor. The tools include network exploitation scripts, backdoors, and frameworks targeting a range of platforms and services, many of which highlight previously unknown vulnerabilities. While the repository itself is provided for educational and research purposes, it also reflects a significant historical moment in cybersecurity, influencing both defensive strategies and awareness of advanced persistent threats. The release offers researchers insight into real-world offensive techniques, though many of the specific exploits are now outdated or patched. EQGRP remains a controversial but important resource for studying the evolution of nation-state-level cyber operations.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    Entity relation diagrams generator

    Entity relation diagrams generator

    Entity Relation Diagrams generation tool

    ERAlchemy generates Entity Relation (ER) diagram from databases or from SQLAlchemy models. The simplest way to install eralchemy on OSX is by using Homebrew. ERAlchemy requires GraphViz to generate the graphs and Python. Both are available for Windows, Mac and Linux. Thanks to it's modular architecture, it can be connected to other ORMs/ODMs/OGMs/O*Ms. Some tests require a local postgres database with a schema named test in a database named test all owned by a user named postgres with a password of postgres. ERAlchemy was inspired by erd, though it is able to render the ER diagram directly from the database and not just only from the ER markup language.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Evaluate

    Evaluate

    A library for easily evaluating machine learning models and datasets

    Evaluate is a library that makes evaluating and comparing models and reporting their performance easier and more standardized.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    Everywhere

    Everywhere

    Context-aware desktop AI assistant that understands screen content

    Everywhere is a context-aware desktop AI assistant designed to interact directly with the content displayed on a user’s screen. It distinguishes itself from traditional AI tools by eliminating the need for manual input methods such as copying text or taking screenshots, instead allowing users to invoke assistance instantly through a shortcut. It can analyze on-screen information in real time and provide contextual responses, making it useful for tasks like troubleshooting errors, summarizing articles, translating text, and refining written content. It integrates with multiple large language model providers and supports various tools, enabling flexible and extensible AI-powered workflows. Everywhere features a modern design with interactive elements such as markdown rendering, keyboard shortcuts, and voice input capabilities. Additionally, the project emphasizes seamless workflow integration by operating alongside existing applications rather than requiring users to switch.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    EvoAgentX

    EvoAgentX

    Self-evolving AI agent framework for automated workflows

    EvoAgentX is an open source framework for building, evaluating, and continuously improving LLM-based agents and multi-agent workflows. It moves beyond static pipelines by introducing a self-evolving system where agents are automatically generated, tested, and optimised through iterative feedback. Developers can define goals in natural language, while the framework handles workflow creation, execution, and refinement. Its modular architecture supports layered components for agents, workflows, evaluation, and evolution, enabling flexible experimentation and scaling. EvoAgentX integrates optimisation algorithms to refine prompts, tool usage, and workflow structures over time. This allows agents to adapt dynamically instead of relying on fixed logic. It is designed for researchers and developers who want to automate complex agent systems and improve performance through continuous learning cycles, reducing manual orchestration and enabling more efficient development.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    FastKoko

    FastKoko

    Dockerized FastAPI wrapper for Kokoro-82M text-to-speech model

    FastKoko is a self-hosted text-to-speech server built around the Kokoro-82M model and exposed through a FastAPI backend. It is designed to be easy to deploy via Docker, with separate CPU and GPU images so that users can choose between pure CPU inference and NVIDIA GPU acceleration. The project exposes an OpenAI-compatible speech endpoint, which means existing code that talks to the OpenAI audio API can often be pointed at a Kokoro-FastAPI instance with minimal changes. It supports multiple languages and voicepacks and allows phoneme based generation for more accurate pronunciation and prosody. The server also offers per-word timestamped captions, which makes it useful for creating subtitles or aligning audio with text. A built in web UI, API documentation, and debug endpoints for monitoring system status help users explore voices, test requests, and integrate the service into larger systems.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    FinalRecon

    FinalRecon

    All-in-one Python web reconnaissance tool for fast target analysis

    FinalRecon is an all-in-one web reconnaissance tool written in Python that helps security professionals gather information about a target website quickly and efficiently. It combines multiple reconnaissance techniques into a single command-line utility so users do not need to run several separate tools to collect similar data. FinalRecon focuses on providing a fast overview of a web target while maintaining accuracy in the collected results. It includes modules for gathering server information, analyzing SSL certificates, performing WHOIS lookups, and crawling website resources. FinalRecon can also enumerate DNS records, discover subdomains, search for directories and files, and scan common network ports. Historical URLs and resources can be retrieved from archived sources to help analyze changes in a website over time. Designed primarily for penetration testers and security researchers, FinalRecon simplifies the reconnaissance phase of security assessments.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    Flask-GraphQL

    Flask-GraphQL

    Adds GraphQL support to your Flask application

    Adds GraphQL support to your Flask application. This will add /graphql endpoint to your app and enable the GraphiQL IDE. If you are using the Schema type of Graphene library, be sure to use the graphql_schema attribute to pass as schema on the GraphQLView view. Otherwise, the GraphQLSchema from graphql-core is the way to go. The GraphQLSchema object that you want the view to execute when it gets a valid request. A value to pass as the context_value to graphql execute function. By default is set to dict with request object at key request. The root_value you want to provide to graphql execute.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    FuzzyWuzzy

    FuzzyWuzzy

    Fuzzy string matching in Python

    We’ve made it our mission to pull in event tickets from every corner of the internet, showing you them all on the same screen so you can compare them and get to your game/concert/show as quickly as possible. Of course, a big problem with most corners of the internet is labeling. One of our most consistently frustrating issues is trying to figure out whether two ticket listings are for the same real-life event (that is, without enlisting the help of our army of interns). To pick an example completely at random, Cirque du Soleil has a show running in New York called “Zarkana”. When we scour the web to find tickets for sale, mostly those tickets are identified by a title, date, time, and venue. We’ve built up a library of “fuzzy” string matching routines to help us along. And good news! We’re open sourcing it. The library is called “Fuzzywuzzy”, the code is pure python, and it depends only on the (excellent) difflib python library.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    GLM-4.6V

    GLM-4.6V

    GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning

    GLM-4.6V represents the latest generation of the GLM-V family and marks a major step forward in multimodal AI by combining advanced vision-language understanding with native “tool-call” capabilities, long-context reasoning, and strong generalization across domains. Unlike many vision-language models that treat images and text separately or require intermediate conversions, GLM-4.6V allows inputs such as images, screenshots or document pages directly as part of its reasoning pipeline — and can output or act via tools seamlessly, bridging perception and execution. Its architecture supports a very large context window (on the order of 128K tokens during training), which lets it handle complex multimodal inputs like long documents, multi-page reports, or video transcripts, while maintaining coherence across extended content. In benchmarks and internal evaluations, GLM-4.6V achieves state-of-the-art (SoTA) performance among models of comparable parameter scale on multimodal reasoning.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    GTFOBins

    GTFOBins

    GTFOBins is a curated list of Unix binaries

    GTFOBins is a curated catalog of Unix / POSIX system binaries and how they can be misused to bypass restrictions, escalate privileges, exfiltrate data, spawn shells, or otherwise act as “living off the land” tools in a compromised environment. It collects documented techniques for how everyday binaries (e.g. awk, bash, tar, scp) can be abused under constrained conditions. Indexed list of Unix binaries and documented misuse techniques. Examples of command invocations to exploit misconfigurations. Scenarios for privilege escalation, file transfer, and process spawning. Community contributions to add or refine binary techniques.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    Gemini Fullstack LangGraph Quickstart

    Gemini Fullstack LangGraph Quickstart

    Get started w/ building Fullstack Agents using Gemini 2.5 & LangGraph

    gemini-fullstack-langgraph-quickstart is a fullstack reference application from Google DeepMind’s Gemini team that demonstrates how to build a research-augmented conversational AI system using LangGraph and Google Gemini models. The project features a React (Vite) frontend and a LangGraph/FastAPI backend designed to work together seamlessly for real-time research and reasoning tasks. The backend agent dynamically generates search queries based on user input, retrieves information via the Google Search API, and performs reflective reasoning to identify knowledge gaps. It then iteratively refines its search until it produces a comprehensive, well-cited answer synthesized by the Gemini model. The repository provides both a browser-based chat interface and a command-line script (cli_research.py) for executing research queries directly. For production deployment, the backend integrates with Redis and PostgreSQL to manage persistent memory, streaming outputs, & background task coordination.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    Google Cloud Platform Python Samples

    Google Cloud Platform Python Samples

    Code samples used on cloud.google

    Google Cloud Platform Python Samples repository is a large, curated collection of Python code examples that demonstrate how to use a wide range of Google Cloud services in real-world scenarios. It serves as a practical companion to official documentation, providing runnable snippets that illustrate how to authenticate, configure environments, and interact with APIs across products such as storage, AI services, and data processing tools. The repository is organized into product-specific directories, allowing developers to quickly locate examples relevant to their use case and adapt them into production workflows. It emphasizes hands-on learning by guiding users through setup steps such as creating virtual environments, installing dependencies, and running scripts locally. These samples are designed to accelerate development by showing best practices for connecting services, handling data, and managing cloud resources programmatically.
    Downloads: 3 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB