Open Source Python Software - Page 90

Python Software

Python Clear Filters

Browse free open source Python Software and projects below. Use the toggles on the left to filter open source Python Software by OS, license, language, programming language, and project status.

  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Granite TSFM

    Granite TSFM

    Foundation Models for Time Series

    granite-tsfm collects public notebooks, utilities, and serving components for IBM’s Time Series Foundation Models (TSFM), giving practitioners a practical path from data prep to inference for forecasting and anomaly-detection use cases. The repository focuses on end-to-end workflows: loading data, building datasets, fine-tuning forecasters, running evaluations, and serving models. It documents the currently supported Python versions and points users to where the core TSFM models are hosted and how to wire up service components. Issues and examples in the tracker illustrate common tasks such as slicing inference windows or using pipeline helpers that return pandas DataFrames, grounding the library in day-to-day time-series operations. The ecosystem around TSFM also includes a community cookbook of “recipes” that showcase capabilities and patterns. Overall, the repo is designed as a hands-on companion for teams adopting time-series foundation models in production-leaning settings.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Graph of Thoughts

    Graph of Thoughts

    Official Implementation of "Graph of Thoughts

    Graph of Thoughts is an open-source framework that implements a novel reasoning paradigm for large language models by organizing reasoning steps as a structured graph instead of a simple linear chain. Traditional reasoning methods such as chain-of-thought generate sequential reasoning steps, but Graph of Thoughts introduces a more flexible structure where multiple reasoning paths can be explored and evaluated simultaneously. In this framework, problems are modeled as a graph of operations where nodes represent reasoning steps and edges represent dependencies between them. The framework executes these operations using a large language model as the reasoning engine while evaluating intermediate results to guide the search process. This approach enables models to explore multiple reasoning strategies in parallel and choose the most promising solutions during problem solving.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    HASS Configurator

    HASS Configurator

    Configuration UI for Home Assistant

    The HASS Configurator is a small web app (you access it via a web browser) that provides a filesystem browser and text-editor to modify files on the machine the configurator is running on. It has been created to allow easy configuration of Home Assistant. It is powered by Ace editor, which supports syntax highlighting for various code/markup languages. YAML files (the default language for Home Assistant configuration files) will be automatically checked for syntax errors while editing. The configurator fetches JavaScript libraries, CSS and fonts from CDNs. Hence it does NOT work when your client device is offline. And it is only available for Python 3.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    HY-Motion 1.0

    HY-Motion 1.0

    HY-Motion model for 3D character animation generation

    HY-Motion 1.0 is an open-source, large-scale AI model suite developed by Tencent’s Hunyuan team that generates high-quality 3D human motion from simple text prompts, enabling the automatic production of fluid, diverse, and semantically accurate animations without manual keyframing or rigging. Built on advanced deep learning architectures that combine Diffusion Transformer (DiT) and flow matching techniques, HY-Motion scales these approaches to the billion-parameter level, resulting in strong instruction-following capabilities and richer motion outputs compared to existing open-source models. The training strategy for the HY-Motion series includes extensive pre-training on thousands of hours of varied motion data, fine-tuning on curated high-quality datasets, and reinforcement learning with human feedback, which improves both the plausibility and adaptability of generated motion sequences.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Handcalcs

    Handcalcs

    Python library for converting Python calculations into rendered latex

    Handcalcs is a Python library that auto-renders calculation code in Jupyter notebooks or LaTeX documents with step-by-step symbolic substitution, giving output a “handwritten” feel. It supports cell magics and auto-LaTeX generation via configurable output options.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    Higher

    Higher

    higher is a pytorch library

    higher is a specialized library designed to extend PyTorch’s capabilities by enabling higher-order differentiation and meta-learning through differentiable optimization loops. It allows developers and researchers to compute gradients through entire optimization processes, which is essential for tasks like meta-learning, hyperparameter optimization, and model adaptation. The library introduces utilities that convert standard torch.nn.Module instances into “stateless” functional forms, so parameter updates can be treated as differentiable operations. It also provides differentiable implementations of common optimizers like SGD and Adam, making it possible to backpropagate through an arbitrary number of inner-loop optimization steps. By offering a clear and flexible interface, higher simplifies building complex learning algorithms that require gradient tracking across multiple update levels. Its design ensures compatibility with existing PyTorch models.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Hindsight

    Hindsight

    Hindsight: Agent Memory That Learns

    Hindsight is an advanced, open-source memory system for AI agents designed to enable long-term learning, reasoning, and consistency across interactions by treating memory as a first-class component of intelligence rather than a simple retrieval layer. It addresses one of the core limitations of modern AI agents, which is their inability to retain and meaningfully use past experiences over time, by introducing a structured, biomimetic memory architecture inspired by how human memory works. Instead of relying solely on vector similarity or basic retrieval techniques, Hindsight organizes information into distinct categories such as facts, experiences, beliefs, and observations, allowing agents to differentiate between raw data and inferred knowledge. The system operates through three core mechanisms—retain, recall, and reflect—which respectively handle storing information, retrieving relevant context, and generating new insights based on accumulated experience.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Homemade Machine Learning

    Homemade Machine Learning

    Python examples of popular machine learning algorithms

    homemade-machine-learning is a repository by Oleksii Trekhleb containing Python implementations of classic machine-learning algorithms done “from scratch”, meaning you don’t rely heavily on high-level libraries but instead write the logic yourself to deepen understanding. Each algorithm is accompanied by mathematical explanations, visualizations (often via Jupyter notebooks), and interactive demos so you can tweak parameters, data, and observe outcomes in real time. The purpose is pedagogical: you’ll see linear regression, logistic regression, k-means clustering, neural nets, decision trees, etc., built in Python using fundamentals like NumPy and Matplotlib, not hidden behind API calls. It is well suited for learners who want to move beyond library usage to understand how algorithms operate internally—how cost functions, gradients, updates and predictions work.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    Hugging Face - Speech To Speech

    Hugging Face - Speech To Speech

    Open speech-to-speech models and pipelines by Hugging Face toolkit AI

    This project from Hugging Face focuses on enabling direct speech-to-speech processing using modern machine learning models. It provides tools and reference implementations that allow audio input to be transformed into audio output without requiring an intermediate text representation. Hugging Face - Speech To Speech builds on recent advances in speech modeling, combining components such as speech recognition, translation, and synthesis into unified pipelines. It is designed to help researchers and developers experiment with multilingual and cross-lingual voice applications. It integrates with the broader Hugging Face ecosystem, making it easier to load pretrained models and run inference. It also serves as a foundation for building real-time or batch audio transformation systems. Overall, it highlights an emerging approach to voice technology that reduces latency and preserves more of the original speech characteristics.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 10
    INS

    INS

    Inspiration database for Internet practitioners with no ads

    INS is described as a kind of “inspiration database for internet workers” — a repository that collects and curates interesting websites, tools, links, or resources that might inspire developers, designers, or any knowledge workers. It aims to operate without ads, focusing purely on the content and resource quality, and leverages automation (e.g. GitHub Actions) to check link validity or site load speed, ensuring that listed resources remain accessible over time. For people in tech who constantly seek new tools, articles, or creative inspiration, ins serves as a living catalogue that can be browsed, contributed to, and relied upon for discovering useful or thought-provoking material without the noise of ads or clickbait. Because it is open-source, users can fork the database, contribute new entries, or adapt it for their personal tool-lists.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Image Super-Resolution (ISR)

    Image Super-Resolution (ISR)

    Super-scale your images and run experiments with Residual Dense

    The goal of this project is to upscale and improve the quality of low-resolution images. This project contains Keras implementations of different Residual Dense Networks for Single Image Super-Resolution (ISR) as well as scripts to train these networks using content and adversarial loss components. Docker scripts and Google Colab notebooks are available to carry training and prediction. Also, we provide scripts to facilitate training on the cloud with AWS and Nvidia-docker with only a few commands. When training your own model, start with only PSNR loss (50+ epochs, depending on the dataset) and only then introduce GANS and feature loss. This can be controlled by the loss weights argument. The weights used to produce these images are available directly when creating the model object. ISR is compatible with Python 3.6 and is distributed under the Apache 2.0 license.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    Image classification models for Keras

    Image classification models for Keras

    Keras code and weights files for popular deep learning models

    All architectures are compatible with both TensorFlow and Theano, and upon instantiation the models will be built according to the image dimension ordering set in your Keras configuration file at ~/.keras/keras.json. For instance, if you have set image_dim_ordering=tf, then any model loaded from this repository will get built according to the TensorFlow dimension ordering convention, "Width-Height-Depth". Pre-trained weights can be automatically loaded upon instantiation (weights='imagenet' argument in model constructor for all image models, weights='msd' for the music tagging model). Weights are automatically downloaded if necessary, and cached locally in ~/.keras/models/. This repository contains code for the following Keras models, VGG16, VGG19, ResNet50, Inception v3, and CRNN for music tagging.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Interpretable machine learning

    Interpretable machine learning

    Book about interpretable machine learning

    This book is about interpretable machine learning. Machine learning is being built into many products and processes of our daily lives, yet decisions made by machines don't automatically come with an explanation. An explanation increases the trust in the decision and in the machine learning model. As the programmer of an algorithm you want to know whether you can trust the learned model. Did it learn generalizable features? Or are there some odd artifacts in the training data which the algorithm picked up? This book will give an overview over techniques that can be used to make black boxes as transparent as possible and explain decisions. In the first chapter algorithms that produce simple, interpretable models are introduced together with instructions how to interpret the output. The later chapters focus on analyzing complex models and their decisions. In an ideal future, machines will be able to explain their decisions and make a transition into an algorithmic age more human.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    Invenio

    Invenio

    Invenio digital library framework

    Invenio is a highly customizable open-source framework for building large-scale digital repositories and research data platforms. Developed by CERN, it is designed to manage, index, and provide access to metadata-rich content such as publications, datasets, and multimedia files. Invenio provides a modular architecture, making it suitable for libraries, archives, and research institutions.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    KoNLPy

    KoNLPy

    Python package for Korean natural language processing

    KoNLPy is a natural language processing (NLP) library for the Korean language, offering tokenization, morphological analysis, and named entity recognition.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    LLMs-from-scratch

    LLMs-from-scratch

    Implement a ChatGPT-like LLM in PyTorch from scratch, step by step

    LLMs-from-scratch is an educational codebase that walks through implementing modern large-language-model components step by step. It emphasizes building blocks—tokenization, embeddings, attention, feed-forward layers, normalization, and training loops—so learners understand not just how to use a model but how it works internally. The repository favors clear Python and NumPy or PyTorch implementations that can be run and modified without heavyweight frameworks obscuring the logic. Chapters and notebooks progress from tiny toy models to more capable transformer stacks, including sampling strategies and evaluation hooks. The focus is on readability, correctness, and experimentation, making it ideal for students and practitioners transitioning from theory to working systems. By the end, you have a grounded sense of how data pipelines, optimization, and inference interact to produce fluent text.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    LLaVA

    LLaVA

    Visual Instruction Tuning: Large Language-and-Vision Assistant

    Visual instruction tuning towards large language and vision models with GPT-4 level capabilities.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 18
    Lightning Flash

    Lightning Flash

    Flash enables you to easily configure and run complex AI recipes

    Your PyTorch AI Factory, Flash enables you to easily configure and run complex AI recipes for over 15 tasks across 7 data domains. In a nutshell, Flash is the production-grade research framework you always dreamed of but didn't have time to build. All data loading in Flash is performed via a from_* classmethod on a DataModule. Which DataModule to use and which from_* methods are available depends on the task you want to perform. For example, for image segmentation where your data is stored in folders, you would use the from_folders method of the SemanticSegmentationData class. Our tasks come loaded with pre-trained backbones and (where applicable) heads. You can view the available backbones to use with your task using available_backbones. With Flash, swapping among 40+ optimizers and 15 + schedulers recipes are simple.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    LoLLMs WEBUI

    LoLLMs WEBUI

    Local AI WebUI for running and managing large language models offlineA

    lollms-webui is a locally hosted web interface designed to run and manage large language models without relying on external services. It provides users with a centralized environment to interact with multiple AI models, making it suitable for experimentation, development, and personal use. lollms-webui emphasizes offline capability, allowing users to maintain privacy and control over their data while still accessing advanced AI features. It integrates model management tools that help users download, configure, and switch between different language models with ease. It is built to be user-friendly while still offering advanced customization options for power users who want deeper control over model behavior. Additionally, it supports extensibility through plugins or modular components, enabling users to expand functionality as needed. Overall, it serves as a flexible platform for running AI locally with a focus on usability and adaptability.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    MCP Timeplus

    MCP Timeplus

    Execute SQL queries and manage databases seamlessly with Timeplus

    An MCP server designed for integration with Timeplus, enabling real-time data streaming and analytics through natural language interactions. ​
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    MODMAIL

    MODMAIL

    A feature rich discord Modmail bot

    Modmail is similar to Reddit's Modmail, both in functionality and purpose. It serves as a shared inbox for server staff to communicate with their users in a seamless way. This bot is free for everyone and always will be. If you like this project and would like to show your appreciation, you can support us on Patreon, cool benefits included! When a member sends a direct message to the bot, Modmail will create a channel or "thread" into a designated category. All further DM messages will automatically relay to that channel; any available staff can respond within the channel. Schedule tasks in human time, e.g. ?close in 2 hours silently. Editing and deleting messages are synced. Support for the diverse range of message contents (multiple images, files). Paginated commands interfaces via reactions. When you close a thread, Modmail will generate a log link and post it to your log channel.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 22
    Manim Python

    Manim Python

    Animation engine for explanatory math videos

    Manim is a Python library and animation engine designed for creating precise, programmatic mathematical visuals—famously used by 3Blue1Brown. It enables developers and educators to script animations using code and produce high-quality explanatory math videos.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    Mashumaro

    Mashumaro

    Fast and well tested serialization library on top of dataclasses

    When using data classes, you often need to dump and load objects based on the schema you have. Mashumaro not only lets you save and load things in different ways, but it also does it super quickly.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    MaxKB

    MaxKB

    Open-source platform for building enterprise-grade agents

    MaxKB (Max Knowledge Brain) is an open-source platform for building enterprise-grade AI agents with strong knowledge retrieval, RAG pipelines, and workflow orchestration. It focuses on practical deployments such as customer support, internal knowledge bases, research assistants, and education, bundling tools for data ingestion, chunking, embedding, retrieval, and answer synthesis. The system exposes flexible tool-use (including MCP), supports multi-model backends, and provides dashboards for dataset management and evaluation. It’s backed by an active org that also builds adjacent ops tooling, and there’s a dedicated documentation repo for configuration and contribution. Community posts describe “self-host your ChatGPT-style assistant” positioning, with integrations and workflows to move from demo to production. Security advisories are tracked publicly, with upgrade guidance when issues arise.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    Moshi

    Moshi

    A speech-text foundation model for real time dialogue

    Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audio codec. Mimi processes 24 kHz audio, down to a 12.5 Hz representation with a bandwidth of 1.1 kbps, in a fully streaming manner (latency of 80ms, the frame size), yet performs better than existing, non-streaming, codecs like SpeechTokenizer (50 Hz, 4kbps), or SemantiCodec (50 Hz, 1.3kbps). Moshi models two streams of audio: one corresponds to Moshi, and the other one to the user. At inference, the stream from the user is taken from the audio input, and the one for Moshi is sampled from the model's output. Along these two audio streams, Moshi predicts text tokens corresponding to its own speech, its inner monologue, which greatly improves the quality of its generation. A small Depth Transformer models inter codebook dependencies for a given time step, while a large, 7B parameter Temporal Transformer models the temporal dependencies.
    Downloads: 3 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB