Open Source Python Machine Learning Software - Page 13

Python Machine Learning Software

View 446 business solutions

Browse free open source Python Machine Learning Software and projects below. Use the toggles on the left to filter open source Python Machine Learning Software by OS, license, language, programming language, and project status.

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Featuretools

    Featuretools

    An open source python library for automated feature engineering

    An open source Python framework for automated feature engineering. Featuretools automatically creates features from temporal and relational datasets. Featuretools uses DFS for automated feature engineering. You can combine your raw data with what you know about your data to build meaningful features for machine learning and predictive modeling. Featuretools provides APIs to ensure only valid data is used for calculations, keeping your feature vectors safe from common label leakage problems. You can specify prediction times row-by-row. Featuretools come with a library of low-level functions that can be stacked to create features. You can build and share your own custom primitives to be reused on any dataset. Featuretools works alongside tools you already use to build machine learning pipelines. You can load in pandas' data frames and automatically create meaningful features in a fraction of the time it would take to do so manually.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Finance

    Finance

    150+ quantitative finance Python programs

    Finance is a repository that compiles structured notes and educational material related to financial analysis, markets, and quantitative finance concepts. The project focuses on explaining key principles used in finance and investment analysis, including topics such as financial statements, valuation models, portfolio theory, and financial markets. The repository is designed as a study reference for students and professionals who want to understand financial systems and the analytical frameworks used in financial decision-making. It organizes concepts into structured documents that explain both theoretical principles and practical calculations used in finance. The materials often include definitions, formulas, conceptual explanations, and examples to help readers understand how financial models and instruments function in real markets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    FineSplice

    FineSplice

    Enhanced splice junction detection and estimation from RNA-Seq data

    FineSplice is a Python wrapper to TopHat2 geared towards a reliable identification of expressed exon junctions from RNA-Seq data, at enhanced detection precision with small loss in sensitivity. Following alignment with TopHat2 using known transcript annotations, FineSplice takes as input the resulting BAM file and outputs a confident set of expressed splice junctions with the corresponding read counts. Potential false positives arising from spurious alignments are filtered out via a semi-supervised anomaly detection strategy based on logistic regression. Multiple mapping reads with a unique location after filtering are rescued and reallocated to the most reliable candidate location. FineSplice requires Python 2.x (>= 2.6) with the following modules installed: pysam (http://code.google.com/p/pysam/) and scikit-learn (http://scikit-learn.org/). For further details check out our publication: Nucl. Acids Res. (2014) doi: 10.1093/nar/gku166
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Five video classification methods

    Five video classification methods

    Code that accompanies my blog post outlining five video classification

    Classifying video presents unique challenges for machine learning models. As I’ve covered in my previous posts, video has the added (and interesting) property of temporal features in addition to the spatial features present in 2D images. While this additional information provides us more to work with, it also requires different network architectures and, often, adds larger memory and computational demands.We won’t use any optical flow images. This reduces model complexity, training time, and a whole whack load of hyperparameters we don’t have to worry about. Every video will be subsampled down to 40 frames. So a 41-frame video and a 500-frame video will both be reduced to 40 frames, with the 500-frame video essentially being fast-forwarded. We won’t do much preprocessing. A common preprocessing step for video classification is subtracting the mean, but we’ll keep the frames pretty raw from start to finish.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 5
    FlexLLMGen

    FlexLLMGen

    Running large language models on a single GPU

    FlexLLMGen is an open-source inference engine designed to run large language models efficiently on limited hardware resources such as a single GPU. The system focuses on high-throughput generation workloads where large batches of text must be processed quickly, such as large-scale data extraction or document analysis tasks. Instead of requiring expensive multi-GPU systems, the framework uses techniques such as memory offloading, compression, and optimized batching to run large models on commodity hardware. The architecture distributes computation and memory usage across the GPU, CPU, and disk in order to maximize the number of tokens processed during inference. This design allows organizations to deploy powerful language models for high-volume tasks without the infrastructure costs typically associated with large-scale AI systems. The project is particularly useful for workloads that prioritize throughput over latency, including benchmarking experiments and large corpus analysis.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Foolbox

    Foolbox

    Python toolbox to create adversarial examples

    Foolbox: Fast adversarial attacks to benchmark the robustness of machine learning models in PyTorch, TensorFlow, and JAX. Foolbox 3 is built on top of EagerPy and runs natively in PyTorch, TensorFlow, and JAX. Foolbox provides a large collection of state-of-the-art gradient-based and decision-based adversarial attacks. Catch bugs before running your code thanks to extensive type annotations in Foolbox. Foolbox is a Python library that lets you easily run adversarial attacks against machine learning models like deep neural networks. It is built on top of EagerPy and works natively with models in PyTorch, TensorFlow, and JAX.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Forecasting Best Practices

    Forecasting Best Practices

    Time Series Forecasting Best Practices & Examples

    Time series forecasting is one of the most important topics in data science. Almost every business needs to predict the future in order to make better decisions and allocate resources more effectively. This repository provides examples and best practice guidelines for building forecasting solutions. The goal of this repository is to build a comprehensive set of tools and examples that leverage recent advances in forecasting algorithms to build solutions and operationalize them. Rather than creating implementations from scratch, we draw from existing state-of-the-art libraries and build additional utilities around processing and featuring the data, optimizing and evaluating models, and scaling up to the cloud. The examples and best practices are provided as Python Jupyter notebooks and R markdown files and a library of utility functions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    GNNPCSAFT

    GNNPCSAFT

    Smart Thermodynamic Modeling with Graph Neural Networks

    The GNNPCSAFT app is an implementation of our project that focuses on using Graph Neural Networks (GNN) to estimate the pure-component parameters of the Equation of State PC-SAFT. We developed this app so the scientific community can access the model's results easily. In this app, the estimated pure-component parameters can be used to calculate thermodynamic properties and compare them with experimental data from the ThermoML Archive. To install the GNNPCSAFT app, download the appropriate latest release from the Files, unzip the file, and run the executable for your operating system (Linux or Windows). More info on github repository.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    GNNPCSAFT Web App

    GNNPCSAFT Web App

    Smart Thermodynamic Modeling with Graph Neural Networks

    The GNNPCSAFT Web App is an implementation of our project that focuses on using Graph Neural Networks (GNN) to estimate the pure-component parameters of the Equation of State PC-SAFT. We developed this app so the scientific community can access the model's results easily. In this app, the estimated pure-component parameters can be used to calculate thermodynamic properties and compare them with experimental data from the ThermoML Archive. More info on github repository.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    GPU Puzzles

    GPU Puzzles

    Solve puzzles. Learn CUDA

    GPU Puzzles is an educational project designed to teach GPU programming concepts through interactive coding exercises and puzzles. Instead of presenting traditional lecture-style explanations, the project immerses learners directly in hands-on programming tasks that demonstrate how GPU computation works. The exercises are implemented using Python with the Numba CUDA interface, which allows Python code to compile into GPU kernels that run on CUDA-enabled hardware. By solving progressively more complex puzzles, learners gain a practical understanding of how parallel algorithms operate on graphics processing units. The project emphasizes experimentation and problem solving, encouraging learners to discover GPU programming techniques through trial and exploration. It can be run in cloud environments such as Google Colab, making it easy for beginners to start experimenting without configuring local GPU hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    GPflow

    GPflow

    Gaussian processes in TensorFlow

    GPflow is a package for building Gaussian process models in Python. It implements modern Gaussian process inference for composable kernels and likelihoods. GPflow builds on TensorFlow 2.4+ and TensorFlow Probability for running computations, which allows fast execution on GPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    pygpr is a collection of algorithms that can be used to perform Gaussian process regression and global optimization.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Gluon CV Toolkit

    Gluon CV Toolkit

    Gluon CV Toolkit

    GluonCV provides implementations of state-of-the-art (SOTA) deep learning algorithms in computer vision. It aims to help engineers, researchers, and students quickly prototype products, validate new ideas and learn computer vision. It features training scripts that reproduce SOTA results reported in latest papers, a large set of pre-trained models, carefully designed APIs and easy-to-understand implementations and community support. From fundamental image classification, object detection, semantic segmentation and pose estimation, to instance segmentation and video action recognition. The model zoo is the one-stop shopping center for many models you are expecting. GluonCV embraces a flexible development pattern while is super easy to optimize and deploy without retaining a heavyweight deep learning framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    GluonNLP

    GluonNLP

    NLP made easy

    GluonNLP is a toolkit that helps you solve NLP problems. It provides easy-to-use tools that helps you load the text data, process the text data, and train models. To facilitate both the engineers and researchers, we provide command-line-toolkits for downloading and processing the NLP datasets. Gluon NLP makes it easy to evaluate and train word embeddings. Here are examples to evaluate the pre-trained embeddings included in the Gluon NLP toolkit as well as example scripts for training embeddings on custom datasets. Fasttext models trained with the library of Facebook research are exported both in text and a binary format. Unlike the text format, the binary format preserves information about subword units and consequently supports the computation of word vectors for words unknown during training (and not included in the text format). Besides training new fastText embeddings with Gluon NLP it is also possible to load the binary format into a Block provided by the Gluon NLP toolkit.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    GluonTS

    GluonTS

    Probabilistic time series modeling in Python

    GluonTS is a Python package for probabilistic time series modeling, focusing on deep learning based models. GluonTS requires Python 3.6 or newer, and the easiest way to install it is via pip. We train a DeepAR-model and make predictions using the simple "airpassengers" dataset. The dataset consists of a single time-series, containing monthly international passengers between the years 1949 and 1960, a total of 144 values (12 years * 12 months). We split the dataset into train and test parts, by removing the last three years (36 months) from the train data. Thus, we will train a model on just the first nine years of data. Python has the notion of extras – dependencies that can be optionally installed to unlock certain features of a package. We make extensive use of optional dependencies in GluonTS to keep the amount of required dependencies minimal. To still allow users to opt-in to certain features, we expose many extra dependencies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Google Research: Language

    Google Research: Language

    Shared repository for open-sourced projects from the Google AI Lang

    Google Research: Language is a shared repository maintained by Google Research that contains open-source projects developed by the Google AI Language team. The repository hosts multiple subprojects related to natural language processing, machine learning, and large-scale language understanding systems. Many of the projects included in the repository correspond to research papers released by Google researchers and provide implementations of new NLP algorithms or experimental frameworks. These implementations often explore advanced techniques such as language modeling, semantic understanding, information retrieval, and multilingual text processing. The repository functions as a collaborative hub where different research initiatives can publish their code, enabling the broader community to reproduce experiments and build upon published work.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Guild AI

    Guild AI

    Experiment tracking, ML developer tools

    Guild AI is an open-source experiment tracking toolkit designed to bring systematic control to machine learning workflows, enabling users to build better models faster. It automatically captures every detail of training runs as unique experiments, facilitating comprehensive tracking and analysis. Users can compare and analyze runs to deepen their understanding and incrementally improve models. Guild AI simplifies hyperparameter tuning by applying state-of-the-art algorithms through straightforward commands, eliminating the need for complex trial setups. It also supports the automation of pipelines, accelerating model development, reducing errors, and providing measurable results. The toolkit is platform-agnostic, running on all major operating systems and integrating seamlessly with existing software engineering tools. Guild AI supports various remote storage types, including Amazon S3, Google Cloud Storage, Azure Blob Storage, and SSH servers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Gym

    Gym

    Toolkit for developing and comparing reinforcement learning algorithms

    Gym by OpenAI is a toolkit for developing and comparing reinforcement learning algorithms. It supports teaching agents, everything from walking to playing games like Pong or Pinball. Open source interface to reinforce learning tasks. The gym library provides an easy-to-use suite of reinforcement learning tasks. Gym provides the environment, you provide the algorithm. You can write your agent using your existing numerical computation library, such as TensorFlow or Theano. It makes no assumptions about the structure of your agent, and is compatible with any numerical computation library, such as TensorFlow or Theano. The gym library is a collection of test problems — environments — that you can use to work out your reinforcement learning algorithms. These environments have a shared interface, allowing you to write general algorithms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    HDBSCAN

    HDBSCAN

    A high performance implementation of HDBSCAN clustering

    HDBSCAN - Hierarchical Density-Based Spatial Clustering of Applications with Noise. Performs DBSCAN over varying epsilon values and integrates the result to find a clustering that gives the best stability over epsilon. This allows HDBSCAN to find clusters of varying densities (unlike DBSCAN), and be more robust to parameter selection. In practice this means that HDBSCAN returns a good clustering straight away with little or no parameter tuning -- and the primary parameter, minimum cluster size, is intuitive and easy to select. HDBSCAN is ideal for exploratory data analysis; it's a fast and robust algorithm that you can trust to return meaningful clusters (if there are any).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20

    HYBRYD

    Library written in C with Python API for IPv6 networking

    This project is a rewritten of an initial project that I've called GLUE and created in 2005. I'm trying to readapt it for Python 2.7.3 and GCC 4.6.3 The library has to be build as a simple Python extension using >python setup.py install and allows to create different kind of servers, clients or hybryds (clients-servers) over (TCP/UDP) using the Ipv6 Protocol. The architecture of the code is based on brain architecture. Will put an IPv6 adress active available as soon as possible so that you can download pieces of codes. The aim of that coding was to use primary linux commands easily codable and make an object of an IPv6 connection. Moreover, the model is full-state!
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Haiku

    Haiku

    JAX-based neural network library

    Haiku is a library built on top of JAX designed to provide simple, composable abstractions for machine learning research. Haiku is a simple neural network library for JAX that enables users to use familiar object-oriented programming models while allowing full access to JAX’s pure function transformations. Haiku is designed to make the common things we do such as managing model parameters and other model state simpler and similar in spirit to the Sonnet library that has been widely used across DeepMind. It preserves Sonnet’s module-based programming model for state management while retaining access to JAX’s function transformations. Haiku can be expected to compose with other libraries and work well with the rest of JAX. Similar to Sonnet modules, Haiku modules are Python objects that hold references to their own parameters, other modules, and methods that apply functions on user inputs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Haiku Sonnet for JAX

    Haiku Sonnet for JAX

    JAX-based neural network library

    Haiku is a library built on top of JAX designed to provide simple, composable abstractions for machine learning research. JAX is a numerical computing library that combines NumPy, automatic differentiation, and first-class GPU/TPU support. Haiku is a simple neural network library for JAX that enables users to use familiar object-oriented programming models while allowing full access to JAX's pure function transformations. Haiku provides two core tools: a module abstraction, hk.Module, and a simple function transformation, hk.transform. hk.Modules are Python objects that hold references to their own parameters, other modules, and methods that apply functions on user inputs. hk.transform turns functions that use these object-oriented, functionally "impure" modules into pure functions that can be used with jax.jit, jax.grad, jax.pmap, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Hamilton DAGWorks

    Hamilton DAGWorks

    Helps scientists define testable, modular, self-documenting dataflow

    Hamilton is a lightweight Python library for directed acyclic graphs (DAGs) of data transformations. Your DAG is portable; it runs anywhere Python runs, whether it's a script, notebook, Airflow pipeline, FastAPI server, etc. Your DAG is expressive; Hamilton has extensive features to define and modify the execution of a DAG (e.g., data validation, experiment tracking, remote execution). To create a DAG, write regular Python functions that specify their dependencies with their parameters. As shown below, it results in readable code that can always be visualized. Hamilton loads that definition and automatically builds the DAG for you. Hamilton brings modularity and structure to any Python application moving data: ETL pipelines, ML workflows, LLM applications, RAG systems, BI dashboards, and the Hamilton UI allows you to automatically visualize, catalog, and monitor execution.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    High-Level Training Utilities Pytorch

    High-Level Training Utilities Pytorch

    High-level training, data augmentation, and utilities for Pytorch

    Contains significant improvements, bug fixes, and additional support. Get it from the releases, or pull the master branch. This package provides a few things. A high-level module for Keras-like training with callbacks, constraints, and regularizers. Comprehensive data augmentation, transforms, sampling, and loading. Utility tensor and variable functions so you don't need numpy as often. Have any feature requests? Submit an issue! I'll make it happen. Specifically, any data augmentation, data loading, or sampling functions. ModuleTrainer. The ModuleTrainer class provides a high-level training interface that abstracts away the training loop while providing callbacks, constraints, initializers, regularizers, and more. You also have access to the standard evaluation and prediction functions. Torchsample provides a wide range of callbacks, generally mimicking the interface found in Keras.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Hivemind

    Hivemind

    Decentralized deep learning in PyTorch. Built to train models

    Hivemind is a PyTorch library for decentralized deep learning across the Internet. Its intended usage is training one large model on hundreds of computers from different universities, companies, and volunteers. Distributed training without a master node: Distributed Hash Table allows connecting computers in a decentralized network. Fault-tolerant backpropagation: forward and backward passes succeed even if some nodes are unresponsive or take too long to respond. Decentralized parameter averaging: iteratively aggregate updates from multiple workers without the need to synchronize across the entire network. Train neural networks of arbitrary size: parts of their layers are distributed across the participants with the Decentralized Mixture-of-Experts. If you have succesfully trained a model or created a downstream repository with the help of our library, feel free to submit a pull request that adds your project to the list.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB