Open Source Python Machine Learning Software - Page 23

Python Machine Learning Software

View 446 business solutions

Browse free open source Python Machine Learning Software and projects below. Use the toggles on the left to filter open source Python Machine Learning Software by OS, license, language, programming language, and project status.

  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    TensorFlow Quantum

    TensorFlow Quantum

    Open-source Python framework for hybrid quantum-classical ml learning

    TensorFlow Quantum is an open-source software framework designed for building and training hybrid quantum-classical machine learning models within the TensorFlow ecosystem. The framework enables researchers and developers to represent quantum circuits as data and integrate them directly into machine learning workflows. By combining classical deep learning techniques with quantum algorithms, the platform allows experimentation with quantum machine learning methods that may offer advantages for certain computational tasks. TensorFlow Quantum integrates with the Cirq quantum computing framework to define and manipulate quantum circuits, while leveraging TensorFlow’s infrastructure for optimization, automatic differentiation, and large-scale computation. The library also supports high-performance simulation of quantum circuits, enabling researchers to test and evaluate quantum models even without direct access to quantum hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    TensorFlow Ranking

    TensorFlow Ranking

    Learning to rank in TensorFlow

    TensorFlow Ranking is a library for Learning-to-Rank (LTR) techniques on the TensorFlow platform. Commonly used loss functions including pointwise, pairwise, and listwise losses. Commonly used ranking metrics like Mean Reciprocal Rank (MRR) and Normalized Discounted Cumulative Gain (NDCG). Multi-item (also known as groupwise) scoring functions. LambdaLoss implementation for direct ranking metric optimization. Unbiased Learning-to-Rank from biased feedback data. We envision that this library will provide a convenient open platform for hosting and advancing state-of-the-art ranking models based on deep learning techniques, and thus facilitate both academic research and industrial applications. We provide a demo, with no installation required, to get started on using TF-Ranking. This demo runs on a colaboratory notebook, an interactive Python environment. Using sparse features and embeddings in TF-Ranking.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    TensorFlow on Raspberry Pi

    TensorFlow on Raspberry Pi

    TensorFlow for Raspberry Pi

    TensorFlow on Raspberry Pi.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    TensorFlowOnSpark

    TensorFlowOnSpark

    TensorFlowOnSpark brings TensorFlow programs to Apache Spark clusters

    By combining salient features from the TensorFlow deep learning framework with Apache Spark and Apache Hadoop, TensorFlowOnSpark enables distributed deep learning on a cluster of GPU and CPU servers. It enables both distributed TensorFlow training and inferencing on Spark clusters, with a goal to minimize the amount of code changes required to run existing TensorFlow programs on a shared grid.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5

    TensorImage

    Image classification library for easily training and deploying models

    (Visit our github repository at https://github.com/TensorImage/tensorimage for more information) TensorImage is and open source package for image classification. It has a wide range of data augmentation operations that can be performed over training data to prevent overfitting and increase testing accuracy. TensorImage is easy to use and manage as all files, trained models and data are organized within a workspace directory, which you can change at any time in the configuration file, therefore being able have an indefinite amount of workspace directories for different purposes. Moreover, TensorImage can also be used to classify on thousands of images with trained image classification models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    TensorLayer

    TensorLayer

    Deep learning and reinforcement learning library for scientists

    TensorLayer is a novel TensorFlow-based deep learning and reinforcement learning library designed for researchers and engineers. It provides an extensive collection of customizable neural layers to build advanced AI models quickly, based on this, the community open-sourced mass tutorials and applications. TensorLayer is awarded the 2017 Best Open Source Software by the ACM Multimedia Society. This project can also be found at OpenI and Gitee. 3.0.0 has been pre-released, the current version supports TensorFlow, MindSpore and PaddlePaddle (partial) as the backends, allowing users to run the code on different hardware like Nvidia-GPU and Huawei-Ascend. In the future, it will support TensorFlow, MindSpore, PaddlePaddle, PyTorch and other backends. TensorLayer has a high-level layer/model abstraction which is effortless to learn. You can learn how deep learning can benefit your AI tasks in minutes through the massive examples.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    TensorNets

    TensorNets

    High level network definitions with pre-trained weights in TensorFlow

    High level network definitions with pre-trained weights in TensorFlow (tested with 2.1.0 >= TF >= 1.4.0). Applicability. Many people already have their own ML workflows and want to put a new model on their workflows. TensorNets can be easily plugged together because it is designed as simple functional interfaces without custom classes. Manageability. Models are written in tf.contrib.layers, which is lightweight like PyTorch and Keras, and allows for ease of accessibility to every weight and end-point. Also, it is easy to deploy and expand a collection of pre-processing and pre-trained weights. Readability. With recent TensorFlow APIs, more factoring and less indenting can be possible. For example, all the inception variants are implemented as about 500 lines of code in TensorNets while 2000+ lines in official TensorFlow models. Reproducibility. You can always reproduce the original results with simple APIs including feature extractions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Tensorpack

    Tensorpack

    A Neural Net Training Interface on TensorFlow, with focus on speed

    Tensorpack is a neural network training interface based on TensorFlow v1. Uses TensorFlow in the efficient way with no extra overhead. On common CNNs, it runs training 1.2~5x faster than the equivalent Keras code. Your training can probably gets faster if written with Tensorpack. Scalable data-parallel multi-GPU / distributed training strategy is off-the-shelf to use. Squeeze the best data loading performance of Python with tensorpack.dataflow. Symbolic programming (e.g. tf.data) does not offer the data processing flexibility needed in research. Tensorpack squeezes the most performance out of pure Python with various auto parallelization strategies. There are too many symbolic function wrappers already. Tensorpack includes only a few common layers. You can use any TF symbolic functions inside Tensorpack.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Texar

    Texar

    Toolkit for Machine Learning, Natural Language Processing

    Texar is a toolkit aiming to support a broad set of machine learning, especially natural language processing and text generation tasks. Texar provides a library of easy-to-use ML modules and functionalities for composing whatever models and algorithms. The tool is designed for both researchers and practitioners for fast prototyping and experimentation. Texar was originally developed and is actively contributed by Petuum and CMU in collaboration with other institutes. A mirror of this repository is maintained by Petuum Open Source. Two Versions, (Mostly) Same Interfaces. Texar-TensorFlow (this repo) and Texar-PyTorch have mostly the same interfaces. Both further combine the best design of TF and PyTorch. Rich Pre-trained Models, Rich Usage with Uniform Interfaces. BERT, GPT2, XLNet, etc, for encoding, classification, generation, and composing complex models with other Texar components!
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    TextAttack

    TextAttack

    Python framework for adversarial attacks, and data augmentation

    Generating adversarial examples for NLP models. TextAttack is a Python framework for adversarial attacks, data augmentation, and model training in NLP.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Texthero

    Texthero

    Text preprocessing, representation and visualization from zero to hero

    Texthero is a python package to work with text data efficiently. It empowers NLP developers with a tool to quickly understand any text-based dataset and it provides a solid pipeline to clean and represent text data, from zero to hero.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Tez

    Tez

    Tez is a super-simple and lightweight Trainer for PyTorch

    Tez is a super-simple and lightweight Trainer for PyTorch. It also comes with many utils that you can use to tackle over 90% of deep learning projects in PyTorch. tez (तेज़ / تیز) means sharp, fast & active. This is a simple, to-the-point, library to make your PyTorch training easy. This library is in early-stage currently! So, there might be breaking changes. Currently, tez supports cpu, single gpu and multi-gpu & tpu training. More coming soon! Using tez is super-easy. We don't want you to be far away from pytorch. So, you do everything on your own and just use tez to make a few things simpler.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    The Deep Review

    The Deep Review

    A collaboratively written review paper on deep learning, genomics, etc

    This repository is home to the Deep Review, a review article on deep learning in precision medicine. The Deep Review is collaboratively written on GitHub using a tool called Manubot (see below). The project operates on an open contribution model, welcoming contributions from anyone. To see what's incoming, check the open pull requests. For project discussion and planning see the Issues. As of writing, we are aiming to publish an update of the deep review. We will continue to make project preprints available on bioRxiv or another preprint service and aim to continue publishing the finished reviews in a peer-reviewed venue as well. Like the initial release, we are planning for an open and collaborative effort. New contributors are welcome and will be listed as version 2.0 authors. Manubot is a system for writing scholarly manuscripts via GitHub. Manubot automates citations and references, versions manuscripts using git, and enables collaborative writing via GitHub.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    The GAN Zoo

    The GAN Zoo

    A list of all named GANs

    The GAN Zoo is an open-source repository that compiles a comprehensive list of Generative Adversarial Network models published in research literature. The project began as a community effort to track the rapidly growing number of GAN architectures appearing in machine learning papers. Because new GAN models are frequently introduced in research publications, the repository serves as a convenient catalog that organizes them in one location. The list includes references to many GAN variants along with links to their original research papers and sometimes implementation code. Users can browse the dataset or explore a tabular version that allows filtering by year or searching for specific model names. The repository encourages contributions from the community so that newly published GAN architectures can be added to the list.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    The Hundred-Page Machine Learning Book

    The Hundred-Page Machine Learning Book

    The Python code to reproduce illustrations from Machine Learning Book

    The Hundred-Page Machine Learning Book is the official companion repository for The Hundred-Page Machine Learning Book written by machine learning researcher Andriy Burkov. The repository contains Python code used to generate the figures, visualizations, and illustrative examples presented in the book. Its purpose is to help readers better understand the concepts explained in the text by allowing them to run and experiment with the underlying code themselves. The book itself provides a concise overview of machine learning theory and practice, covering topics such as supervised learning, unsupervised learning, neural networks, and optimization algorithms. The repository complements these explanations by offering practical implementations that demonstrate how various algorithms behave when applied to data. Readers can explore the scripts to reproduce diagrams and observe how mathematical concepts translate into working code.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    TimeMixer

    TimeMixer

    Decomposable Multiscale Mixing for Time Series Forecasting

    TimeMixer is a deep learning framework designed for advanced time series forecasting and analysis using a multiscale neural architecture. The model focuses on decomposing time series data into multiple temporal scales in order to capture both short-term seasonal patterns and long-term trends. Instead of relying on traditional recurrent or transformer-based architectures, TimeMixer is implemented as a fully multilayer perceptron–based model that performs temporal mixing across different resolutions of the data. The architecture introduces specialized components such as Past-Decomposable-Mixing blocks, which extract information from historical sequences at different scales, and Future-Multipredictor-Mixing modules that combine predictions from multiple forecasting paths. This design allows the model to integrate complementary information across scales and produce more accurate predictions for complex temporal patterns.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Top deep learning Github repositories

    Top deep learning Github repositories

    Top 200 deep learning Github repositories sorted by stars

    Top-Deep-Learning is a curated repository that aggregates some of the most influential and widely used deep learning projects available on GitHub. Instead of providing its own machine learning models or frameworks, the project functions as an organized index that helps users discover high-quality deep learning repositories across different application domains. The repository categorizes projects related to neural networks, computer vision, natural language processing, reinforcement learning, and other areas of artificial intelligence. By collecting popular open-source implementations in one place, the project simplifies the process of exploring cutting-edge tools and research implementations for deep learning practitioners. The curated lists are particularly helpful for developers who want to quickly identify well-maintained projects with strong community support.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    TorchAudio

    TorchAudio

    Data manipulation and transformation for audio signal processing

    The aim of torchaudio is to apply PyTorch to the audio domain. By supporting PyTorch, torchaudio follows the same philosophy of providing strong GPU acceleration, having a focus on trainable features through the autograd system, and having consistent style (tensor names and dimension names). Therefore, it is primarily a machine learning library and not a general signal processing library. The benefits of PyTorch can be seen in torchaudio through having all the computations be through PyTorch operations which makes it easy to use and feel like a natural extension.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    TorchGAN

    TorchGAN

    Research Framework for easy and efficient training of GANs

    The torchgan package consists of various generative adversarial networks and utilities that have been found useful in training them. This package provides an easy-to-use API which can be used to train popular GANs as well as develop newer variants. The core idea behind this project is to facilitate easy and rapid generative adversarial model research. TorchGAN is a Pytorch-based framework for designing and developing Generative Adversarial Networks. This framework has been designed to provide building blocks for popular GANs and also to allow customization for cutting-edge research. Using TorchGAN's modular structure allows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    TorchMetrics AI

    TorchMetrics AI

    Machine learning metrics for distributed, scalable PyTorch application

    TorchMetrics is a collection of 100+ PyTorch metrics implementations and an easy-to-use API to create custom metrics.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Torchreid

    Torchreid

    Deep learning person re-identification in PyTorch

    Torchreid is a library for deep-learning person re-identification, written in PyTorch and developed for our ICCV’19 project, Omni-Scale Feature Learning for Person Re-Identification. In "deep-person-reid/scripts/", we provide a unified interface to train and test a model. See "scripts/main.py" and "scripts/default_config.py" for more details. The folder "configs/" contains some predefined configs which you can use as a starting point. The code will automatically (download and) load the ImageNet pretrained weights. After the training is done, the model will be saved as "log/osnet_x1_0_market1501_softmax_cosinelr/model.pth.tar-250". Under the same folder, you can find the tensorboard file. Different from the same-domain setting, here we replace random_erase with color_jitter. This can improve the generalization performance on the unseen target dataset.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22

    Training Image Operators from Samples

    Tools to train Image Operators automatically from a set of samples.

    TRIOS - Training Image Operators from Samples is a set of tools to bring Image Processing closer to scientists in general. It is capable of estimating an operator between two images using only pairs of samples that contain an input image and the desired output. The operator is saved to a file and can be applied to any image.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Transfer Learning Repo

    Transfer Learning Repo

    Transfer learning / domain adaptation / domain generalization

    Transfer Learning Repo is an open-source repository that compiles resources, code implementations, and academic references related to transfer learning and its related research areas. The project functions as a large knowledge hub that organizes papers, tutorials, datasets, and software implementations across topics such as domain adaptation, domain generalization, multi-task learning, and few-shot learning. The repository includes surveys and theoretical explanations that help readers understand how transfer learning methods allow models trained in one domain to adapt to new tasks or datasets. In addition to academic references, the project provides practical code implementations of many transfer learning algorithms so that researchers can reproduce experiments or build their own applications. The repository also catalogs well-known scholars, research laboratories, and datasets relevant to transfer learning studies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Transformer Reinforcement Learning X

    Transformer Reinforcement Learning X

    A repo for distributed training of language models with Reinforcement

    trlX is a distributed training framework designed from the ground up to focus on fine-tuning large language models with reinforcement learning using either a provided reward function or a reward-labeled dataset. Training support for Hugging Face models is provided by Accelerate-backed trainers, allowing users to fine-tune causal and T5-based language models of up to 20B parameters, such as facebook/opt-6.7b, EleutherAI/gpt-neox-20b, and google/flan-t5-xxl. For models beyond 20B parameters, trlX provides NVIDIA NeMo-backed trainers that leverage efficient parallelism techniques to scale effectively.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Trax

    Trax

    Deep learning with clear code and speed

    Trax is an end-to-end library for deep learning that focuses on clear code and speed. It is actively used and maintained in the Google Brain team. Run a pre-trained Transformer, create a translator in a few lines of code. Features and resources, API docs, where to talk to us, how to open an issue and more. Walkthrough, how Trax works, how to make new models and train on your own data. Trax includes basic models (like ResNet, LSTM, Transformer) and RL algorithms (like REINFORCE, A2C, PPO). It is also actively used for research and includes new models like the Reformer and new RL algorithms like AWR. Trax has bindings to a large number of deep learning datasets, including Tensor2Tensor and TensorFlow datasets. You can use Trax either as a library from your own python scripts and notebooks or as a binary from the shell, which can be more convenient for training large models. It runs without any changes on CPUs, GPUs and TPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB