Open Source Python Machine Learning Software - Page 24

Python Machine Learning Software

View 446 business solutions

Browse free open source Python Machine Learning Software and projects below. Use the toggles on the left to filter open source Python Machine Learning Software by OS, license, language, programming language, and project status.

  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    U-Net Fusion RFI

    U-Net Fusion RFI

    U-Net for RFI Detection based on @jakeret's implementation

    See original code here: https://github.com/jakeret/tf_unet Currently this project is based on Tensorflow 1.13 code base and there are no plans to transfer to TF version 2. The primary improvements to this code base include a training and evaluation framework, along with a fusion based approach to detection, combining a number of models (currently hard coded to two trained models) along with Sum Threshold as an additional "expert." Additional work is being done to add custom layers to this model for further experimentation, including Squeeze/Excitation layers (unimplemented.) Sum Threshold (in fusion as an expert, and in testing as a comparison) requires the use of AOFlagger by Andre Offringa. You can find this code at https://gitlab.com/aroffringa/aoflagger. This project will use the aoflagger program within the code, so you may need to ensure that any environment variables are set for aoflagger before use. cite: https://sourceforge.net/p/u-net-fusion-rfi/wiki/cite/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    UnionML

    UnionML

    Build and deploy machine learning microservices

    Creating ML apps should be simple and frictionless. UnionML is an open-source Python framework built on top of Flyte™, unifying the complex ecosystem of ML tools into a single interface. Combine the tools that you love using a simple, standardized API so you can stop writing so much boilerplate and focus on what matters: the data and the models that learn from them. Fit the rich ecosystem of tools and frameworks into a common protocol for machine learning. Using industry-standard machine learning methods, implement endpoints for fetching data, training models, serving predictions (and much more) to write a complete ML stack in one place. Data science, ML engineering, and MLOps practitioners can all gather around UnionML apps as a way of defining a single source of truth about your ML system’s behavior. This helps you maintain consistent code across your ML stack, from training to prediction logic.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3

    Unsupervised Random Forest

    On-line Unsupervised Random Forest

    This tool uses Random Forest and PAM to cluster observations and to calculate the dissimilarity between observations. It supports on-line prediction of new observations (no need to retrain); and supports datasets that contain both continuous (e.g. CPU load) and categorical (e.g. VM instance type) features. In particular, we use an unsupervised formulation of the Random Forest algorithm to calculate similarities and provide them as input to a clustering algorithm. For the sake of efficiency and meeting the dynamism requirement of autonomic clouds, our methodology consists of two steps: (i) off-line clustering and (ii) on-line prediction. RF+PAM can: Cluster observations (Unsupervised Learning) Calculate the dissimilarity between 2 or more observations (how different two observations are)
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    UpTrain

    UpTrain

    Your open-source LLM evaluation toolkit

    Get scores for factual accuracy, context retrieval quality, guideline adherence, tonality, and many more. You can’t improve what you can’t measure. UpTrain continuously monitors your application's performance on multiple evaluation criterions and alerts you in case of any regressions with automatic root cause analysis. UpTrain enables fast and robust experimentation across multiple prompts, model providers, and custom configurations, by calculating quantitative scores for direct comparison and optimal prompt selection. Hallucinations have plagued LLMs since their inception. By quantifying degree of hallucination and quality of retrieved context, UpTrain helps to detect responses with low factual accuracy and prevent them before serving to the end-users. Unleash unparalleled power with a single line of code and tailor every detail as per as your use-case.
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Python framework for video processing and content analysis using CUDA for acceleration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6

    Virtual Botmaster

    Simulate Botnet NetFlow traffic for research analysis

    Simulate Botnet NetFlow traffic for research analysis
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    YOLOv4-large

    YOLOv4-large

    Scaled-YOLOv4: Scaling Cross Stage Partial Network

    YOLOv4-large is an open-source implementation of the Scaled-YOLOv4 object detection architecture, designed to improve both the accuracy and scalability of real-time computer vision models. The project provides a PyTorch implementation of the Scaled-YOLOv4 framework, which extends the original YOLOv4 architecture using Cross Stage Partial (CSP) networks and new scaling techniques. Unlike earlier object detection systems that only scale depth or width, this architecture scales multiple aspects of the neural network including structure, resolution, and channel configuration. This scaling strategy enables the model to adapt to different hardware environments while maintaining a strong balance between speed and detection accuracy. The repository includes multiple model variants such as YOLOv4-tiny, YOLOv4-CSP, and large-scale configurations designed for high-performance detection tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 10
    Yann
    Yann is Yet Another Neural Network. Yann is a library to create fast neural networks. It is also a GUI to easily create, edit, train, execute and investigate networks. Multiple topologies, runtime properties and ensemble learning are supported.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Yellowbrick

    Yellowbrick

    Visual analysis and diagnostic tools to facilitate ML selection

    Yellowbrick extends the Scikit-Learn API to make model selection and hyperparameter tuning easier. Under the hood, it’s using Matplotlib. Yellowbrick is a suite of visual diagnostic tools called "Visualizers" that extend the scikit-learn API to allow human steering of the model selection process. In a nutshell, Yellowbrick combines scikit-learn with matplotlib in the best tradition of the scikit-learn documentation, but to produce visualizations for your machine learning workflow.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in TensorFlow 2.0

    YoloV3 Implemented in Tensorflow 2.0

    YoloV3 Implemented in TensorFlow 2.0 is built using TensorFlow 2.0. The project provides a modern deep learning implementation of the popular YOLOv3 algorithm, which is widely used for real-time object detection in images and video streams. YOLOv3 works by dividing an image into grid regions and predicting bounding boxes and class probabilities simultaneously, allowing objects to be detected quickly and efficiently. The repository includes training scripts, inference tools, and configuration files that make it possible to train custom object detection models on user-defined datasets. It also demonstrates how to integrate the model with TensorFlow’s high-level APIs such as Keras for easier experimentation and model development. The project supports both pretrained models and full training pipelines, enabling researchers and developers to adapt YOLOv3 for tasks such as surveillance, robotics, autonomous driving, and image analysis.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Zylthra

    Zylthra

    Zylthra: A PyQt6 app to generate synthetic datasets with DataLLM.

    Welcome to Zylthra, a powerful Python-based desktop application built with PyQt6, designed to generate synthetic datasets using the DataLLM API from data.mostly.ai. This tool allows users to create custom datasets by defining columns, configuring generation parameters, and saving setups for reuse, all within a sleek, dark-themed interface.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    anaGo

    anaGo

    Bidirectional LSTM-CRF and ELMo for Named-Entity Recognition

    anaGo is a Python library for sequence labeling(NER, PoS Tagging,...), implemented in Keras. anaGo can solve sequence labeling tasks such as named entity recognition (NER), part-of-speech tagging (POS tagging), semantic role labeling (SRL) and so on. Unlike traditional sequence labeling solver, anaGo doesn't need to define any language-dependent features. Thus, we can easily use anaGo for any language. In anaGo, the simplest type of model is the Sequence model. Sequence model includes essential methods like fit, score, analyze and save/load. For more complex features, you should use the anaGo modules such as models, preprocessing and so on.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    AStro inFER - a rule miner and executer
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    auto-sklearn

    auto-sklearn

    Automated machine learning with scikit-learn

    auto-sklearn is an automated machine learning toolkit and a drop-in replacement for a scikit-learn estimator. auto-sklearn frees a machine learning user from algorithm selection and hyperparameter tuning. It leverages recent advantages in Bayesian optimization, meta-learning and ensemble construction. Auto-sklearn 2.0 includes latest research on automatically configuring the AutoML system itself and contains a multitude of improvements which speed up the fitting the AutoML system. auto-sklearn 2.0 works the same way as regular auto-sklearn. auto-sklearn is licensed the same way as scikit-learn, namely the 3-clause BSD license.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    auto_ml

    auto_ml

    Automated machine learning for analytics & production

    auto_ml is designed for production. Here's an example that includes serializing and loading the trained model, then getting predictions on single dictionaries, roughly the process you'd likely follow to deploy the trained model. Before you go any further, try running the code. Load up some data (either a DataFrame, or a list of dictionaries, where each dictionary is a row of data). Make a column_descriptions dictionary that tells us which attribute name in each row represents the value we’re trying to predict. Pass all that into auto_ml, and see what happens! You can pass in your own function to perform feature engineering on the data. This will be called as the first step in the pipeline that auto_ml builds out. You will be passed the entire X dataset (not the y dataset), and are expected to return the entire X dataset. The advantage of including it in the pipeline is that it will then be applied to any data you want predictions on later.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    automl-gs

    automl-gs

    Provide an input CSV and a target field to predict, generate a model

    Give an input CSV file and a target field you want to predict to automl-gs, and get a trained high-performing machine learning or deep learning model plus native Python code pipelines allowing you to integrate that model into any prediction workflow. No black box: you can see exactly how the data is processed, and how the model is constructed, and you can make tweaks as necessary. automl-gs is an AutoML tool which, unlike Microsoft's NNI, Uber's Ludwig, and TPOT, offers a zero code/model definition interface to getting an optimized model and data transformation pipeline in multiple popular ML/DL frameworks, with minimal Python dependencies (pandas + scikit-learn + your framework of choice). automl-gs is designed for citizen data scientists and engineers without a deep statistical background under the philosophy that you don't need to know any modern data preprocessing and machine learning engineering techniques to create a powerful prediction workflow.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    autoresearch

    autoresearch

    AI agents autonomously run and improve ML experiments overnight

    autoresearch is an experimental framework that enables AI agents to autonomously conduct machine learning research by iteratively modifying and training models. Created by Andrej Karpathy, the project allows an agent to edit the model training code, run short experiments, evaluate results, and repeat the process without human intervention. Each experiment runs for a fixed five-minute training window, enabling rapid iteration and consistent comparison across architectural or hyperparameter changes. The system centers on a simple workflow where the agent modifies a single training file while human researchers guide the process through a program.md instruction file. Designed to run on a single GPU, it keeps the research loop minimal and self-contained to make autonomous experimentation practical. Over time, the agent logs experiments, evaluates improvements, and gradually evolves the model through automated trial-and-error.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    bulbea

    bulbea

    Deep Learning based Python Library for Stock Market Prediction

    bulbea is an open-source Python library designed for financial analysis and stock market prediction using machine learning and deep learning techniques. The library provides tools for retrieving financial time series data, preprocessing market data, and training predictive models that estimate future price movements. bulbea integrates common machine learning frameworks such as TensorFlow and Keras to build neural network models capable of learning patterns in historical financial data. It includes utilities for splitting datasets, normalizing time series, and training models such as recurrent neural networks that can capture temporal dependencies in market behavior. The library also incorporates sentiment analysis capabilities that analyze social media data, particularly from Twitter, to estimate public sentiment toward financial assets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    captcha_break

    captcha_break

    Identification codes

    This project will use Keras to build a deep convolutional neural network to identify the captcha verification code. It is recommended to use a graphics card to run the project. The following visualization codes are jupyter notebookall done in . If you want to write a python script, you can run it normally with a little modification. Of course, you can also remove these visualization codes. captcha is a library written in python to generate verification codes. It supports image verification codes and voice verification codes. We use its function of generating image verification codes. First, we set our verification code format to numbers and capital letters, and generate a string of verification codes. It is well known that tensorflow occupies all video memory by default, which is not conducive to us conducting multiple experiments at the same time, so we can use the following code when tensorflow uses the video memory it needs instead of directly occupying all video memory.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    a distributed engine for abstract neural network development via natural-language programming
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    cintruder

    cintruder

    CIntruder - OCR Bruteforcing Toolkit

    Captcha Intruder is an automatic pentesting tool to bypass captchas. -> CIntruder-v0.4 (.zip) -> md5 = 6326ab514e329e4ccd5e1533d5d53967 -> CIntruder-v0.4 (.tar.gz) ->md5 = 2256fccac505064f3b84ee2c43921a68 --------------------------------------------
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    cracking-the-data-science-interview

    cracking-the-data-science-interview

    A Collection of Cheatsheets, Books, Questions, and Portfolio

    Cracking the Data Science Interview is an open educational repository that collects study materials, resources, and reference links for preparing for data science interviews. The project organizes content across many fundamental areas of data science, including statistics, probability, SQL, machine learning, and deep learning. It includes cheat sheets that summarize important technical concepts commonly discussed during technical interviews. The repository also provides links to recommended books, tutorials, practice platforms, and blog posts that help learners strengthen their theoretical and practical skills. In addition to conceptual study materials, the project includes interview question banks and case study prompts that simulate real hiring scenarios. The resource is particularly useful for candidates preparing for technical interviews in data science, machine learning, or analytics roles.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    crème de la crème of AI courses

    crème de la crème of AI courses

    This repository is a curated collection of links to various courses

    crème de la crème of AI courses is an open-source repository that serves as a curated directory of high-quality educational resources related to artificial intelligence, machine learning, and modern data science. The project aggregates links to online courses, tutorials, lecture series, and learning materials from universities, research labs, and independent educators. The repository organizes courses by topic, difficulty level, format, and release year, allowing learners to quickly identify relevant material depending on their experience and interests. Topics covered include deep learning, natural language processing, computer vision, large language models, linear algebra, reinforcement learning, and machine learning engineering. Because the repository links to well-known educational content such as university lecture series and professional training materials, it functions as a structured roadmap for individuals who want to develop expertise in artificial intelligence.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB