Open Source Mac Machine Learning Software - Page 5

Machine Learning Software for Mac

View 59 business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Horovod

    Horovod

    Distributed training framework for TensorFlow, Keras, PyTorch, etc.

    Horovod was originally developed by Uber to make distributed deep learning fast and easy to use, bringing model training time down from days and weeks to hours and minutes. With Horovod, an existing training script can be scaled up to run on hundreds of GPUs in just a few lines of Python code. Horovod can be installed on-premise or run out-of-the-box in cloud platforms, including AWS, Azure, and Databricks. Horovod can additionally run on top of Apache Spark, making it possible to unify data processing and model training into a single pipeline. Once Horovod has been configured, the same infrastructure can be used to train models with any framework, making it easy to switch between TensorFlow, PyTorch, MXNet, and future frameworks as machine learning tech stacks continue to evolve. Start scaling your model training with just a few lines of Python code. Scale up to hundreds of GPUs with upwards of 90% scaling efficiency.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Lepton AI

    Lepton AI

    A Pythonic framework to simplify AI service building

    A Pythonic framework to simplify AI service building. Cutting-edge AI inference and training, unmatched cloud-native experience, and top-tier GPU infrastructure. Ensure 99.9% uptime with comprehensive health checks and automatic repairs.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    MONAI

    MONAI

    AI Toolkit for Healthcare Imaging

    The MONAI framework is the open-source foundation being created by Project MONAI. MONAI is a freely available, community-supported, PyTorch-based framework for deep learning in healthcare imaging. It provides domain-optimized foundational capabilities for developing healthcare imaging training workflows in a native PyTorch paradigm. Project MONAI also includes MONAI Label, an intelligent open source image labeling and learning tool that helps researchers and clinicians collaborate, create annotated datasets, and build AI models in a standardized MONAI paradigm. MONAI is an open-source project. It is built on top of PyTorch and is released under the Apache 2.0 license. Aiming to capture best practices of AI development for healthcare researchers, with an immediate focus on medical imaging. Providing user-comprehensible error messages and easy to program API interfaces. Provides reproducibility of research experiments for comparisons against state-of-the-art implementations.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    MediaPipe Solutions

    MediaPipe Solutions

    Cross-platform, customizable ML solutions

    MediaPipe is an open-source framework developed by Google for building cross-platform machine learning pipelines that process audio, video, and other streaming data in real time. The system provides developers with tools and reusable components that allow them to combine multiple machine learning models with preprocessing and postprocessing logic into efficient perception pipelines. These pipelines can run on a wide variety of platforms including mobile devices, desktop systems, web browsers, and embedded edge devices. MediaPipe is widely used in computer vision and multimedia applications such as hand tracking, face detection, pose estimation, object recognition, and gesture analysis. The framework includes prebuilt solutions that developers can quickly integrate into applications as well as lower-level APIs that allow custom pipeline construction.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    Metaflow

    Metaflow

    A framework for real-life data science

    Metaflow is a human-friendly Python library that helps scientists and engineers build and manage real-life data science projects. Metaflow was originally developed at Netflix to boost productivity of data scientists who work on a wide variety of projects from classical statistics to state-of-the-art deep learning.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    MuseGAN

    MuseGAN

    An AI for Music Generation

    MuseGAN is a deep learning research project designed to generate symbolic music using generative adversarial networks. The system focuses specifically on generating multi-track polyphonic music, meaning that it can simultaneously produce multiple instrument parts such as drums, bass, piano, guitar, and strings. Instead of generating raw audio, the model operates on piano-roll representations of music, which encode notes as time-pitch matrices for each instrument track. This representation allows the neural network to capture rhythmic patterns, harmonic relationships, and structural dependencies across instruments. The architecture is based on convolutional GAN models that learn temporal musical structure and inter-track relationships from training data. The project was trained using the Lakh Pianoroll Dataset, a large collection of multitrack musical sequences derived from MIDI files.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    PandaOCR

    PandaOCR

    Multifunctional OCR Image and Text Recognition

    At present, the newly refactored PandaOCR.Pro professional version has been released. It is faster and more stable, with richer interfaces and easier operation. It is recommended for you to use it! The normal version will continue to be maintained, and all interfaces will be retained but no new functions will be added. The reason why the version number of the professional version starts from 5.x is that the normal version will be updated in the future, so a period of version number is reserved. You can continue to use the regular version for free as before, without worrying about deactivating the regular version after the launch of the professional version. If you have higher needs, you can try the professional version. You can also use the Baidu API interface without activation. Support shortcut keys and screen corner trigger screenshot recognition function, convenient and fast.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    ROOT

    ROOT

    Analyzing, storing and visualizing big data, scientifically

    ROOT is a unified software package for the storage, processing, and analysis of scientific data: from its acquisition to the final visualization in the form of highly customizable, publication-ready plots. It is reliable, performant and well supported, easy to use and obtain, and strives to maximize the quantity and impact of scientific results obtained per unit cost, both of human effort and computing resources. ROOT provides a very efficient storage system for data models, that demonstrated to scale at the Large Hadron Collider experiments: Exabytes of scientific data are written in columnar ROOT format. ROOT comes with histogramming capabilities in an arbitrary number of dimensions, curve fitting, statistical modeling, and minimization, to allow the easy setup of a data analysis system that can query and process the data interactively or in batch mode, as well as a general parallel processing framework, RDataFrame, that can considerably speed up an analysis.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    SentencePiece

    SentencePiece

    Unsupervised text tokenizer for Neural Network-based text generation

    SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabulary size is predetermined prior to the neural model training. SentencePiece implements subword units (e.g., byte-pair-encoding (BPE) [Sennrich et al.]) and unigram language model [Kudo.]) with the extension of direct training from raw sentences. SentencePiece allows us to make a purely end-to-end system that does not depend on language-specific pre/postprocessing. Purely data driven, sentencePiece trains tokenization and detokenization models from sentences. Pre-tokenization (Moses tokenizer/MeCab/KyTea) is not always required. SentencePiece treats the sentences just as sequences of Unicode characters. There is no language-dependent logic.
    Downloads: 3 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    TTS

    TTS

    Deep learning for text to speech

    TTS is a library for advanced Text-to-Speech generation. It's built on the latest research, was designed to achieve the best trade-off among ease-of-training, speed, and quality. TTS comes with pre-trained models, tools for measuring dataset quality, and is already used in 20+ languages for products and research projects. Released models in PyTorch, Tensorflow and TFLite. Tools to curate Text2Speech datasets underdataset_analysis. Demo server for model testing. Notebooks for extensive model benchmarking. Modular (but not too much) code base enabling easy testing for new ideas. Text2Spec models (Tacotron, Tacotron2, Glow-TTS, SpeedySpeech). Speaker Encoder to compute speaker embeddings efficiently. Vocoder models (MelGAN, Multiband-MelGAN, GAN-TTS, ParallelWaveGAN, WaveGrad, WaveRNN). If you are only interested in synthesizing speech with the released TTS models, installing from PyPI is the easiest option.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Thinc

    Thinc

    A refreshing functional take on deep learning

    Thinc is a lightweight deep learning library that offers an elegant, type-checked, functional-programming API for composing models, with support for layers defined in other frameworks such as PyTorch, TensorFlow and MXNet. You can use Thinc as an interface layer, a standalone toolkit or a flexible way to develop new models. Previous versions of Thinc have been running quietly in production in thousands of companies, via both spaCy and Prodigy. We wrote the new version to let users compose, configure and deploy custom models built with their favorite framework. Switch between PyTorch, TensorFlow and MXNet models without changing your application, or even create mutant hybrids using zero-copy array interchange. Develop faster and catch bugs sooner with sophisticated type checking. Trying to pass a 1-dimensional array into a model that expects two dimensions? That’s a type error. Your editor can pick it up as the code leaves your fingers.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    Transformers

    Transformers

    State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX

    Hugging Face Transformers provides APIs and tools to easily download and train state-of-the-art pre-trained models. Using pre-trained models can reduce your compute costs, carbon footprint, and save you the time and resources required to train a model from scratch. These models support common tasks in different modalities. Text, for tasks like text classification, information extraction, question answering, summarization, translation, text generation, in over 100 languages. Images, for tasks like image classification, object detection, and segmentation. Audio, for tasks like speech recognition and audio classification. Transformers provides APIs to quickly download and use those pretrained models on a given text, fine-tune them on your own datasets and then share them with the community on our model hub. At the same time, each python module defining an architecture is fully standalone and can be modified to enable quick research experiments.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    Turing.jl

    Turing.jl

    Bayesian inference with probabilistic programming

    Bayesian inference with probabilistic programming.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    autoresearch

    autoresearch

    AI agents autonomously run and improve ML experiments overnight

    autoresearch is an experimental framework that enables AI agents to autonomously conduct machine learning research by iteratively modifying and training models. Created by Andrej Karpathy, the project allows an agent to edit the model training code, run short experiments, evaluate results, and repeat the process without human intervention. Each experiment runs for a fixed five-minute training window, enabling rapid iteration and consistent comparison across architectural or hyperparameter changes. The system centers on a simple workflow where the agent modifies a single training file while human researchers guide the process through a program.md instruction file. Designed to run on a single GPU, it keeps the research loop minimal and self-contained to make autonomous experimentation practical. Over time, the agent logs experiments, evaluates improvements, and gradually evolves the model through automated trial-and-error.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    Clustering Variation looks for a good subset of attributes in order to improve the classification accuracy of supervised learning techniques in classification problems with a huge number of attributes involved. It first creates a ranking of attributes based on the Variation value, then divide into two groups, last using Verification method to select the best group.
    Downloads: 28 This Week
    Last Update:
    See Project
  • 16
    SMILI

    SMILI

    Scientific Visualisation Made Easy

    The Simple Medical Imaging Library Interface (SMILI), pronounced 'smilie', is an open-source, light-weight and easy-to-use medical imaging viewer and library for all major operating systems. The main sMILX application features for viewing n-D images, vector images, DICOMs, anonymizing, shape analysis and models/surfaces with easy drag and drop functions. It also features a number of standard processing algorithms for smoothing, thresholding, masking etc. images and models, both with graphical user interfaces and/or via the command-line. See our YouTube channel for tutorial videos via the homepage. The applications are all built out of a uniform user-interface framework that provides a very high level (Qt) interface to powerful image processing and scientific visualisation algorithms from the Insight Toolkit (ITK) and Visualisation Toolkit (VTK). The framework allows one to build stand-alone medical imaging applications quickly and easily.
    Leader badge
    Downloads: 68 This Week
    Last Update:
    See Project
  • 17
    This project contains weka packages of neural networks algorithms implementations like Learning Vector Quantizer (LVQ) and Self-organizing Maps (SOM). For more information about weka, please visit http://www.cs.waikato.ac.nz/~ml/weka/
    Leader badge
    Downloads: 68 This Week
    Last Update:
    See Project
  • 18
    AI Cheatsheets

    AI Cheatsheets

    Essential Cheat Sheets for deep learning and machine learning research

    cheatsheets-ai is an open-source repository that collects essential cheat sheets covering many tools and concepts used in machine learning, deep learning, and data science. The project aims to provide quick-reference materials that help engineers, researchers, and students review key techniques and frameworks without reading extensive documentation. It compiles cheat sheets for widely used libraries and technologies such as TensorFlow, Keras, NumPy, Pandas, Scikit-learn, Matplotlib, and PySpark. These materials summarize common functions, workflows, and best practices in a concise visual format that makes them easy to consult during development or study sessions. The repository functions as a centralized library where users can quickly access reference materials for both machine learning theory and practical programming tools. Many of the cheat sheets are available as downloadable PDFs and images, allowing learners to keep them as quick references while working on projects.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    AI-Tutorials/Implementations Notebooks

    AI-Tutorials/Implementations Notebooks

    Codes/Notebooks for AI Projects

    AI-Tutorials/Implementations Notebooks repository is a comprehensive collection of artificial intelligence tutorials and implementation examples intended for developers, students, and researchers who want to learn by building practical AI projects. The repository contains numerous Jupyter notebooks and code samples that demonstrate modern techniques in machine learning, deep learning, data science, and large language model workflows. It includes implementations for a wide range of AI topics such as computer vision, agent systems, federated learning, distributed systems, adversarial attacks, and generative AI. Many of the tutorials focus on building AI agents, multi-agent systems, and workflows that integrate language models with external tools or APIs. The codebase acts as a hands-on learning resource, allowing users to experiment with new frameworks, architectures, and machine learning workflows through guided examples.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    AIGC-Interview-Book

    AIGC-Interview-Book

    AIGC algorithm engineer interview secrets

    AIGC-Interview-Book is a large educational repository designed to help engineers prepare for technical interviews related to artificial intelligence and generative AI roles. The project compiles knowledge from industry practitioners and researchers into a structured reference covering the AI ecosystem. Topics included in the repository span large language models, generative AI systems, traditional deep learning methods, reinforcement learning, computer vision, natural language processing, and machine learning theory. In addition to technical concepts, the repository also contains interview preparation materials such as practice questions, hiring insights, and career advice for AI engineers. The materials are organized so readers can study fundamental topics as well as advanced research areas that frequently appear in technical interviews.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 21
    BackgroundMattingV2

    BackgroundMattingV2

    Real-Time High-Resolution Background Matting

    Official repository for the paper Real-Time High-Resolution Background Matting. Our model requires capturing an additional background image and produces state-of-the-art matting results at 4K 30fps and HD 60fps on an Nvidia RTX 2080 TI GPU.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    Causal ML

    Causal ML

    Uplift modeling and causal inference with machine learning algorithms

    Causal ML is a Python package that provides a suite of uplift modeling and causal inference methods using machine learning algorithms based on recent research [1]. It provides a standard interface that allows users to estimate the Conditional Average Treatment Effect (CATE) or Individual Treatment Effect (ITE) from experimental or observational data. Essentially, it estimates the causal impact of intervention T on outcome Y for users with observed features X, without strong assumptions on the model form. An important lever to increase ROI in an advertising campaign is to target the ad to the set of customers who will have a favorable response in a given KPI such as engagement or sales. CATE identifies these customers by estimating the effect of the KPI from ad exposure at the individual level from A/B experiments or historical observational data.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    CodeSearchNet

    CodeSearchNet

    Datasets, tools, and benchmarks for representation learning of code

    CodeSearchNet is a large-scale dataset and research benchmark designed to advance the development of systems that retrieve source code using natural language queries. The project was created through collaboration between GitHub and Microsoft Research and aims to support research on semantic code search and program understanding. The dataset contains millions of pairs of source code functions and corresponding documentation comments extracted from open-source repositories. These pairs allow machine learning models to learn relationships between natural language descriptions and programming code. The dataset currently covers several widely used programming languages, including Python, JavaScript, Ruby, Go, Java, and PHP. In addition to the dataset itself, the repository includes baseline models, evaluation tools, and instructions for building code retrieval systems that can map user queries to relevant code snippets.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Daft

    Daft

    Distributed DataFrame for Python designed for the cloud

    Daft is a framework for ETL, analytics and ML/AI at scale. Its familiar Python Dataframe API is built to outperform Spark in performance and ease of use. Daft plugs directly into your ML/AI stack through efficient zero-copy integrations with essential Python libraries such as Pytorch and Ray. It also allows requesting GPUs as a resource for running models. Daft runs locally with a lightweight multithreaded backend. When your local machine is no longer sufficient, it scales seamlessly to run out-of-core on a distributed cluster. Underneath its Python API, Daft is built in blazing fast Rust code. Rust powers Daft’s vectorized execution and async I/O, allowing Daft to outperform frameworks such as Spark.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    DeepCTR-Torch

    DeepCTR-Torch

    Easy-to-use,Modular and Extendible package of deep-learning models

    DeepCTR-Torch is an easy-to-use, Modular and Extendible package of deep-learning-based CTR models along with lots of core components layers that can be used to build your own custom model easily.It is compatible with PyTorch.You can use any complex model with model.fit() and model.predict(). With the great success of deep learning, DNN-based techniques have been widely used in CTR estimation tasks. The data in the CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Low-order Extractor learns feature interaction through product between vectors. Factorization-Machine and it’s variants are widely used to learn the low-order feature interaction. High-order Extractor learns feature combination through complex neural network functions like MLP, Cross Net, etc.
    Downloads: 2 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB