Open Source Federated Learning Frameworks

Browse free open source Federated Learning Frameworks and projects below. Use the toggles on the left to filter open source Federated Learning Frameworks by OS, license, language, programming language, and project status.

  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • Stay in Flow. Let Zenflow Handle the Heavy Lifting. Icon
    Stay in Flow. Let Zenflow Handle the Heavy Lifting.

    Your AI engineering control center. Zenflow turns specs into shipped features using parallel agents and multi-repo intelligence.

    Zenflow is your engineering control center, turning specs into shipped features. Parallel agents handle coding, testing, and refactoring with real repo context. Multi-agent workflows remove bottlenecks and automate routine work so developers stay focused and in flow.
    Try free now
  • 1
    Flower

    Flower

    Flower: A Friendly Federated Learning Framework

    A unified approach to federated learning, analytics, and evaluation. Federate any workload, any ML framework, and any programming language. Federated learning systems vary wildly from one use case to another. Flower allows for a wide range of different configurations depending on the needs of each individual use case. Flower originated from a research project at the University of Oxford, so it was built with AI research in mind. Many components can be extended and overridden to build new state-of-the-art systems. Different machine learning frameworks have different strengths. Flower can be used with any machine learning framework, for example, PyTorch, TensorFlow, Hugging Face Transformers, PyTorch Lightning, scikit-learn, JAX, TFLite, MONAI, fastai, MLX, XGBoost, Pandas for federated analytics, or even raw NumPy for users who enjoy computing gradients by hand.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    NErlNet

    NErlNet

    Nerlnet is a framework for research and development

    NErlNet is a research-grade framework for distributed machine learning over IoT and edge devices. Built with Erlang (Cowboy HTTP), OpenNN, and Python (Flask), it enables simulation of clusters on a single machine or real deployment across heterogeneous devices.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    Appfl

    Appfl

    Advanced Privacy-Preserving Federated Learning framework

    APPFL (Advanced Privacy-Preserving Federated Learning) is a Python framework enabling researchers to easily build and benchmark privacy-aware federated learning solutions. It supports flexible algorithm development, differential privacy, secure communications, and runs efficiently on HPC and multi-GPU setups.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Awesome-FL

    Awesome-FL

    Comprehensive and timely academic information on federated learning

    A “awesome” curated list of federated learning (FL) academic resources: research papers, tools, frameworks, datasets, tutorials, and workshops. A hub for FL knowledge maintained by the academic community.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Grafana: The open and composable observability platform Icon
    Grafana: The open and composable observability platform

    Faster answers, predictable costs, and no lock-in built by the team helping to make observability accessible to anyone.

    Grafana is the open source analytics & monitoring solution for every database.
    Learn More
  • 5
    FATE

    FATE

    An industrial grade federated learning framework

    FATE (Federated AI Technology Enabler) is the world's first industrial grade federated learning open source framework to enable enterprises and institutions to collaborate on data while protecting data security and privacy. It implements secure computation protocols based on homomorphic encryption and multi-party computation (MPC). Supporting various federated learning scenarios, FATE now provides a host of federated learning algorithms, including logistic regression, tree-based algorithms, deep learning and transfer learning. FATE became open-source in February 2019. FATE TSC was established to lead FATE open-source community, with members from major domestic cloud computing and financial service enterprises. FedAI is a community that helps businesses and organizations build AI models effectively and collaboratively, by using data in accordance with user privacy protection, data security, data confidentiality and government regulations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    FEDML Open Source

    FEDML Open Source

    The unified and scalable ML library for large-scale training

    A Unified and Scalable Machine Learning Library for Running Training and Deployment Anywhere at Any Scale. TensorOpera AI is the next-gen cloud service for LLMs & Generative AI. It helps developers to launch complex model training, deployment, and federated learning anywhere on decentralized GPUs, multi-clouds, edge servers, and smartphones, easily, economically, and securely. Highly integrated with TensorOpera open source library, TensorOpera AI provides holistic support of three interconnected AI infrastructure layers: user-friendly MLOps, a well-managed scheduler, and high-performance ML libraries for running any AI jobs across GPU Clouds. A typical workflow is shown in the figure above. When a developer wants to run a pre-built job in Studio or Job Store, TensorOperaLaunch swiftly pairs AI jobs with the most economical GPU resources, and auto-provisions, and effortlessly runs the job, eliminating complex environment setup and management.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    FL4Health

    FL4Health

    Library to facilitate federated learning research

    FL4Health is a Vector Institute toolkit for building modular, clinically-focused FL pipelines. Tailored for healthcare, it supports privacy-preserving FL, heterogeneous data settings, integrated reporting, and clear API design.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    FLEXible

    FLEXible

    Federated Learning (FL) experiment simulation in Python

    FLEXible (Federated Learning Experiments) is a Python framework offering tools to simulate FL with deep learning. It includes built-in datasets (MNIST, CIFAR10, Shakespeare), supports TensorFlow/PyTorch, and has extensions for adversarial attacks, anomaly detection, and decision trees.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    FedLab

    FedLab

    A flexible Federated Learning Framework based on PyTorch

    A Python-based framework for federated learning simulation, emphasizing modularity, communication efficiency, and algorithmic flexibility. Supports both server- and client-side customization for research and development purposes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Dominate AI Search Results Icon
    Dominate AI Search Results

    Generative Al is shaping brand discovery. AthenaHQ ensures your brand leads the conversation.

    AthenaHQ is a cutting-edge platform for Generative Engine Optimization (GEO), designed to help brands optimize their visibility and performance across AI-driven search platforms like ChatGPT, Google AI, and more.
    Learn More
  • 10
    Fedhf

    Fedhf

    A Flexible Federated Learning Simulator

    FedHF is a Python-based simulator for flexible, heterogeneous, and asynchronous federated learning research. It provides configurable resource models, supports asynchronous protocols, and accelerates experimentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Flexe

    Flexe

    The open source federated learning for vehicular network simulation

    Flexe is a FL simulator designed for connected and autonomous vehicles (CAVs). It enables horizontal/vertical/transfer FL schemes and simulates realistic wireless and vehicular dynamics. Separate Python client (PyFlexe) available.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    NVIDIA FLARE

    NVIDIA FLARE

    NVIDIA Federated Learning Application Runtime Environment

    NVIDIA Federated Learning Application Runtime Environment NVIDIA FLARE is a domain-agnostic, open-source, extensible SDK that allows researchers and data scientists to adapt existing ML/DL workflows(PyTorch, TensorFlow, Scikit-learn, XGBoost etc.) to a federated paradigm. It enables platform developers to build a secure, privacy-preserving offering for a distributed multi-party collaboration. NVIDIA FLARE is built on a componentized architecture that allows you to take federated learning workloads from research and simulation to real-world production deployment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Pfl Research

    Pfl Research

    Simulation framework for accelerating research

    A fast, modular Python framework released by Apple for privacy-preserving federated learning (PFL) simulation. Integrates with TensorFlow, PyTorch, and classical ML, and offers high-speed distributed simulation (7–72× faster than alternatives).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Substra

    Substra

    Low-level Python library used to interact with a Substra network

    An open-source framework supporting privacy-preserving, traceable federated learning and machine learning orchestration. Offers a Python SDK, high-level FL library (SubstraFL), and web UI to define datasets, models, tasks, and orchestrate secure, auditable collaborations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Taorluath is a Service-Oriented Learning Architecture based on a WAFFLE Bus design methodology which is the result of the fusion of the concepts behind the Wide Area Freely Federated Learning Environment (WAFFLE) and the Enterprise Service Bus (ESB).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Xfl

    Xfl

    An Efficient and Easy-to-use Federated Learning Framework

    XFL is a lightweight, high-performance federated learning framework supporting both horizontal and vertical FL. It integrates homomorphic encryption, DP, secure MPC, and optimizes network resilience. Compatible with major ML libraries and deployable via Docker or Conda.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next

Open Source Federated Learning Frameworks Guide

Open source federated learning frameworks are collaborative software tools that enable decentralized machine learning across multiple devices or institutions without sharing raw data. These frameworks are built to support privacy-preserving model training by allowing participants to keep data locally while contributing to a shared global model. Commonly used in sensitive domains such as healthcare, finance, and edge computing, federated learning reduces the risk of data breaches and enables compliance with privacy regulations like GDPR or HIPAA.

Notable open source federated learning frameworks include TensorFlow Federated (TFF), Flower, FedML, and PySyft. TensorFlow Federated, developed by Google, integrates with the broader TensorFlow ecosystem and offers both simulation and deployment capabilities. Flower is designed to be modular and scalable, supporting cross-device and cross-silo learning with a flexible architecture. FedML focuses on distributed AI research and practical applications with cloud, edge, and mobile support. PySyft emphasizes privacy-enhancing technologies like differential privacy and secure multi-party computation, making it suitable for research on secure federated systems.

These frameworks typically provide APIs for model training, client orchestration, aggregation strategies, and secure communication. They are often community-driven, fostering rapid experimentation and innovation. By using open source platforms, developers and researchers can build, test, and deploy federated learning systems with reduced development time and increased transparency. As interest in privacy-aware machine learning continues to grow, open source federated learning frameworks are playing a critical role in democratizing access to secure and distributed AI technologies.

Open Source Federated Learning Frameworks Features

  • Differential Privacy: Adds noise to model updates to protect individual user data from being reconstructed.
  • Secure Aggregation: Ensures the server only sees combined model updates, preventing exposure of individual client contributions.
  • Homomorphic Encryption / SMPC: Allows secure computations over encrypted data, supporting trustless environments.
  • Cross-Device & Cross-Silo FL: Supports both lightweight edge devices and large institutions, depending on the deployment need.
  • Modular and Extensible Design: Enables developers to plug in custom algorithms or modify core FL components easily.
  • Flexible Topologies: Offers centralized, decentralized (peer-to-peer), or hierarchical setups to scale across environments.
  • Model Initialization and Distribution: Handles pushing initial or pretrained models to all clients before training begins.
  • Client Update Collection: Gathers local training results from clients, either in sync or as clients become available.
  • Model Aggregation Techniques: Applies strategies like weighted averaging or robust aggregation to combine updates into a global model.
  • Simulated FL Environments: Allows researchers to prototype and test algorithms locally before live deployment.
  • Training Analytics and Logging: Tracks metrics like accuracy, loss, participation rate, and system performance.
  • Client Selection Strategies: Chooses clients dynamically based on criteria like availability, reliability, or system health.
  • Communication Compression: Minimizes bandwidth by compressing updates with techniques like quantization or sparsity.
  • Asynchronous Training Support: Lets clients operate independently, preventing slow ones from blocking progress.
  • Edge Optimization: Tailors execution to resource-constrained devices like mobile phones and IoT hardware.
  • Non-IID Data Support: Handles skewed, non-uniform client data with specialized FL algorithms for fairness and accuracy.
  • Personalization Mechanisms: Adapts the global model to better fit each client's unique local data and use case.
  • Data Partitioning Tools: Simulates realistic user data distributions in local testing environments.
  • Framework Compatibility: Works with major ML libraries like PyTorch, TensorFlow, and JAX.
  • Docker and Kubernetes Integration: Supports scalable deployment in production using containers and orchestration platforms.
  • API and SDK Access: Provides tools for building FL capabilities into mobile apps or backend systems.
  • TensorFlow Federated (TFF): Great for research; integrates with TensorFlow and supports simulation.
  • PySyft: Emphasizes privacy with support for encrypted computations and federated data science workflows.
  • Flower (FLWR): Simple, scalable, and framework-agnostic, making it useful across academic and production settings.
  • FedML: Full-stack ecosystem for mobile and edge AI, including simulation tools and public datasets.
  • FATE: Built for enterprise use in finance and healthcare, supports complex setups like vertical FL and secure computation.

Different Types of Open Source Federated Learning Frameworks

  • Horizontal Federated Learning (HFL): Designed for settings where clients share the same feature space but have different data samples. Common in cross-device learning (e.g., mobile phones). These frameworks focus on model averaging, privacy preservation, and handling data heterogeneity across many clients.
  • Vertical Federated Learning (VFL): Used when different parties hold different features about the same users. Ideal for inter-organizational collaboration (e.g., banks, hospitals). These frameworks include tools for entity alignment, encrypted computation, and feature-partitioned model training.
  • Federated Transfer Learning (FTL): Applied when both features and samples differ across participants, but there's limited overlap. These frameworks leverage transfer learning, shared embeddings, and representation alignment to allow learning between disparate domains or data sources.
  • Cross-Platform Federated Learning: Supports training across varied hardware and operating systems, such as embedded devices, IoT, and smartphones. These frameworks focus on compatibility, efficient communication, and robust execution across device types.
  • Customizable Federated Learning Toolkits: Built for researchers and developers who want flexibility. These frameworks provide modular APIs, simulation environments, and extensibility to test new algorithms, client strategies, and system configurations.
  • Privacy-Enhanced Federated Learning: Focused on strict privacy and security, often for regulated industries. These include implementations of differential privacy, secure aggregation, and encrypted gradient exchange to ensure compliance and data protection.
  • Domain-Specific Federated Learning: Tailored to particular fields such as healthcare, genomics, computer vision, or NLP. These frameworks offer specialized preprocessing, model templates, and evaluation tools suited to the unique challenges of each domain.

Advantages of Open Source Federated Learning Frameworks

  • Data privacy and security: FL frameworks keep data on local devices and use techniques like differential privacy or secure aggregation, reducing exposure and supporting compliance with regulations like GDPR and HIPAA.
  • Transparency and auditability: Open source code allows anyone to inspect, audit, or improve it, reducing the risk of hidden vulnerabilities and increasing trust among users and stakeholders.
  • Customizability and flexibility: Users can modify the framework to suit specific use cases, devices, regulatory environments, or infrastructure needs—something proprietary platforms often limit.
  • Cost-effectiveness: No licensing fees mean organizations can deploy federated learning at scale without high upfront costs, making it especially valuable for startups, academia, and nonprofits.
  • Community support and innovation: A global contributor base keeps the technology evolving quickly, offering patches, new features, integrations, and shared best practices.
  • Compatibility with existing ML tools: Many FL frameworks integrate with PyTorch, TensorFlow, and other libraries, enabling easy adoption for teams already using mainstream ML tools.
  • Scalability across device types: These frameworks support a wide variety of hardware, from smartphones to edge servers, helping distribute training tasks based on each device’s capability.
  • Academic and research-friendly: Free access and modifiability make them ideal for testing new privacy-preserving techniques or optimization methods in federated environments.
  • Support for data sovereignty: FL enables model training across regions without moving data, allowing organizations to comply with national or institutional data governance laws.
  • Resilience and fault tolerance: Because training is decentralized, it can continue even when some nodes go offline—ideal for real-world deployments with intermittent connectivity.

Types of Users That Use Open Source Federated Learning Frameworks

  • Academic researchers and students: Use FL frameworks to prototype new algorithms, study data heterogeneity, and publish reproducible research in privacy-preserving machine learning.
  • Enterprise ML engineers and data scientists: Deploy FL systems across internal data silos to train models without violating data privacy or compliance rules.
  • Mobile and edge AI developers: Integrate FL into smartphones, IoT devices, or embedded systems to enable local model training with user data while preserving privacy.
  • Privacy and security researchers: Focus on enhancing FL security by implementing differential privacy, secure aggregation, and testing against adversarial attacks.
  • AI startups and innovators: Leverage FL as a privacy-first differentiator, allowing them to offer collaborative AI solutions to industries like finance, healthcare, and telecom.
  • Healthcare data scientists: Train models across hospitals or labs while maintaining compliance with HIPAA and GDPR, enabling better diagnosis without centralizing patient data.
  • Government and defense researchers: Use FL to enable cross-agency collaboration on sensitive data, ensuring national security and data compartmentalization.
  • Open source contributors and community developers: Build and maintain FL frameworks by adding features, fixing bugs, writing documentation, and expanding compatibility.
  • Data governance and compliance officers: Oversee FL implementations to ensure regulatory compliance, data handling transparency, and auditability, often in partnership with technical teams.

How Much Do Open Source Federated Learning Frameworks Cost?

Open source federated learning frameworks are generally free to use, since they are released under permissive licenses such as MIT, Apache 2.0, or GPL. These licenses typically allow users to access, modify, and distribute the software at no cost. However, while the framework itself might be free, organizations still incur costs in integrating and maintaining it within their infrastructure. These include expenses related to setting up the computing environment, ensuring secure communication among decentralized nodes, and customizing the software for specific machine learning workflows. Additionally, ongoing updates, debugging, and community support management require time and technical expertise.

The overall cost of using an open source federated learning framework thus depends heavily on internal resource allocation and project scale. For smaller teams with in-house machine learning capabilities, these tools can offer significant savings compared to commercial platforms. On the other hand, large-scale deployments often require investment in specialized personnel, such as machine learning engineers, security experts, and DevOps professionals. Moreover, depending on the use case—like healthcare or finance—compliance and data privacy enforcement may add regulatory overhead. In short, while the software is free, the total cost of ownership involves hidden infrastructure, personnel, and operational expenses.

What Software Do Open Source Federated Learning Frameworks Integrate With?

Open source federated learning frameworks can integrate with a range of software types that support decentralized data processing, secure communication, model management, and orchestration. One major category of software is machine learning libraries and tools, such as TensorFlow, PyTorch, and Scikit-learn. These are often wrapped or extended by federated learning frameworks like TensorFlow Federated or PySyft to support distributed model training without centralizing data. These integrations allow developers to define and run standard training workflows across multiple data silos.

Another key type of software that integrates with federated learning systems is orchestration and infrastructure platforms. Kubernetes, Docker, and other container orchestration tools are frequently used to deploy and manage the components of federated systems across distributed devices or nodes. These enable scalable, reproducible deployments and streamline the management of resources across edge devices, enterprise servers, or cloud platforms.

Security-focused software, such as encryption libraries and secure multiparty computation (SMPC) tools, also integrate with federated learning frameworks. These are critical to ensuring privacy and confidentiality in training workflows. Homomorphic encryption and differential privacy libraries, for example, can be used to enhance the security of model updates and gradients shared between nodes and aggregators.

Data pipeline tools like Apache Beam, Apache Kafka, and custom ETL systems are often used in conjunction with federated learning setups to preprocess, stream, and manage data locally before it's used in training. These integrations support real-time or batch learning from decentralized data sources while maintaining compliance with privacy and regulatory requirements.

Monitoring, analytics, and MLOps platforms such as MLflow, TensorBoard, or custom dashboards can also be integrated to track training metrics, visualize model performance, and manage federated learning experiments. These tools help maintain oversight over the federated lifecycle and ensure accountability, even in distributed environments.

Together, these software integrations enable open source federated learning frameworks to operate effectively across a wide variety of use cases, from healthcare and finance to mobile devices and edge computing.

What Are the Trends Relating to Open Source Federated Learning Frameworks?

  • Growing adoption across privacy-sensitive sectors: Federated learning is increasingly used in healthcare, finance, and IoT, where data privacy regulations prevent centralized training.
  • Edge computing integration: FL is being adapted for deployment on edge devices such as smartphones and wearables, enabling local training and reducing the need to transmit raw data.
  • Frameworks tailored for different settings: Tools are distinguishing between cross-silo (enterprise-focused) and cross-device (consumer-scale) FL, each with distinct performance and security optimizations.
  • Emergence of versatile open source frameworks: Frameworks like Flower (flexible and production-ready), FedML (scalable with MLOps tools), TensorFlow Federated (research-friendly), OpenFL (Intel’s privacy-first approach), and FATE (enterprise-grade, China-centric) are leading the space.
  • Security and privacy advancements: Many frameworks now support differential privacy, secure aggregation, and homomorphic encryption to protect model updates, while also introducing defense mechanisms against poisoning and inference attacks.
  • Ecosystem integration and MLOps support: Frameworks are aligning with MLOps practices, including support for CI/CD, versioning, monitoring, and real-world deployment pipelines.
  • Improved tooling for simulation and benchmarking: Tools increasingly support realistic, non-IID (non-identically distributed) data simulations, crucial for evaluating FL in practical conditions.
  • Modular, extensible architectures: FL frameworks offer pluggable communication layers (like gRPC or MQTT), flexible aggregation strategies (e.g., FedAvg, FedProx), and sometimes multi-language API support, allowing greater customization.
  • Reproducible research support: Standardized datasets (like LEAF and FedScale), leaderboard integration, and pre-configured experiment templates help researchers compare FL methods more efficiently.
  • Global collaboration and open standards: Efforts toward standardized FL protocols, international localization, and collaboration via initiatives like OpenMined and the Linux Foundation are expanding the ecosystem.
  • Better visualization and monitoring: Real-time dashboards, client tracking, explainability tools, and federated analytics are being integrated to improve observability and compliance-readiness.
  • Scalability and efficiency improvements: Techniques like dynamic client sampling, model compression (quantization/sparsity), and adaptive scheduling make it easier to run FL in large, resource-constrained environments.
  • Emergence of advanced learning techniques: Frameworks are exploring personalized FL (e.g., client-specific tuning), multitask FL (supporting related but distinct goals), and meta-learning to enhance convergence and generalization.

How Users Can Get Started With Open Source Federated Learning Frameworks

Selecting the right open source federated learning (FL) framework involves aligning your technical and operational requirements with the design principles, capabilities, and ecosystem support of the available tools. The process begins by understanding your specific use case—whether it involves mobile devices, edge computing, healthcare, finance, or IoT environments. This influences the architecture you’ll need: client-server, peer-to-peer, or cross-silo vs cross-device. Once the use case is clear, consider the programming language and infrastructure compatibility. Many frameworks are built in Python, but if your system architecture depends on other languages or platforms, such as Java or C++, you'll need to make sure the framework supports or interoperates well with those.

Scalability and fault tolerance are also critical. The framework should support asynchronous and synchronous training modes and must be able to handle dropped clients or unreliable networks. Tools like TensorFlow Federated (TFF) are ideal for research purposes but may not scale well for production workloads. In contrast, frameworks like Flower or FedML are designed with more practical deployment in mind, providing better abstractions for scaling up federated computations.

Security and privacy features are non-negotiable in most FL applications. You should assess whether the framework supports secure aggregation, differential privacy, homomorphic encryption, or other mechanisms to safeguard both model updates and raw data. This is especially vital in regulated industries where data sensitivity is paramount. Evaluate the documentation, code quality, and the strength of the open source community around the framework. Active communities provide quicker bug fixes, better support, and faster integration of cutting-edge features. Some frameworks, like PySyft or OpenFL, are backed by academic and commercial institutions and offer robust governance models, while others may be less mature or not actively maintained.

Finally, consider extensibility and integration with your existing machine learning pipeline. The framework should integrate smoothly with major ML libraries like PyTorch, TensorFlow, or scikit-learn. It should also offer customization hooks—such as for defining aggregation strategies or incorporating new privacy-preserving techniques—without requiring deep modifications to the core framework. A dry-run or small-scale proof of concept using shortlisted frameworks can help uncover hidden complexities before committing to a full rollout.

By aligning these factors with your technical environment and business priorities, you can identify a federated learning framework that is sustainable, scalable, and secure for your intended application.