Foundational Models and Federated Learning: Survey, Taxonomy, Challenges and Practical Insights

AI in healthcare
Published: arXiv: 2509.05142v1
Authors

Cosmin-Andrei Hatfaludi Alex Serban

Abstract

Federated learning has the potential to unlock siloed data and distributed resources by enabling collaborative model training without sharing private data. As more complex foundational models gain widespread use, the need to expand training resources and integrate privately owned data grows as well. In this article, we explore the intersection of federated learning and foundational models, aiming to identify, categorize, and characterize technical methods that integrate the two paradigms. As a unified survey is currently unavailable, we present a literature survey structured around a novel taxonomy that follows the development life-cycle stages, along with a technical comparison of available methods. Additionally, we provide practical insights and guidelines for implementing and evolving these methods, with a specific focus on the healthcare domain as a case study, where the potential impact of federated learning and foundational models is considered significant. Our survey covers multiple intersecting topics, including but not limited to federated learning, self-supervised learning, fine-tuning, distillation, and transfer learning. Initially, we retrieved and reviewed a set of over 4,200 articles. This collection was narrowed to more than 250 thoroughly reviewed articles through inclusion criteria, featuring 42 unique methods. The methods were used to construct the taxonomy and enabled their comparison based on complexity, efficiency, and scalability. We present these results as a self-contained overview that not only summarizes the state of the field but also provides insights into the practical aspects of adopting, evolving, and integrating foundational models with federated learning.

Paper Summary

Problem
The problem this research paper addresses is how to integrate foundational models (FMs) with federated learning (FL) to unlock siloed data and distributed resources without sharing private data. This integration is important because FMs require more computational resources and diverse data, which are often siloed due to privacy concerns.
Key Innovation
The key innovation of this paper is a comprehensive literature survey that categorizes articles using a novel taxonomy based on the stage where FMs are used (e.g., pre-training or inference) and the type of FL method used. This survey provides insights into the practical aspects of adopting, evolving, and integrating FMs with FL.
Practical Impact
The practical impact of this research is that it can help healthcare providers and other organizations integrate siloed data to improve diagnostic algorithms and develop robust collaborative ML models without sharing private data. This has significant potential to improve patient outcomes and reduce costs.
Analogy / Intuitive Explanation
Think of FMs as pre-trained language models that need to be fine-tuned for specific tasks, like medical diagnosis. FL is like a collaboration platform where multiple hospitals can share their own data and train the model without sharing it with each other. By combining these two approaches, researchers can develop robust ML models that are tailored to specific healthcare needs while preserving patient privacy.
Paper Information
Categories:
cs.LG
Published Date:

arXiv ID:

2509.05142v1

Quick Actions