International Workshop on Federated Learning in the Age of Foundation Models
In Conjunction with IJCAI 2024 (FL@FM-IJCAI'24)

Final Submission Deadline: April 26, 2024 (23:59:59 AoE)
Notification Due: May 25, 2024 (23:59:59 AoE)
Workshop Date: Monday, August 05, 2024
Venue: International Convention Center (ICC), Jeju Island, South Korea

Call for Papers

Foundation models (FMs) are typically associated with large language models (LLMs), like ChatGPT, and are characterized by their scale and broad applicability. While these models provide transformative capabilities, they also introduce significant challenges, particularly concerning dis-tributed model management and related data privacy, efficiency, and scalability. The training of foundation models is data and resource intensive and the conventional methods are typically centralized; this creates significant challenges including regulatory and privacy concerns in real-world use cases. These include distributed training data, computational resources to manage distributed data repositories, and development of and alignment with regulatory guidelines (e.g., GDPR) that restrict sharing sensitive data.

Federated learning (FL) is an emerging paradigm that can mitigate these challenges by training a global but distributed model using distributed data. The extensive application of machine learning to analyze and draw insight from real-world, distributed, and sensitive data necessitates familiarity with and adoption of this relevant and timely topic within the general scientific community. As FL allows self-interested data owners to collaboratively train models, end-users can become co-creators of AI solutions. By adopting federated learning approaches, we can leverage distributed data and computing power available across different sources while respecting user privacy.

The rise of FMs amplifies the importance and relevance of FL as a crucial research direction. With FMs becoming the norm in machine learning development, the focus shifts from model architecture design to tackling the issues surrounding privacy-preserving and distributed learning. Advancements in FL methods have the potential to unlock the use of FMs, enabling efficient and scalable training while safeguarding sensitive data.

FMs such as GPT-4 encoded with vast knowledge and powerful emergent abilities have achieved remarkable success in various natural language processing and computer vision tasks. Grounding FMs by adapting them to domain-specific tasks or augmenting them with domain-specific knowledge enables us to exploit the full potential of FMs. However, grounding FMs faces several challenges, stemming primarily from constrained computing resources, data privacy, model heterogeneity, and model ownership. Federated Transfer Learning (FTL), the combination of FL and transfer learning, provides promising solutions to address these challenges. In recent years, the need for grounding FMs leveraging FTL, coined FTL-FM, has arisen strongly in both academia and industry.

With this in mind, we invite original research contributions, position papers, and work-in-progress reports on various aspects of federated learning in the era of foundation models. Since the emergence of foundation models has been a rela-tively recent phenomenon, their full impact on federated learning has not yet been well explored or understood. We hope to provide a platform to facilitate interaction among students, scholars, and industry professionals from around the world to discuss the latest advancements, share insights, and identify future directions in this exciting field.

This workshop aims to bring together academic researchers and industry practitioners to address open issues in this interdisciplinary research area. For industry participants, we intend to create a forum to communicate problems are practically relevant. For academic participants, we hope to make it easier to become productive in this area. The workshop will focus on the theme of combining FL with FM to open up opportunities to address new challenges. The workshop topics include but are not limited to:
Theory and algorithmic foundations:
  • Impact of heterogeneity in FL of large models
  • Multi-stage model training (e.g., base model + fine tuning)
  • Optimization advances in FL (e.g., beyond first-order and local methods)
  • Prompt tuning in federated settings
  • Self-supervised learning in federated settings
Leveraging foundation models to improve federated learning:
  • Adaptive aggregation strategies for FL in heterogeneous environments
  • Foundation model enhanced FL knowledge distillation
  • Overcoming data interoperability challenges using foundation models
  • Personalization of FL with foundation models
Federated learning for training and tuning foundation models:
  • Fairness, bias, and interpretability challenges in FL with foundation models
  • Federated transfer learning with foundation models
  • FL techniques for training large-scale foundation models
  • Hardware for FL with foundation models
  • Optimization algorithms for federated training of foundation models
  • Privacy-preserving mechanisms in FL with foundation models
  • Resource-efficient FL with foundation models
  • Security and robustness considerations in FL with foundation models
  • Systems and infrastructure for FL with foundation models
  • Vertical federated learning with foundation models
  • Vulnerabilities of FL with foundation models

More information on previous workshops can be found here.

Submission Instructions

Each submission can be up to 7 pages of contents plus up to 2 additional pages of references and acknowledgements. The submitted papers must be written in English and in PDF format according to the IJCAI'24 template. All submitted papers will be under a single-blind peer review for their novelty, technical quality and impact. The submissions can contain author details. Submission will be accepted via the Easychair submission website.

Based on the requirement from IJCAI'24, at least one author of each accepted paper must travel to the IJCAI venue in person. In addition, multiple submissions of the same paper to more than one IJCAI workshop are forbidden.

Easychair submission site:

For enquiries, please email to:

Post Workshop Publications

Selected papers will be invited to be included to an edited book on FL@FM to be published by the LNAI series by SpringerNature.

Organizing Committee

Steering Co-Chairs     General Co-Chairs     Program Co-Chairs

Qiang Yang

Yaochu Jin
(Westlake U)

Randy Goebel
(U Alberta)

Lixin Fan

Xiaoxiao Li

Han Yu
Publicity Co-Chairs     Publications Co-Chairs

Alysa Ziying Tan

Jiankai Sun

Ying Wei

Zengxiang Li

Program Committee

  • Alysa Ziying Tan (Alibaba-NTU Singapore Joint Research Institute)
  • Anran Li (Yale University, USA)
  • Chun-Yin Huang (University of British Colubia, Canada)
  • Guojun Zhang (Huawei Noah's Ark Lab, China)
  • Hongyi Peng (Nanyang Technological University, Singapore)
  • Jiangtian Nie (Nanyang Technological University, Singapore)
  • Jiankai Sun (The Ohio State University, USA)
  • Jihong Park (Deakin University, Australia)
  • Jinhyun So (University of Southern California, USA)
  • Jun Luo (Huawei Noah's Ark Lab, China)
  • Peng Zhang (Guangzhou University, China)
  • Rui Liu Nanyang (Technological University, Singapore)
  • Shiqiang Wang (IBM Thomas J. Watson Research Center, USA)
  • Siwei Feng (Soochow University, China)
  • Songze Li (Southeast University, China)
  • Wei Yang Bryan Lim (Nanyang Technological University, Singapore)
  • Wenlong Deng (University of British Colubia, Canada)
  • Xi Chen (Huawei, Canada)
  • Xiaohu Wu (Beijing University of Posts and Telecommunications, China)
  • Xiaoli Tang (Nanyang Technological University, Singapore)
  • Xu Guo (Nanyang Technological University, Singapore)
  • Yanci Zhang (Nanyang Technological University, Singapore)
  • Yang Zhang (Nanjing University of Aeronautics and Astronautics, China)
  • Yulan Gao (Nanyang Technological University, Singapore)
  • Yuxin Shi (Nanyang Technological University, Singapore)
  • Zelei Liu (Unicom (Shanghai) Industrial Internet Co. Ltd., China)
  • Zhuan Shi (EPFL, Switzerland)
  • Zichen Chen (University of California Santa Barbara, USA)

Organized by