Federated Learning in the Age of Foundation Models
Summer School @ Singapore 2024 (FL@FM-Singapore'24)


Date: May 13-14, 2024
Venue: Infuse Level 14, A*STAR, Connexis South Tower, FP 1, Singapore (138632)
Register here for free, before May 10, 2024!
Lunch & coffee breaks provided!

Program


Day 1: Monday, May 13, 2024
  
Time Activity
  
09:00 – 09:30 Registration
09:30 – 10:10 Keynote Talk 1: Trustworthy Federated Learning: Research and Applications, by Qiang Yang
10:10 – 10:40 Keynote Talk 2: Advancing Time Series Sensor Data Analytics: Harnessing Federated Learning and Foundation Models, by Xiaoli Li
10:40 – 11:00 Coffee Break
11:00 – 11:30 Keynote Talk 3: Introduction to Federated Recommendations, by Guodong Long
11:30 – 12:00 Keynote Talk 4: Rethinking Benchmarks for Machine Learning Systems, by Bingsheng He
12:00 – 14:00 Lunch
14:00 – 15:10 Briefings Session 1 (10 min per talk, including Q&A)
  1. Anran Li: Historical Embedding-Guided Efficient Large-Scale Federated Graph Learning
  2. Yanci Zhang: LR-XFL: Logical Reasoning-based Explainable Federated Learning
  3. Alysa Tan: FL-CLIP: Bridging Plasticity and Stability in Pre-Trained Federated Class-Incremental Learning Models
  4. Xiaoli Tang: Dual Calibration-based Personalised Federated Learning
  5. Liping Yi: FedSSA: Semantic Similarity-based Aggregation for Efficient Model-Heterogeneous Personali
  6. Hongyan Chang: Bias Propagation in Federated Learning
  7. Sixu Hu: Communication-Efficient Generalized Neuron Matching for Federated Learning
15:10 – 15:30 Coffee Break
15:30 – 16:40 Briefings Session 2 (10 min per talk, including Q&A)
  1. Xianjie Guo: Sample Quality Heterogeneity-aware Federated Causal Discovery through Adaptive Variable Space Selection
  2. Pengwei Xing: Federated Neuro-Symbolic Learning
  3. Hongyi Peng: FedCal: Achieving Local and Global Calibration in Federated Learning via Aggregated Parameterized Scaler
  4. Yuxin Shi: Fairness-Aware Job Scheduling for Multi-Job Federated Learning
  5. Jiewei Chen: Enabling Foundation Models: A Distributed Collaboration Framework via Graph-based Federated Learning
  6. Lulu Wang: Reconciliation of Privacy Protection and Security Defense in Federated Learning
  7. Yiqun Diao: Towards addressing label skews in one-shot federated learning
16:40 – 17:00 Break
17:00 – 17:30 Keynote Talk 5: Federated Continual Learning via Prompt-based Dual Knowledge Transfer, by Ying Wei
17:30 – 17:50 Keynote Talk 6: Model Ownership in Federated Learning, by Chee Seng Chan
17:50 – 18:20 Formation of Discussion Groups
   

Day 2: Tuesday, May 14, 2024
  
Time Activity
  
09:00 – 09:30 Registration
09:30 – 10:00 Keynote Talk 7: Trustworthy Federated Learning and Large Language Models, by Xiaojin Zhang
10:00 – 10:30 Coffee Break
10:30 – 11:00 Training Session: How to look for a research topic for Trustworthy Federated Learning, by Lixin Fan
11:00 – 12:00 Group Discussions
12:00 – 14:00 Lunch
14:00 – 15:00 Group Discussions
15:00 – 15:30 Coffee Break
15:30 – 17:00 Group Discussions
17:00 – 18:00 Student Summarization
   

Overview

Foundation models (FMs) are typically associated with large language models (LLMs), like ChatGPT, and are characterized by their scale and broad applicability. While these models provide transformative capabilities, they also introduce significant challenges, particularly concerning dis-tributed model management and related data privacy, efficiency, and scalability. The training of foundation models is data and resource intensive and the conventional methods are typically centralized; this creates significant challenges including regulatory and privacy concerns in real-world use cases. These include distributed training data, computational resources to manage distributed data repositories, and development of and alignment with regulatory guidelines (e.g., GDPR) that restrict sharing sensitive data.

Federated learning (FL) is an emerging paradigm that can mitigate these challenges by training a global but distributed model using distributed data. The extensive application of machine learning to analyze and draw insight from real-world, distributed, and sensitive data necessitates familiarity with and adoption of this relevant and timely topic within the general scientific community. As FL allows self-interested data owners to collaboratively train models, end-users can become co-creators of AI solutions. By adopting federated learning approaches, we can leverage distributed data and computing power available across different sources while respecting user privacy.

The rise of FMs amplifies the importance and relevance of FL as a crucial research direction. With FMs becoming the norm in machine learning development, the focus shifts from model architecture design to tackling the issues surrounding privacy-preserving and distributed learning. Advancements in FL methods have the potential to unlock the use of FMs, enabling efficient and scalable training while safeguarding sensitive data.

FMs such as GPT-4 encoded with vast knowledge and powerful emergent abilities have achieved remarkable success in various natural language processing and computer vision tasks. Grounding FMs by adapting them to domain-specific tasks or augmenting them with domain-specific knowledge enables us to exploit the full potential of FMs. However, grounding FMs faces several challenges, stemming primarily from constrained computing resources, data privacy, model heterogeneity, and model ownership. Federated Transfer Learning (FTL), the combination of FL and transfer learning, provides promising solutions to address these challenges. In recent years, the need for grounding FMs leveraging FTL, coined FTL-FM, has arisen strongly in both academia and industry.

In this summer school, we bring together academic researchers and industry practitioners to present their latest research and discuss visions towards addressing open issues in this interdisciplinary research area. It will focus on the theme of combining FL with FM to open up opportunities to address new challenges. It will be a useful platform for students working in related areas to interact closely with leading scientists in the field.


Invited Speakers

   

Title: Trustworthy Federated Learning: Research and Applications

Speaker: Qiang Yang, Chief AI Officer (CAIO), WeBank / Professor Emeritus, Hong Kong University of Science and Technology (HKUST), Hong Kong, China

Biography
Qiang Yang is the head of the AI Department at WeBank (Chief AI Officer) and Professor Emeritus at the Computer Science and Engineering (CSE) Department of the Hong Kong University of Science and Technology (HKUST), where he was a former head of CSE Department and founding director of the Big Data Institute (2015-2018). His research interests include AI, machine learning, and data mining, especially in transfer learning, automated planning, federated learning, and case-based reasoning. He is a fellow of several international societies, including ACM, AAAI, IEEE, IAPR, and AAAS. He received his Ph.D. in Computer Science in 1989 and his M.Sc. in Astrophysics in 1985, both from the University of Maryland, College Park. He obtained his B.Sc. in Astrophysics from Peking University in 1982. He had been a faculty member at the University of Waterloo (1989-1995) and Simon Fraser University (1995-2001). He was the founding Editor-in-Chief of the ACM Transactions on Intelligent Systems and Technology (ACM TIST) and IEEE Transactions on Big Data (IEEE TBD). He served as the President of International Joint Conference on AI (IJCAI, 2017-2019) and an executive council member of Association for the Advancement of AI (AAAI, 2016-2020). Qiang Yang is a recipient of several awards, including the 2004/2005 ACM KDDCUP Championship, the ACM SIGKDD Distinguished Service Award (2017), and AAAI Innovative Application Awards (2018, 2020 and 2022). He was the founding director of Huawei's Noah's Ark Lab (2012-2014) and a co-founder of 4Paradigm Corp, an AI platform company. He is an author of several books including Intelligent Planning (Springer), Crafting Your Research Future (Morgan & Claypool), and Constraint-based Design Recovery for Software Engineering (Springer).

   

Title: Advancing Time Series Sensor Data Analytics: Harnessing Federated Learning and Foundation Models

Speaker: Xiaoli Li, Department Head (Machine Intellection), Institute of Infocomm Research (I2R), A*STAR, Singapore

Biography
Dr. Li Xiaoli is the Department Head and Principal Scientist of the Machine Intellection (MI) department at the Institute for Infocomm Research (I2R), A*STAR, Singapore. He also holds adjunct full professor position at School of Computer Science and Engineering, Nanyang Technological University. He has been a member of ITSC (Information Technology Standards Committee) from ESG Singapore and IMDA since 2020 and has served as joint lab directors with a few major industry partners. He holds the title of IEEE Fellow and is also recognised as a Fellow of the Asia-Pacific Artificial Intelligence Association (AAIA). Dr. Li also served as a health innovation expert panel member for the Ministry of Health (MOH), as well as an AI advisor for the Smart Nation and Digital Government Office (SNDGO), Prime Minister's Office, highlighting his extensive involvement in key Government and industry initiatives. His research interests include AI, data mining, machine learning, and bioinformatics. He has been serving as the Chair of many leading AI/data mining/machine learning related conferences & workshops (including KDD, ICDM, SDM, PKDD/ECML, ACML, PAKDD, WWW, IJCAI, AAAI, ACL, and CIKM). He currently serves as editor-in-chief of Annual Review of Artificial Intelligence, and associate editor of IEEE Transactions on Artificial Intelligence, Knowledge and Information Systems, and Machine Learning with Applications (Elsevier).

   

Title: Introduction to Federated Recommendations

Speaker: Guodong Long, Associate Professor, University of Technology Sydney (UTS), Australia

Biography
Dr. Guodong Long is an Associate Professor in School of Computer Science, Faculty of Engineering and IT (FEIT), University of Technology Sydney (UTS), Australia. He is one of the core members of the Research Australian Artificial Intelligence Institute (AAII). He is currently leading a research group to conduct application-driven research on machine learning and data science. Particularly, his research interests focus on several application domains, such as NLP, Healthcare, Smart Home, Education and Social Media. He is dedicated on exploring the blue-sky research ideas with real-world value and impact. His group's research is funded by multiple sources of industry grants and ARC grants. He has published more than 100 papers on top-tier conferences including ICLR, ICML, NeurIPS, AAAI, IJCAI, ACL, KDD, WebConf, and journals including IEEE TPAMI, TKDE and TNNLS. His publications attract more than 10k citations. He will serve as a general co-chair for the WebConf 2025 to be hosted in Sydney.

   

Title: Rethinking Benchmarks for Machine Learning Systems

Speaker: Bingsheng He, Professor and Vice-Dean (Research), National University of Singapore (NUS), Singapore

Biography
Dr. Bingsheng He is currently a Professor and Vice-Dean (Research) at School of Computing, National University of Singapore. Before that, he was a faculty member in Nanyang Technological University, Singapore (2010-2016), and held a research position in the System Research group of Microsoft Research Asia (2008-2010), where his major research was building high performance cloud computing systems for Microsoft. He got the Bachelor degree in Shanghai Jiao Tong University (1999-2003), and the Ph.D. degree in Hong Kong University of Science & Technology (2003-2008). His current research interests include cloud computing, database systems and high performance computing. He has been a winner for industry faculty awards from Microsoft/NVIDIA/Xilinx/Alibaba. His work also won multiple recognitions as “Best papers” collection or awards in top forums such as SIGMOD 2008, VLDB 2013 (demo), IEEE/ACM ICCAD 2017, PACT 2018, IEEE TPDS 2019, FPGA 2021 and VLDB 2023 (industry). Since 2010, he has (co-)chaired a number of international conferences and workshops, including IEEE CloudCom 2014/2015, BigData Congress 2018, ICDCS 2020 and ICDE 2024. He has served in editor board of international journals, including IEEE Transactions on Cloud Computing (IEEE TCC), IEEE Transactions on Parallel and Distributed Systems (IEEE TPDS), IEEE Transactions on Knowledge and Data Engineering (TKDE), Springer Journal of Distributed and Parallel Databases (DAPD) and ACM Computing Surveys (CSUR). He is an ACM Distinguished member (class of 2020).

   

Title: Federated Continual Learning via Prompt-based Dual Knowledge Transfer

Speaker: Ying Wei, Nanyang Assistant Professor, Nanyang Technological University (NTU), Singapore

Biography
I am currently a Nanyang Assistant Professor with School of Computer Science and Engineering, Nanyang Technological University. I am generally interested in developing algorithms that equip machines with more general intelligence via knowledge transfer and compositionality. This includes allowing continuous transfer and adaptation of the knowledge previous learned (nowadays in LLMs) to quickly learn the current task with minimal human supervision, and autonomously evaluating the success of knowledge transfer. I am also passionate about applying these algorithms into real-world applications with small data, e.g., drug discovery. Previously, I was an Assistant Professor at Department of Computer Science, City University of Hong Kong and a senior researcher at Tencent AI Lab. I completed my Ph.D. in Computer Science and Engineering at Hong Kong University of Science and Technology under the supervision of Professor Qiang Yang, and my B.S. in Automation at Huazhong University of Science and Technology. I have also spent time interning at Microsoft Research Asia.

   

Title: Model Ownership in Federated Learning

Speaker: Chee Seng Chan, Professor, Universiti Malaya, Malaysia

Biography
My research interests include computer vision and machine learning, where I lead a young and energetic research team that has published more than 100 papers in related top peer-review conferences and journals (e.g. CVPR, NeurIPS, TPAMI, TIP etc). I was the founding Chair for IEEE Computational Intelligence Society, Malaysia chapter. Also currently, I serve as the Associate Editor of Pattern Recognition (Elsevier), and have co-organized several conferences/workshops/tutorials/challenges related to computer vision/machine learning. I was the recipient of Top Research Scientists Malaysia (TRSM) in 2022, Young Scientists Network Academy of Sciences Malaysia (YSN-ASM) in 2015 and Hitachi Research Fellowship in 2013. Besides that, I am also a senior member (IEEE), Professional Engineer (BEM) and Chartered Engineer (IET). During 2020-2022, I was seconded to the Ministry of Science, Technology and Innovation (MOSTI) as the Lead of PICC Unit under COVID19 Immunisation Task Force (CITF), as well as the Undersecretary for Division of Data Strategic and Foresight.

   

Title: Trustworthy Federated Learning and Large Language Models

Speaker: Xiaojin Zhang, Assistant Professor, Huazhong University of Science and Technology, China

Biography
Xiaojin Zhang is an assistant professor at Huazhong University of Science and Technology (HUST). She served as a postdoctoral researcher at the Hong Kong University of Science and Technology (HKUST) from August 2021 to August 2023. She was supervised by Professor Qiang Yang and Professor Kai Chen. Her primary research focus is trustworthy federated learning. She obtained her Ph.D. degree in Computer Science and Engineering from the Chinese University of Hong Kong (CUHK) in 2021. Her Ph.D. advisors were Professor Shengyu Zhang and Professor Siu On Chan and her research mainly revolved around theoretical computer science. Zhang Xiaojin earned her Master's degree in Engineering from the Institute of Computing Technology at the University of Chinese Academy of Sciences (UCAS) in 2017. Under the guidance of Professor Simin He, her research focused on bioinformatics applications. She completed her undergraduate studies in the Experimental Class of Mathematical Economics at the School of Mathematics and Statistics, Shandong University, in 2014, obtaining dual Bachelor's degrees in Mathematics and Finance. Zhang Xiaojin has served as a reviewer, program committee member, and session chair for conferences such as UAI, IJCAI, NeurIPS, ICLR, ECC, ICMLA, SmartCloud, Algorithms, TIST, TETCI, TMC, Information Sciences, and Neural Networks. She has received awards such as the National Scholarship and the first prize in the National Mathematical Modeling Competition. Her research findings have been published in top international conferences and journals such as ICML, AAAI, TCS, ICDE, UAI, TIST, and Nature Communications.


Briefings



Anran Li (NTU)

"Historical Embedding-Guided Efficient Large-Scale Federated Graph Learning," SIGMOD'24


Yanci Zhang (NTU)

"LR-XFL: Logical Reasoning-based Explainable Federated Learning," AAAI-24


Alysa Ziying Tan (NTU)

"FL-CLIP: Bridging Plasticity and Stability in Pre-Trained Federated Class-Incremental Learning Models," ICME'24
 


Xiaoli Tang (NTU)

"Dual Calibration-based Personalised Federated Learning," IJCAI'24


Liping Yi (NTU)

"FedSSA: Semantic Similarity-based Aggregation for Efficient Model-Heterogeneous Personalized Federated Learning," IJCAI'24


Xianjie Guo (NTU)

"Sample Quality Heterogeneity-aware Federated Causal Discovery through Adaptive Variable Space Selection," IJCAI'24
 


Yuxin Shi (NTU)

"Fairness-Aware Job Scheduling for Multi-Job Federated Learning," ICASSP'24


Pengwei Xing (NTU)

"Federated Neuro-Symbolic Learning," ICML'24


Hongyi Peng (NTU)

"FedCal: Achieving Local and Global Calibration in Federated Learning via Aggregated Parameterized Scaler," ICML'24
 


Jiewei Chen (SUTD)

"Enabling Foundation Models: A Distributed Collaboration Framework via Graph-based Federated Learning"


Lulu Wang (SUTD)

"Reconciliation of Privacy Protection and Security Defense in Federated Learning"


Hongyan Chang (NUS)

"Bias Propagation in Federated Learning"
 


Yiqun Diao (NUS)

"Towards addressing label skews in one-shot federated learning"


Sixu Hu (NUS)

"Communication-Efficient Generalized Neuron Matching for Federated Learning"

Organizing Committee



Lixin Fan
(WeBank)
   

Sin Gee Teo
(A*STAR)
   

Xiaoxiao Li
(UBC)
   

Han Yu
(NTU)
   

Organized by