We are interested in developing a unified machine learning framework for effectively training machine learning models from many small data sources such as mobile devices. This is a commonly encountered situation in mobile computing scenarios, where data is scarce and distributed while the tasks are distinct. In this paper, we propose a federated few-shot learning (FedFSL) framework to learn a few-shot classification model that can classify unseen data classes with only a few labeled samples. With the federated learning strategy, FedFSL can utilize many data sources while keeping data privacy and communication efficiency. To tackle the issue of obtaining misaligned decision boundaries produced by client models, we propose to regularize local updates by minimizing the divergence of client models. We also formulate the training in an adversarial fashion and optimize the client models to produce a discriminative feature space that can better represent unseen data samples. We demonstrate the intuitions and conduct experiments to show our approaches outperform baselines by more than 10% in learning benchmark vision tasks and 5% in language tasks.
Song WangXingbo FuKaize DingChen ChenHuiyuan ChenJundong Li
Yunfeng ZhaoGuoxian YuJun WangCarlotta DomeniconiMaozu GuoXiangliang ZhangLizhen Cui
Wenke HuangMang YeBo DuXiang Gao
Anastasiya DanilenkaKarolina BogackaKatarzyna Wasielewska-Michniewska
Dongqi CaiShangguang WangYaozong WuFelix Xiaozhu LinMengwei Xu