论文标题
部分可观测时空混沌系统的无模型预测
Collaborating Heterogeneous Natural Language Processing Tasks via Federated Learning
论文作者
论文摘要
近年来,对个人私人文本数据的隐私问题越来越多,促进了联邦学习(FL)的发展。但是,现有有关在NLP中应用FL的研究不适合与参与者与异构或私人学习目标协调参与者。在这项研究中,我们通过提出分配对比(表示为ATC)框架,进一步扩大了NLP中FL的应用范围,这使得具有异质NLP任务的客户能够构建FL课程并相互学习有用的知识。具体来说,建议客户首先执行服务器分配的统一任务,而不是使用自己的学习目标,这被称为分配培训阶段。之后,在对比培训阶段,客户以不同的本地学习目标培训,并与其他贡献一致且有用的模型更新的客户交流知识。我们对涵盖自然语言理解(NLU)和自然语言生成(NLG)任务的六个广泛使用的数据集进行了广泛的实验,与各种基线方法相比,提议的ATC框架可取得重大改进。源代码可在\ url {https://github.com/alibaba/federatedscope/tree/master/master/federatedscope/nlp/hetero_tasks}中获得。
The increasing privacy concerns on personal private text data promote the development of federated learning (FL) in recent years. However, the existing studies on applying FL in NLP are not suitable to coordinate participants with heterogeneous or private learning objectives. In this study, we further broaden the application scope of FL in NLP by proposing an Assign-Then-Contrast (denoted as ATC) framework, which enables clients with heterogeneous NLP tasks to construct an FL course and learn useful knowledge from each other. Specifically, the clients are suggested to first perform local training with the unified tasks assigned by the server rather than using their own learning objectives, which is called the Assign training stage. After that, in the Contrast training stage, clients train with different local learning objectives and exchange knowledge with other clients who contribute consistent and useful model updates. We conduct extensive experiments on six widely-used datasets covering both Natural Language Understanding (NLU) and Natural Language Generation (NLG) tasks, and the proposed ATC framework achieves significant improvements compared with various baseline methods. The source code is available at \url{https://github.com/alibaba/FederatedScope/tree/master/federatedscope/nlp/hetero_tasks}.