首页> 外文会议>International Joint Conference on Neural Networks >Learning Private Neural Language Modeling with Attentive Aggregation
【24h】

Learning Private Neural Language Modeling with Attentive Aggregation

机译:通过专注聚合学习专用神经语言建模

获取原文
获取外文期刊封面目录资料

摘要

Mobile keyboard suggestion is typically regarded as a word-level language modeling problem. Centralized machine learning techniques require the collection of massive user data for training purposes, which may raise privacy concerns in relation to users’ sensitive data. Federated learning (FL) provides a promising approach to learning private language modeling for intelligent personalized keyboard suggestions by training models on distributed clients rather than training them on a central server. To obtain a global model for prediction, existing FL algorithms simply average the client models and ignore the importance of each client during model aggregation. Furthermore, there is no optimization for learning a well-generalized global model on the central server. To solve these problems, we propose a novel model aggregation with an attention mechanism considering the contribution of client models to the global model, together with an optimization technique during server aggregation. Our proposed attentive aggregation method minimizes the weighted distance between the server model and client models by iteratively updating parameters while attending to the distance between the server model and client models. Experiments on two popular language modeling datasets and a social media dataset show that our proposed method outperforms its counterparts in terms of perplexity and communication cost in most settings of comparison.
机译:移动键盘建议通常被视为单词级语言建模问题。集中式机器学习技术需要出于训练目的而收集大量用户数据,这可能会引起与用户敏感数据有关的隐私问题。联合学习(FL)通过在分布式客户端上训练模型而不是在中央服务器上训练模型,为学习智能个性化键盘建议的私有语言模型提供了一种有前途的方法。为了获得用于预测的全局模型,现有的FL算法仅对客户端模型进行平均,而在模型聚合过程中忽略每个客户端的重要性。此外,对于在中央服务器上学习通用化的全局模型没有优化。为了解决这些问题,我们提出了一种新的模型聚合方法,其中考虑了客户端模型对全局模型的贡献,并采用了一种关注机制,并在服务器聚合过程中采用了优化技术。我们提出的专心聚合方法通过迭代更新参数,同时注意服务器模型和客户端模型之间的距离,从而最小化了服务器模型和客户端模型之间的加权距离。在两个流行语言建模数据集和一个社交媒体数据集上进行的实验表明,在大多数比较设置中,我们提出的方法在困惑度和沟通成本方面均优于同类方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号