首页> 外国专利> METHOD FOR SERVING PARAMETER EFFICIENT NLP MODELS THROUGH ADAPTIVE ARCHITECTURES

METHOD FOR SERVING PARAMETER EFFICIENT NLP MODELS THROUGH ADAPTIVE ARCHITECTURES

机译:通过自适应架构提供参数高效NLP模型的方法

摘要

A machine learning system executed by a processor may generate predictions for a variety of natural language processing (NLP) tasks. The machine learning system may include a single deployment implementing a parameter efficient transfer learning architecture. The machine learning system may use adapter layers to dynamically modify a base model to generate a plurality of fine-tuned models. Each fine-tuned model may generate predictions for a specific NLP task. By transferring knowledge from the base model to each fine-tuned model, the ML system achieves a significant reduction in the number of tunable parameters required to generate a fine- tuned NLP model and decreases the fine-tuned model artifact size. Additionally, the ML system reduces training times for fine-tuned NLP models, promotes transfer learning across NLP tasks with lower labeled data volumes, and enables easier and more computationally efficient deployments for multi-task NLP.
机译:由处理器执行的机器学习系统可以为各种自然语言处理(NLP)任务生成预测。机器学习系统可以包括实现参数有效传输学习架构的单个部署。机器学习系统可以使用适配器层动态修改基础模型以生成多个微调模型。每个细调模型可以生成特定NLP任务的预测。通过将知识从基础模型转移到每个微调模型,ML系统实现了生成微调NLP模型所需的可调参数的数量的显着降低,并降低了微调模型伪影尺寸。此外,ML系统减少了微调NLP模型的训练时间,促进了具有较低标记数据卷的NLP任务的转移学习,并实现了用于多任务NLP的更容易和更具计算的能力部署。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号