首页> 外文学位 >Graphical Models for Heterogeneous Transfer Learning and Co-reference Resolution.
【24h】

Graphical Models for Heterogeneous Transfer Learning and Co-reference Resolution.

机译:异构转移学习和共同参考解析的图形模型。

获取原文
获取原文并翻译 | 示例

摘要

Traditional supervised machine learning requires labeled data for a specific problem of interest. There have been many attempts to reduce this requirement such as approaches based on semi-supervised learning. In recent years, people have started to consider a new strategy known as transfer learning, where labeled data from an old problem (called the source task) is used to assist the learning of a new but related problem (the target task).;In this thesis, we mainly consider an extreme case of transfer learning that we denote as heterogeneous transfer learning - where the feature spaces of the source task and the target tasks are disjoint. We first consider the cross-lingual text classification task, where we need to train a classifier for Chinese but we only have labeled data in English. We adapt the structural correspondence learning (SCL) algorithm for the problem. Furthermore, we generalize the SCL algorithm as a multi-task transfer learning strategy and propose the use of a restricted Boltzmann machine (RBM), a special type of probabilistic graphical models, as an implementation. We also give some preliminary theoretical analysis for the strategy by combining previous work on general transfer learning and multi-task learning.;Finally, we study the problem of co-reference resolution using another kind of graphical models, the conditional random field (CRF). We show that a previously proposed ranking approach, which produces state of the art results, can be viewed as a special case of the model. We go on to show how using a CRF allows us to easily incorporate other NLP tasks such as non-anaphoric identification and noun phrase boundary detection.
机译:传统的监督式机器学习需要针对感兴趣的特定问题标记数据。已经进行了许多尝试来减少这种需求,例如基于半监督学习的方法。近年来,人们开始考虑一种称为转移学习的新策略,其中使用来自旧问题(称为源任务)的标记数据来协助学习新的但相关的问题(目标任务)。在本论文中,我们主要考虑转移学习的极端情况,我们将其称为异构转移学习-源任务和目标任务的特征空间是不相交的。我们首先考虑跨语言文本分类任务,我们需要训练中文的分类器,但只有英文标签数据。我们针对该问题采用了结构对应学习(SCL)算法。此外,我们将SCL算法概括为一种多任务转移学习策略,并提出了一种使用受限Boltzmann机器(RBM)(一种特殊类型的概率图形模型)的实现方式。我们还通过结合先前在通用转移学习和多任务学习方面的工作,对该策略进行了初步的理论分析。最后,我们使用另一种图形模型(条件随机场(CRF))研究共参考分辨率的问题。 。我们表明,可以将先前提出的产生最新技术结果的排名方法视为模型的特殊情况。我们继续展示使用CRF如何使我们能够轻松地合并其他NLP任务,例如非回指识别和名词短语边界检测。

著录项

  • 作者

    Wei, Bin.;

  • 作者单位

    University of Rochester.;

  • 授予单位 University of Rochester.;
  • 学科 Computer Science.
  • 学位 Ph.D.
  • 年度 2011
  • 页码 100 p.
  • 总页数 100
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号