【24h】

Deep Bayesian Natural Language Processing

机译:深贝叶斯自然语言处理

获取原文
获取外文期刊封面目录资料

摘要

This introductory tutorial addresses the advances in deep Bayesian learning for natural language with ubiquitous applications ranging from speech recognition (Saon and Chien, 2012; Chan et al., 2016) to document summarization (Chang and Chien, 2009), text classification (Blei et al., 2003; Zhang et al., 2015), text segmentation (Chien and Chueh, 2012), information cxtraction (Narasimhan et al., 2016), image caption generation (Vinyals et al., 2015; Xu et al., 2015), sentence generation (Li et al., 2016), dialogue control (Zhao and Eskenazi, 2016), sentiment classification, recommendation system, question answering (Sukhbaatar et al., 2015) and machine translation (Bahdanau el al., 2014), to name a few. Traditionally, "deep learning" is taken to be a learning process where the inference or optimization is based on the real-valued deterministic model. The "semantic structure" in words, sentences, entities, actions and documents drawn from a large vocabulary may not be well expressed or correctly optimized in mathematical logic or computer programs. The "distribution function" in discrete or continuous latent variable model for natural language may not be properly decomposed or estimated. This tutorial addresses the fundamentals of statistical models and neural networks, and focus on a series of advanced Bayesian models and deep models including hierarchical Dirichlet process (Teh et al., 2006), Chinese restaurant process (Blei et al., 2010), hierarchical Pitman-Yor process (Teh, 2006), Indian buffet process (Ghahra-mani and Griffiths, 2005), recurrent neural network (Mikolov et al., 2010; Van Den Oord et al., 2016), long short-term memory (Hochreiter and Schmidhuber, 1997; Cho et al., 2014), sequence-to-sequence model (Sutskever et al., 2014), variational auto-encoder (Kingma and Welling, 2014), generative adversarial network (Goodfellow et al., 2014) , attention mechanism (Chorowski et al., 2015; Seo et al., 2016), memory-augmented neural network (Graves et al., 2014; Sukhbaatar et al., 2015) , skip neural network (Campos et al., 2018), stochastic neural network (Bengio et al., 2014; Miao et al., 2016), predictive state neural network (Downey et al., 2017) and policy neural network (Mnih et al., 2015; Yu et al., 2017). We present how these models are connected and why they work for a variety of applications on symbolic and complex patterns in natural language. The variational inference and sampling method are formulated to tackle the optimization for complicated models (Rezende et al., 2014). The word and sentence cmbcddings, clustering and co-clustering arc merged with linguistic and semantic constraints. A series of case studies and domain applications are presented to tackle different issues in deep Bayesian processing, learning and understanding. At last, we will point out a number of directions and outlooks for future studies.
机译:本入门教程介绍了针对自然语言的深度贝叶斯学习的进展,其应用范围广泛,从语音识别(Saon和Chien,2012; Chan等,2016)到文档摘要(Chang和Chien,2009),文本分类(Blei等)。等人,2003; Zhang等人,2015),文本分割(Chien和Chueh,2012),信息截取(Narasimhan等人,2016),图像标题生成(Vinyals等人,2015; Xu等人, 2015),句子生成(Li等,2016),对话控制(Zhao和Eskenazi,2016),情感分类,推荐系统,问答(Sukhbaatar等,2015)和机器翻译(Bahdanau el等,2014) ),仅举几例。传统上,“深度学习”被视为一种学习过程,其中推理或优化是基于实值确定性模型的。从大量词汇中提取的单词,句子,实体,动作和文档中的“语义结构”可能无法在数学逻辑或计算机程序中很好地表达或正确优化。自然语言的离散或连续潜变量模型中的“分布函数”可能无法正确分解或估计。本教程介绍了统计模型和神经网络的基础知识,并着重介绍了一系列高级贝叶斯模型和深层模型,包括层次化Dirichlet过程(Teh等人,2006年),中餐厅过程(Blei等人,2010年),层次结构Pitman-Yor过程(Teh,2006年),印度自助过程(Ghahra-mani和Griffiths,2005年),递归神经网络(Mikolov等人,2010年; Van Den Oord等人,2016年),长期短期记忆( Hochreiter和Schmidhuber,1997; Cho等,2014),序列到序列模型(Sutskever等,2014),变分自动编码器(Kingma和Welling,2014),生成对抗网络(Goodfellow等, 2014年),注意力机制(Chorowski等人,2015年; Seo等人,2016年),记忆增强神经网络(Graves等人,2014年; Sukhbaatar等人,2015年),跳过神经网络(Campos等人。 ,2018),随机神经网络(Bengio等,2014; Miao等,2016),预测状态神经网络(Downey等,2017)和政策神经网络(Mnih et al。,2015; Yu等人,2017)。我们将介绍这些模型如何连接以及它们为何能以自然语言在符号和复杂模式下用于各种应用程序。提出了变分推理和抽样方法来解决复杂模型的优化问题(Rezende等,2014)。单词和句子cmbcddings,聚类和共聚结合了语言和语义约束。提出了一系列案例研究和领域应用程序,以解决深度贝叶斯处理,学习和理解中的不同问题。最后,我们将指出一些未来研究的方向和前景。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号