首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Measuring Sentiment Annotation Complexity of Text
【24h】

Measuring Sentiment Annotation Complexity of Text

机译:测量情绪注释文本的复杂性

获取原文

摘要

The effort required for a human annota-tor to detect sentiment is not uniform for all texts, irrespective of his/her expertise. We aim to predict a score that quantifies this effort, using linguistic properties of the text. Our proposed metric is called Sentiment Annotation Complexity (SAC). As for training data, since any direct judgment of complexity by a human annota-tor is fraught with subjectivity, we rely on cognitive evidence from eye-tracking. The sentences in our dataset are labeled with SAC scores derived from eye-fixation duration. Using linguistic features and annotated SACs, we train a regressor that predicts the SAC with a best mean error rate of 22.02% for five-fold cross-validation. We also study the correlation between a human annotator's perception of complexity and a machine's confidence in polarity determination. The merit of our work lies in (a) deciding the sentiment annotation cost in, for example, a crowdsourcing setting, (b) choosing the right classifier for sentiment prediction.
机译:无论他/她的专业知识如何,所有文本都不统一,对人类Annota-Tor检测情绪所需的努力并不统一。我们的目标是使用文本的语言属性来预测量化这项努力的分数。我们所提出的指标称为情绪注释复杂性(SAC)。至于培训数据,由于人类Annota-tor对复杂性的任何直接判断都充满了主观性,因此我们依赖于追踪的认知证据。我们数据集中的句子标有来自眼固定持续时间的SAC分数。使用语言特征和带注释的囊,我们训练一个回归,以预测囊的最佳误差率为22.02%,对于五倍交叉验证。我们还研究了人类注入者对复杂性的看法与机器对极性决定的信心之间的相关性。我们的工作的优点在于(a)决定情绪注释成本,例如,众包设置,(b)选择右分类器进行情绪预测。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号