首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >Exploring Semi-Supervised Methods for Labeling Support in Multimodal Datasets
【2h】

Exploring Semi-Supervised Methods for Labeling Support in Multimodal Datasets

机译:探索多模式数据集中标签支持的半监督方法

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Working with multimodal datasets is a challenging task as it requires annotations which often are time consuming and difficult to acquire. This includes in particular video recordings which often need to be watched as a whole before they can be labeled. Additionally, other modalities like acceleration data are often recorded alongside a video. For that purpose, we created an annotation tool that enables to annotate datasets of video and inertial sensor data. In contrast to most existing approaches, we focus on semi-supervised labeling support to infer labels for the whole dataset. This means, after labeling a small set of instances our system is able to provide labeling recommendations. We aim to rely on the acceleration data of a wrist-worn sensor to support the labeling of a video recording. For that purpose, we apply template matching to identify time intervals of certain activities. We test our approach on three datasets, one containing warehouse picking activities, one consisting of activities of daily living and one about meal preparations. Our results show that the presented method is able to give hints to annotators about possible label candidates.
机译:使用多峰数据集是一项具有挑战性的任务,因为它需要注释,这些注释通常很耗时且难以获取。特别是其中包括通常需要整体观看的视频记录,然后才能对其进行标记。另外,诸如加速度数据之类的其他形式通常与视频一起记录。为此,我们创建了一个注释工具,可以对视频和惯性传感器数据的数据集进行注释。与大多数现有方法相比,我们专注于半监督标签支持以推断整个数据集的标签。这意味着在标记少量实例之后,我们的系统能够提供标记建议。我们旨在依靠腕戴式传感器的加速度数据来支持视频记录的标签。为此,我们应用模板匹配来识别某些活动的时间间隔。我们在三个数据集上测试了我们的方法,一个数据集包含仓库拣选活动,一个包含日常生活活动,另一个包含膳食准备。我们的结果表明,提出的方法能够为注释者提供有关可能的候选标签的提示。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号