首页> 外文会议>International Conference on Computational Intelligence and Knowledge Economy >A Comparative Study of Deep Transfer Learning Techniques for Cultural (Aeta) Dance Classification utilizing Skeleton-Based Choreographic Motion Capture Data
【24h】

A Comparative Study of Deep Transfer Learning Techniques for Cultural (Aeta) Dance Classification utilizing Skeleton-Based Choreographic Motion Capture Data

机译:利用基于骨架的编排运动捕获数据的文化(AETA)舞蹈分类深度转移学习技术的比较研究

获取原文

摘要

The advancement of motion-sensing technology and depth cameras has led to a vast opportunity in motion analysis and monitoring applications, kinesiology analysis, and safeguarding intangible cultural heritage (ICH). A technique that allows a computer to understand human behavior is necessary to analyze and identify the motion using the motion capture data. The integration of motion sensing technology such as markerless motion capture devices, inertial sensors, and deep learning techniques gives an innovative approach to recording, analyzing, and visually recognizing human choreographic motion. Convolutional Neural Network (CNN) is one of the best-known techniques in learning patterns from images and videos and is most widely used among deep learning architectures for vision applications. This study explored different CNN architecture to determine the best prediction classifier based on its performances, such as VGG19, InceptionV3, and MobileNetV2. This study aims to perform an image classification approach of one of the Philippines’ cultural dances, Aeta dances, utilizing skeleton-based motion capture data using CNN. The test results were assessed based on the generated training accuracy and evaluation of the loss function to assess the models’ overall efficiency. VGG19 produced the highest model cultural dance classification accuracy among the three architectures, which resulted in 98.68% compared to InceptionV3 and MobileNetV2. Thus, the VGG19 model illustrates the optimal transfer learning result implies the best fit model than InceptionV3 and MobileNetV2.
机译:运动传感技术和深度摄像机的进步导致了运动分析和监测应用,运动学分析和保护无形文化遗产(ICH)的绝大机会。允许计算机理解人类行为的技术是使用运动捕获数据分析和识别运动的必要方法。运动传感技术的整合如无价值运动捕获装置,惯性传感器和深度学习技术,提供了一种创新的记录,分析和视觉识别人类编排运动的创新方法。卷积神经网络(CNN)是来自图像和视频的学习模式中最着名的技术之一,并且最广泛地使用视觉应用的深度学习架构。本研究探索了不同的CNN架构,以基于其性能来确定最佳预测分类器,例如VGG19,Inceptionv3和MobileNetv2。本研究旨在利用CNN利用基于骨架的运动捕获数据来执行菲律宾文化舞蹈AETA舞蹈的图像分类方法。基于产生的训练准确性和评估损失功能来评估测试结果,以评估模型的整体效率。 VGG19在三个架构中产生了最高的模型文化舞蹈分类准确性,与Inceptionv3和MobileNetv2相比,98.68%导致了98.68%。因此,VGG19模型说明了最佳转移学习结果意味着比Inceptionv3和MobileNetv2的最佳拟合模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号