首页> 外国专利> METHOD OF PROCESSING VIDEO FOR DETERMINING EMOTION OF A PERSON

METHOD OF PROCESSING VIDEO FOR DETERMINING EMOTION OF A PERSON

机译:用于确定一个人情绪的视频方法

摘要

The multi-modal emotion recognition method using artificial intelligence of the present description is an emotion recognition method that processes an image to determine a person's emotional state, and provides an image and a voice expressing the appearance of a person, wherein the image is the first comprising an imaging unit, a second imaging unit immediately following the first imaging unit, and a third imaging unit immediately following the second imaging unit; The first image unit processes the first image unit to determine the emotional state of the person, wherein the first image unit shows the person's face and at least one hand, and the at least one hand characterized in that no part of the face overlaps; and the second image unit processes the second image unit to determine the emotional state of the person, wherein the second image unit shows the person's face and at least one hand, and the at least one hand is the person Including, wherein the processing of the first image unit includes at least one of the first image unit to determine whether the at least one hand covers the face of the person processing a frame of ; finding a first facial element of the person in the at least one frame of the first image unit; and in a state in which the first facial element is located, the at least one obtaining first facial feature data of the first image unit based on the shape of the first facial element shown in the frame of processing the audio data of the first image unit to obtain determining a state; wherein the processing of the second image unit includes processing at least one frame of the second image unit to determine whether the face of the person is covered by at least one hand. determining whether or not the at least one hand covers the face of the person in the second image unit; and finding the first facial element of the person in at least one frame of the second image unit. obtaining the first facial feature data of the first image unit based on the shape of the first face element shown in the at least one frame of the second image unit in a state where the first facial element is located step, and the audio of the second image unit to obtain the voice characteristic data based on the human voice characteristics in the second image unit processing data, the first facial feature data of the second image unit, the audio feature data of the second image unit, and additional data indicating a location where at least one hand covers a part of the person's face and determining the emotional state of the person for the second image unit based on the plurality of data included therein.
机译:使用本描述的人工智能的多模态情绪识别方法是一种情绪识别方法,其处理图像以确定一个人的情绪状态,并提供表示人类外观的图像和语音,其中图像是第一个包括成像单元,紧接在第一成像单元之后的第二成像单元,以及紧接在第二成像单元之后的第三成像单元;第一图像单元处理第一图像单元以确定人的情绪状态,其中第一图像单元示出了人的脸部和至少一只手,并且至少一只手表征在于面部的任何部分都没有;第二图像单元处理第二图像单元以确定人的情绪状态,其中第二图像单元示出了人的脸部和至少一只手,并且至少一只手是包括的人,其中第一图像单元包括第一图像单元中的至少一个,以确定至少一只手是否覆盖处理框架的人的面部;在第一图像单元的至少一帧中找到人的第一面部元素;并且,在第一面部元件所在的状态下,基于在处理第一图像单元的音频数据的帧中所示的第一面部元件的形状来获得第一图像单元的第一面部特征数据的至少一个获得确定状态;其中,第二图像单元的处理包括处理第二图像单元的至少一帧以确定人的面是否覆盖至少一只手。确定至少一只手是否覆盖了第二图像单元中的人的面;并在第二图像单元的至少一帧中找到人的第一面部元件。基于第一面部元件的第二图像单元的至少一帧中所示的第一面元件的形状获得第一图像单元的第一面部特征数据,其中第一面部元件的步骤和第二个图像单元的状态。图像单元基于第二图像单元处理数据中的人类语音特性获得语音特征数据,第二图像单元的第一面部特征数据,第二图像单元的音频特征数据,以及指示位置的附加数据至少一只手覆盖了人的面部的一部分并基于其中包括的多个数据确定第二图像单元的人的情绪状态。

著录项

  • 公开/公告号KR102290186B1

    专利类型

  • 公开/公告日2021-08-17

    原文格式PDF

  • 申请/专利权人 주식회사 제네시스랩;

    申请/专利号KR20200081613

  • 发明设计人 유대훈;이영복;

    申请日2020-07-02

  • 分类号G06N5/04;G06K9;G10L15/02;

  • 国家 KR

  • 入库时间 2022-08-24 22:17:27

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号