...
首页> 外文期刊>Neurocomputing >Learning fashion compatibility across categories with deep multimodal neural networks
【24h】

Learning fashion compatibility across categories with deep multimodal neural networks

机译:学习时尚兼容性跨多模态神经网络的类别

获取原文
获取原文并翻译 | 示例

摘要

Fashion compatibility is a subjective sense of human for relationships between fashion items, which is essential for fashion recommendation. Recently, it increasingly attracts more and more attentions and has become a very hot research topic. Learning fashion compatibility is a challenging task, since it needs to consider plenty of factors about fashion items, such as color, texture, style and functionality. Unlike low-level visual compatibility (e.g., color, texture), high-level semantic compatibility (e.g., style, functionality) cannot be handled purely based on fashion images. In this paper, we propose a novel multimodal framework to learn fashion compatibility, which simultaneously integrates both semantic and visual embeddings into a unified deep learning model. For semantic embeddings, a multilayered Long Short-Term Memory (LSTM) is employed for discriminative semantic representation learning, while a deep Convolutional Neural Network (CNN) is used for visual embeddings. A fusion module is then constructed to combine semantic and visual information of fashion items, which equivalently transforms semantic and visual spaces into a latent feature space. Furthermore, a new triplet ranking loss with compatible weights is introduced to measure fine-grained relationships between fashion items, which is more consistent with human feelings on fashion compatibility in reality. Extensive experiments conducted on Amazon fashion dataset demonstrate the effectiveness of the proposed method for learning fashion compatibility, which outperforms the state-of-the-art approaches. (C) 2019 Elsevier B.V. All rights reserved.
机译:时尚兼容性是时尚物品之间关系的主观感受,这对于时尚推荐至关重要。最近,它越来越多地吸引了越来越多的关注,并已成为一个非常热门的研究课题。学习时尚兼容性是一个具有挑战性的任务,因为它需要考虑一些关于时尚物品的因素,例如颜色,纹理,风格和功能。与低级视觉兼容性(例如,颜色,纹理),高电平语义兼容性(例如,样式,功能)不能纯粹基于时尚图像来处理。在本文中,我们提出了一种新的多模态框架来学习时尚兼容性,这同时将语义和视觉嵌入的同时集成到统一的深度学习模型中。对于语义嵌入,使用多层的长短期记忆(LSTM)用于鉴别语义表示学习,而深度卷积神经网络(CNN)用于视觉嵌入。然后构建一个融合模块以组合时尚物品的语义和视觉信息,其等效地将语义和视觉空间转换为潜在特征空间。此外,引入了一种具有兼容权重的新的三联排名损失,以测量时尚物品之间的细粒度关系,这与人类对时尚兼容性的人类感受更符合。在亚马逊时装数据集上进行的广泛实验证明了提出的学习时尚兼容性方法的有效性,这优于最先进的方法。 (c)2019 Elsevier B.v.保留所有权利。

著录项

  • 来源
    《Neurocomputing 》 |2020年第28期| 237-246| 共10页
  • 作者单位

    Southwest Jiaotong Univ Sch Informat Sci & Technol Chengdu Peoples R China;

    Southwest Jiaotong Univ Sch Informat Sci & Technol Chengdu Peoples R China;

    Southwest Jiaotong Univ Sch Informat Sci & Technol Chengdu Peoples R China;

    Southwest Jiaotong Univ Sch Informat Sci & Technol Chengdu Peoples R China;

    Southwest Jiaotong Univ Sch Informat Sci & Technol Chengdu Peoples R China;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Fashion compatibility; Deep learning; Neural networks; Multimodal;

    机译:时尚兼容性;深度学习;神经网络;多式联版;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号