论文记录-2017-A review of affective computing: From unimodal analysis to multimodal fusion
论文内容
A review of affective computing: From unimodal analysis to multimodal fusion1
情感计算综述:从单模态分析到多模态融合
以下仅为作者阅读论文时的记录,学识浅薄,如有错误,欢迎指正。
摘要
-
Affective computing is an emerging interdisciplinary research field bringing together researchers and practitioners from various fields, ranging from artificial intelligence, natural language processing, to cognitive and social sciences.
情感计算是一个新兴的跨学科研究领域,汇集了来自人工智能、自然语言处理到认知和社会科学等各个领域的研究人员和从业者。 -
With the proliferation of videos posted online (e.g., on YouTube, Facebook, Twitter) for product reviews, movie reviews, political views, and more, affective computing research has increasingly evolved from conventional unimodal analysis to more complex forms of multimodal analysis.
随着在网上发布的有关产品评论、电影评论、政治观点等视频(例如在YouTube, Facebook, Twitter上)数量的激增,情感计算研究越来越从传统的单模态分析发展为更复杂的多模态分析形式。 -
This is the primary motivation behind our first of its kind, comprehensive literature review of the diverse field of affective computing.
这是我们对不同领域的情感计算进行全面文献综述的第一个主要动机。 -
Furthermore, existing literature surveys lack a detailed discussion of state of the art in multimodal affect analysis frameworks, which this review aims to address.
此外,现有的文献综述缺乏对多模态情感分析框架研究现状的详细讨论,本文旨在解决这些问题。 -
Multimodality is defined by the presence of more than one modality or channel, e.g., visual, audio, text, gestures, and eye gage.
多模态的定义是存在多个模态或通道,例如,视觉、音频、文本、姿态和眼距。 -
In this paper, we focus mainly on the use of audio, visual and text information for multimodal affect analysis, since around 90% of the relevant literature appears to cover these three modalities.
在本文中,我们主要关注使用音频、视觉和文本信息进行多模态情感分析,因为大约90%的相关文献似乎涵盖了这三种模式。 -
Following an overview of different techniques for unimodal affect analysis, we outline existing methods for fusing information from different modalities.
在概述了单模态情感分析的不同技术之后