Paper Abstract and Keywords |
Presentation |
2016-12-20 10:20
Constructing a Japanese multimodal corpus from emotional monologues and dialogues Nurul Lubis (NAIST), Randy Gomez (HRI), Sakriani Sakti (NAIST), Keisuke Nakamura (HRI), Koichiro Yoshino, Satoshi Nakamura (NAIST), Kazuhiro Nakadai (HRI) SP2016-51 |
Abstract |
(in Japanese) |
(See Japanese page) |
(in English) |
To fully incorporate emotion into human-computer interaction, rich sets of labeled emotional data is prerequisite. However, in Japanese, the majority of the existing emotion database is still limited to unimodal and bimodal corpora. To allow more complete observation of emotion occurrences, we construct the first audio-visual emotion corpora in Japanese, collected from 14 native speakers. Preliminary speech emotion recognition experiments on the corpus and achieved an accuracy of 61.42% for five classes of emotion. |
Keyword |
(in Japanese) |
(See Japanese page) |
(in English) |
multimodal / emotion / corpus / Japanese / / / / |
Reference Info. |
IEICE Tech. Rep., vol. 116, no. 378, SP2016-51, pp. 9-10, Dec. 2016. |
Paper # |
SP2016-51 |
Date of Issue |
2016-12-13 (SP) |
ISSN |
Print edition: ISSN 0913-5685 Online edition: ISSN 2432-6380 |
Copyright and reproduction |
All rights are reserved and no part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including photocopy, recording, or any information storage and retrieval system, without permission in writing from the publisher. Notwithstanding, instructors are permitted to photocopy isolated articles for noncommercial classroom use without fee. (License No.: 10GA0019/12GB0052/13GB0056/17GB0034/18GB0034) |
Notes on Review |
This article is a technical report without peer review, and its polished version will be published elsewhere. |
Download PDF |
SP2016-51 |
|