Committee |
Date Time |
Place |
Paper Title / Authors |
Abstract |
Paper # |
EA |
2024-05-22 14:55 |
Online |
Online |
Environmental sound synthesis and creation of dataset using vocal imitations Yuki Okamoto (Ritsumeikan Univ.), Keisuke Imoto (Doshisha Univ.), Shinnosuke Takamichi (The Univ. of Tokyo/Keio Univ.), Ryotaro Nagase, Takahiro Fukumori, Yoichi Yamashita (Ritsumeikan Univ.) |
[more] |
|
EA |
2024-05-22 15:45 |
Online |
Online |
Audio-change Captioning to Explain Machine-sound Anomalies Shunsuke Tsubaki (Doshisha Univ./Hitachi), Yohei Kawaguchi, Tomoya Nishida (Hitachi), Keisuke Imoto (Doshisha Univ.), Yuki Okamoto (Ritsumeikan Univ./Hitachi), Kota Dohi, Takashi Endo (Hitachi) |
[more] |
|
EA |
2024-05-22 16:10 |
Online |
Online |
Incremental Learning for Joint analysis of Acoustic Scenes and Sound Events Kaori Inoue, Yuka Fukumoto, Naoki Koga, Keisuke Imoto (Doshisha Univ.) |
[more] |
|
SIP, SP, EA, IPSJ-SLP [detail] |
2024-03-01 14:25 |
Okinawa |
(Primary: On-site, Secondary: Online) |
[Invited Talk]
Getting Started With Environmental Sound Analysis and Synthesis Keisuke Imoto (Doshisha Univ.) EA2023-118 SIP2023-165 SP2023-100 |
With the development of machine learning techniques and improvements in computing power, audio analysis and synthesis te... [more] |
EA2023-118 SIP2023-165 SP2023-100 p.333 |
SP, IPSJ-SLP, EA, SIP [detail] |
2023-02-28 15:20 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Single-channel environmental sound classification using distance-based sound separation Ryoya Ogura, Sayaka Shiota (Tokyo Metropolitan Univ.), Keisuke Imoto (Doshisha Univ.), Hitoshi Kiya (Tokyo Metropolitan Univ.) |
[more] |
|
SP, IPSJ-SLP, EA, SIP [detail] |
2023-02-28 15:55 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Self-Supervised Learning With Spatial Audio-Visual Recording for Sound Event Localization and Detection Yoto Fujita (Kyoto Univ.), Yoshiaki Bando (AIST), Keisuke Imoto (Doshisha Univ./AIST), Masaki Onihsi (AIST), Yoshii Kazuyoshi (Kyoto Univ.) EA2022-89 SIP2022-133 SP2022-53 |
This paper describes an unsupervised pre-training method for sound event localization and detection (SELD) on multi-chan... [more] |
EA2022-89 SIP2022-133 SP2022-53 pp.78-82 |
SP, IPSJ-SLP, EA, SIP [detail] |
2023-02-28 16:15 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images Hien Ohnaka (NITTC), Shinnosuke Takamichi (UT), Keisuke Imoto (DU), Yuki Okamoto (Rits), Kazuki Fujii, Hiroshi Saruwatari (UT) EA2022-90 SIP2022-134 SP2022-54 |
(To be available after the conference date) [more] |
EA2022-90 SIP2022-134 SP2022-54 pp.83-88 |
SP, IPSJ-SLP, EA, SIP [detail] |
2023-03-01 09:10 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Joint analysis of acoustic scenes and sound events based on semi-supervised learning Ami Igarashi, Shunsuke Tsubaki, Keisuke Imoto (DU) EA2022-103 SIP2022-147 SP2022-67 |
(To be available after the conference date) [more] |
EA2022-103 SIP2022-147 SP2022-67 pp.165-170 |
EA, SIP, SP, IPSJ-SLP [detail] |
2022-03-02 13:25 |
Okinawa |
(Primary: On-site, Secondary: Online) |
[Poster Presentation]
Multi-channel missing signals recovery using autoencoder for acoustic scene classification Yuki Shiroma, Yuma Kinoshita (Tokyo Metro. Univ.), Keisuke Imoto (Doshisha Univ.), Sayaka Shiota, Nobutaka Ono, Hitoshi Kiya (Tokyo Metro. Univ.) EA2021-86 SIP2021-113 SP2021-71 |
[more] |
EA2021-86 SIP2021-113 SP2021-71 pp.140-145 |
SP, EA, SIP |
2020-03-02 13:00 |
Okinawa |
Okinawa Industry Support Center (Cancelled but technical report was issued) |
Learning of Classification Models using Emotion-specific Soft Labels for Speech Emotion Recognition Mayuko Ozawa, Keisuke Imoto, Ryosuke Yamanishi, Yoichi Yamashita (Ritsumeikan Univ.) EA2019-107 SIP2019-109 SP2019-56 |
[more] |
EA2019-107 SIP2019-109 SP2019-56 pp.35-40 |
SP, EA, SIP |
2020-03-03 09:00 |
Okinawa |
Okinawa Industry Support Center (Cancelled but technical report was issued) |
Evaluation of vocal personality and expression for speech synthesized by non-parallel voice conversion with narrative speech Ryotaro Nagase, Keisuke Imoto, Ryosuke Yamanishi, Yoichi Yamashita (Ritsumeikan Univ.) EA2019-138 SIP2019-140 SP2019-87 |
In the technology of voice conversion, reproduction of emotion and intonation, pause is one of the research issues. Howe... [more] |
EA2019-138 SIP2019-140 SP2019-87 pp.213-218 |
EA |
2019-12-12 15:50 |
Fukuoka |
Kyushu Inst. Tech. |
Experimental study of robustness to spatial information in spatial cepstrum Taiga Kawamura, Ryoichi Miyazaki (NITTC), Keisuke Imoto (Ritsumeikan Univ.) EA2019-73 |
In this study, we investigate how the behavior of the spatial cepstrum changes by the sound source and the movement of t... [more] |
EA2019-73 pp.51-56 |
SP |
2018-08-27 11:35 |
Kyoto |
Kyoto Univ. |
[Poster Presentation]
A Study on Representation of Speaker Information for DNN Speech Synthesis Lin Yuhan, Keisuke Imoto, Masahiro Niitsuma, Ryosuke Yamanishi, Yoichi Yamashita (Ritsumeikan Univ.) SP2018-25 |
Recent studies have shown that DNN speech synthesis can generate natural synthesized speech than HMM-based speech synthe... [more] |
SP2018-25 pp.15-18 |
SP, IPSJ-SLP (Joint) |
2018-07-26 13:30 |
Shizuoka |
Sago-Royal-Hotel (Hamamatsu) |
[Invited Talk]
Fundamentals and Research Trends on Acoustic Event and Scene Analysis Keisuke Imoto (Ritsumeikan Univ.) SP2018-16 |
In this report, we review the fundamentals and research trends of acoustic event and scene analysis, which analyzes vari... [more] |
SP2018-16 pp.1-6 |
SIP, EA, SP, MI (Joint) [detail] |
2018-03-20 09:00 |
Okinawa |
|
[Poster Presentation]
Performance evaluation of unknown sound clustering for indoor-environmental sound classification based on self-generated acoustic model Sakiko Mishima, Yukoh Wakabayashi, Takahiro Fukumori, Keisuke Imoto, Masato Nakayama, Takanobu Nishiura (Ritsumeikan Univ.) EA2017-152 SIP2017-161 SP2017-135 |
Indoor-environmental sound classification is useful for surveillance systems which monitor the situations in the dark an... [more] |
EA2017-152 SIP2017-161 SP2017-135 pp.277-280 |
SP |
2017-08-30 11:00 |
Kyoto |
Kyoto Univ. |
[Poster Presentation]
Emotion Recognition in Speech Using Deep Neural Network Li ShiChuan, Tomoki Ishikawa, Masahiro Niitsuma, Keisuke Imoto, Yoichi Yamashita (Ritsumeikan Univ.) SP2017-24 |
Speech conveys not only linguistic information but also paralinguistic and non-linguistic information such as emotions,a... [more] |
SP2017-24 pp.25-28 |