Committee |
Date Time |
Place |
Paper Title / Authors |
Abstract |
Paper # |
SIP, IT, RCS |
2024-01-19 14:55 |
Miyagi |
(Primary: On-site, Secondary: Online) |
Dataset Generation System for Hand Gesture Recognition Using FMCW-MIMO Radar Katsuhisa Kashiwagi (Murata Manufacturing/Yokohama National Univ.), Koichi Ichige (Yokohama National Univ.) IT2023-72 SIP2023-105 RCS2023-247 |
In this paper, we propose a dataset generation system for hand gesture recognition using Frequency Modulated Continuous ... [more] |
IT2023-72 SIP2023-105 RCS2023-247 pp.229-234 |
PRMU, IPSJ-CVIM, IPSJ-DCC, IPSJ-CGVI |
2023-11-17 09:20 |
Tottori |
(Primary: On-site, Secondary: Online) |
Co-speech Gesture Generation with Variational Auto Encoder Shihichi Ka, Koichi Shinoda (Tokyo Tech) PRMU2023-29 |
Co-speech gesture generation is the study of generating gestures from speech. In prior works, deterministic methods lear... [more] |
PRMU2023-29 pp.74-79 |
BioX |
2023-10-13 10:20 |
Okinawa |
Nobumoto Ohama Memorial Hall |
Discrimination between Real and Generated Gestures of Speakers
-- An Attempt to Improve Generalization Performance in Unseen Generation Methods through Self-Supervised Learning -- Geng Mu (AGU), Naoshi Kaneko (TDU), Kazuhiko Sumi (AGU) BioX2023-67 |
Currently, discerning artificially generated misinformation is a critical societal challenge, with research progressing ... [more] |
BioX2023-67 pp.44-49 |
NC, IBISML, IPSJ-BIO, IPSJ-MPS [detail] |
2023-06-30 14:45 |
Okinawa |
OIST Conference Center (Primary: On-site, Secondary: Online) |
Diffusion model with MASKed input for generating gestures during dyadic conversation Yuya Okadome (TUS), Yutaka Nakamura (Riken) NC2023-19 IBISML2023-19 |
In a dyadic conversation scene, it is necessary to consider not only the behavior of one person but not also that of a c... [more] |
NC2023-19 IBISML2023-19 pp.121-128 |
PRMU, IPSJ-CVIM |
2023-05-18 14:15 |
Aichi |
(Primary: On-site, Secondary: Online) |
Streamable gesture generators for low-latency gesture generation Yuna Mitsubayashi, Naoshi Kaneko, Kazuhiko Sumi (Aoyama Gakuin Univ.) PRMU2023-4 |
Currently, conversational agents such as interactive robots are developing remarkably, and not only their dialogue respo... [more] |
PRMU2023-4 pp.16-21 |
PRMU, IPSJ-CVIM |
2023-05-18 14:30 |
Aichi |
(Primary: On-site, Secondary: Online) |
Discriminating between fake and real gestures in automatic gesture generation Geng Mu, Nosh Kaneko, Kazuhiko Sumi (AGU) PRMU2023-5 |
In recent years, gestures play a crucial role in communication with anthropomorphized agents and robots. The use of gest... [more] |
PRMU2023-5 pp.22-26 |
ICSS, IPSJ-SPT |
2023-03-13 12:30 |
Okinawa |
Okinawaken Seinenkaikan (Primary: On-site, Secondary: Online) |
Developing a Hand Gesture Recognition System for VR Devices Mei Suzuki, Ryo Iijima, Kazuki Nomoto (Waseda Univ.), Tetsushi Ohki (Shizuoka Univ./AIP), Tatsuya Mori (Waseda Univ./NICT/AIP) ICSS2022-48 |
We develop a hand gesture recognition scheme using finger movements to recognize the VR user wearing a head-mounted disp... [more] |
ICSS2022-48 pp.1-6 |
ICSS, IPSJ-SPT |
2023-03-13 13:10 |
Okinawa |
Okinawaken Seinenkaikan (Primary: On-site, Secondary: Online) |
sEMG-based Gesture Authentication for Smartwatch Akihisa Watanabe, Ryo Iijima, Tatsuya Mori (Waseda Univ.,) ICSS2022-50 |
We propose a gesture-based user recognition system using sEMG signals for smartwatches.
We designed the system to reali... [more] |
ICSS2022-50 pp.13-18 |
HCGSYMPO (2nd) |
2022-12-14 - 2022-12-16 |
Kagawa |
Onsite (Sunport Takamatsu) and Online (Primary: On-site, Secondary: Online) |
Relationship between linguistic expression and gesture by an avatar to explain a route on a map
-- Analyzing gaze of viewers -- Ayana Ogata, Koh Kakusho (K. G.), Takatsugu Kojima (SUMS) |
For information presentation with various kind of representation including characters, images, speech, etc., speech is o... [more] |
|
SWIM |
2022-11-26 16:55 |
Tokyo |
Kikai-Shinko-Kaikan Bldg. (Primary: On-site, Secondary: Online) |
Object Detection Using Gesture Recognition with YOLO
-- Application to Inventory Management -- Tsukasa Kudo (SIST) SWIM2022-27 |
In object detection using deep learning, it is necessary to prepare training data for each object, which is an applicati... [more] |
SWIM2022-27 pp.40-47 |
MVE |
2022-09-08 15:00 |
Tokyo |
(Primary: On-site, Secondary: Online) |
Gamified English Conversation Learning System with Characters
-- Gesture Scoring in Simulated Face-to-Face Communication Phase -- Naoki Matsumura, Tomoko Yonezawa (Kansai Univ.) MVE2022-11 |
In this study, we focused on speech proactivity evaluation in foreign language conversation with virtual characters in w... [more] |
MVE2022-11 pp.9-14 |
SWIM, SC |
2022-08-26 11:05 |
Online |
Held online |
Object Recognition by Fingertip Gestures Utilizing Transfer Learning of YOLOv5 Tsukasa Kudo (SIST) SWIM2022-12 SC2022-18 |
In object recognition for videos and images, when the target area is small, the area is firstly extracted by object dete... [more] |
SWIM2022-12 SC2022-18 pp.9-15 |
WIT, HI-SIGACI |
2021-12-08 14:05 |
Online |
Online |
Effect of Adaptative Gesture-Interface to Persons with Severe Physical Disabilities Kazuyuki Itoh, Tsuyoshi Nakayama, Takafumi Kinoshita (NRCD), Ikushi Yoda (AIST), Katsuhiro Mizuno (NCNP) WIT2021-32 |
their daily living operations at home. For a muscular dystrophy patient, the system is installed instead of keyboard ope... [more] |
WIT2021-32 pp.1-6 |
HCS |
2020-10-31 15:30 |
Online |
Online |
Attractiveness and impressions of attractive gestures and postures Saika Yoshida, Nobuyuki Watanabe (KIT) HCS2020-42 |
(To be available after the conference date) [more] |
HCS2020-42 pp.2-5 |
PRMU |
2020-10-10 10:30 |
Online |
Online |
Analysis and Classification of Gestures in TED Talks Hitoshi Teshima (Kyushu Univ.), Naoki Wake (Microsoft), Diego Thomas (Kyushu Univ.), Yuta Nakashima (Osaka Univ.), Hiroshi Kawasaki (Kyushu Univ.), Katsushi Ikeuchi (Microsoft) PRMU2020-35 |
Most of automatic gesture generation methods have focused on generating beat gestures only. Among these methods, many ge... [more] |
PRMU2020-35 pp.104-109 |
PRMU, IPSJ-CVIM |
2020-03-16 15:20 |
Kyoto |
(Cancelled but technical report was issued) |
Investigation of Real-Time 3D Head Position Estimation for Gesture Interface for People with Motor Dysfunction Yuki Ozawa (Tsukuba Univ./AIST), Ikushi Yoda (AIST) PRMU2019-75 |
We have developed a gesture interface for people with motor dysfunction using the RGB-D camera. The interface has nine m... [more] |
PRMU2019-75 pp.53-58 |
HCS |
2020-03-05 14:20 |
Online |
Online Meeting |
[Online] The Relationship between Gestures and Words by a Coach who Teaches Advanced Walking Skill Masatoshi Yamada (Tokoha Univ.), Daisuke Sato (SATO SPEED), Sayaka Tohyama (Shizuoka Univ.), Takeshi Matsumura (Tokoha Univ.), Yugo Takeuchi (Shizuoka Univ.) HCS2019-98 |
The purpose of this study is to clarify the relationship between gestures and words by a coach who teaches advanced walk... [more] |
HCS2019-98 pp.77-82 |
HIP |
2019-12-20 11:15 |
Miyagi |
RIEC, Tohoku University |
On the effect of gestures on the memory of speech content Seiji Mio (Niigata Univ), Ryosuke Niimi (Niigata U) HIP2019-76 |
It has been claimed that gestures promote human speech understanding, while it is yet unclear to what extent gestures co... [more] |
HIP2019-76 pp.61-64 |
HCGSYMPO (2nd) |
2019-12-11 - 2019-12-13 |
Hiroshima |
Hiroshima-ken Joho Plaza (Hiroshima) |
Identification of a Learning State Using Face Information from Kinect Takashi Ito (Aoyama Gakuin Univ.), Hikari Fukai, Kenichi Takahashi, Tomoko Kajiyama (Hiroshima City Univ.) |
In actual classes, it is very important for teachers to grasp learning states of students. However, it is difficult to m... [more] |
|
TL |
2019-03-18 11:45 |
Tokyo |
Waseda University |
Study of Real-Time Human Behavior Recognition Using Wearable Device
-- Gesture Recognition by Finger Motion Sensing -- Kei Nakada, Atsushi Ito, Hiroyuki Hatano (Utsunomiya Univ.) TL2018-55 |
According to the spread IoT, various devices are connected to the network and exist around our life space. In order to u... [more] |
TL2018-55 pp.27-31 |