Presentation 2019-12-04
LSTMの前処理としての特徴選択の有効性
Shiori Koga, Tsunenori Mine, Sachio Hirokawa,
PDF Download Page PDF download Page Link
Abstract(in Japanese) (See Japanese page)
Abstract(in English) Among RNNs, especially LSTM is capable of long-term memory and can be expected to acquire information including better contextual information, and has been used for various problems including classification problems. In such a classification problem, various attempts have been made to improve the accuracy of classification by selecting features such as words that contribute to classification from among the words that make up sentences. However, since LSTM attaches importance to the context, no attempt has been made to select words in LSTM input as far as we know. In this study, we show the effectiveness of using feature selection as pre-processing of LSTM input, using Word2Vec and BERT as word expressions in LSTM input sentences. The results show that the classification accuracy improves when feature selection is performed.
Keyword(in Japanese) (See Japanese page)
Keyword(in English) text classification / LSTM / feature selection / Word2Vec / BERT
Paper # NLC2019-31
Date of Issue 2019-11-27 (NLC)

Conference Information
Committee NLC / IPSJ-NL / SP / IPSJ-SLP
Conference Date 2019/12/4(3days)
Place (in Japanese) (See Japanese page)
Place (in English) NHK Science & Technology Research Labs.
Topics (in Japanese) (See Japanese page)
Topics (in English) The 6th Natural Language Processing Symposium & The 21th Spoken Language Symposium
Chair Takeshi Sakaki(Hottolink) / / Hisashi Kawai(NICT)
Vice Chair Mitsuo Yoshida(Toyohashi Univ. of Tech.) / Kazutaka Shimada(Kyushu Inst. of Tech.) / / Akinobu Ri(Nagoya Inst. of Tech.)
Secretary Mitsuo Yoshida(Ryukoku Univ.) / Kazutaka Shimada(NTT) / / Akinobu Ri(Kyoto Univ.) / (Waseda Univ.)
Assistant Takeshi Kobayakawa(NHK) / Hiroki Sakaji(Univ. of Tokyo) / / Tomoki Koriyama(Univ. of Tokyo) / Yusuke Ijima(NTT)

Paper Information
Registration To Technical Committee on Natural Language Understanding and Models of Communication / Special Interest Group on Natural Language / Technical Committee on Speech / Special Interest Group on Spoken Language Processing
Language JPN-ONLY
Title (in Japanese) (See Japanese page)
Sub Title (in Japanese) (See Japanese page)
Title (in English)
Sub Title (in English)
Keyword(1) text classification
Keyword(2) LSTM
Keyword(3) feature selection
Keyword(4) Word2Vec
Keyword(5) BERT
1st Author's Name Shiori Koga
1st Author's Affiliation Kyushu University(Kyushu Univ.)
2nd Author's Name Tsunenori Mine
2nd Author's Affiliation Kyushu University(kyushu Univ.)
3rd Author's Name Sachio Hirokawa
3rd Author's Affiliation Kyushu University(Kyushu Univ.)
Date 2019-12-04
Paper # NLC2019-31
Volume (vol) vol.119
Number (no) NLC-320
Page pp.pp.13-18(NLC),
#Pages 6
Date of Issue 2019-11-27 (NLC)