SP Subject Index 2001

{SP2001-1} S. Ding, M. Otsuka, M. Ashizawa, T. Niitsuma and K. Sugai,
``Blind source separation of real-world acoustic signals based on ICA in time-frequency-domain,''
IEICE, SP2001-1, pp.1--8, Apr. 2001.
{ Blind source separation(BSS), Independent component analysis(ICA), Deconvolution, FastICA, Permutation Problem, Separation of speech signals }

{SP2001-2} T. Kawamura, H. Saruwatari, and K. Shikano,
``Blind Source Separation Based on Fast-Convergence Algorithm with ICA and Beamforming,''
IEICE, SP2001-2, pp.9--16, Apr. 2001.
{ Blind Source Separation, Fast-Convergence, ICA, Beamforming, Reverberation }

{SP2001-3} T. Muto and M. Sugiyama,
``Model-based Voice Decomposition Method under Time Constraint,''
IEICE, SP2001-3, pp.17--24, Apr. 2001.
{ Auditory Scene Analysis, Segregation, Combinatorial Optimization }

{SP2001-4} M. Nakamura, T. Nishiura, A. Lee, H. Saruwatari and K. Shikano,
``Talker Localization on Autonomous Mobile Robot Using a Microphone Array,''
IEICE, SP2001-4, pp.25--32, Apr. 2001.
{ Microphone Array, Autonomous Mobile Robot, Sound source localization, HMM, Sound source discrimination }

{SP2001-5} S. Onishi, H. Kokubo, H. Yamamoto, and Y. Sagisaka,
``Out-of-vocabulary word modeling by using sub-word units in large-vocabulary continuous speech recognition,''
IEICE, SP2001-5, pp.33--40, Apr. 2001.
{ LVCSR, Out-of-vocabulary words, Language model, structured, Hierarchical }

{SP2001-6} Y. Inouye,
``Blind Separation and Recovery of Multiple Source Signals,''
IEICE, SP2001-6, pp.41--48, Apr. 2001.
{ Blind signal separation, Blind equalization, MIMO-FIR systems, Independent component analysis. higher-order statistics(higher-order cumurants). }

{SP2001-7} H. Saruwatari,
``Blind Source Separation Using Array Signal Processing,''
IEICE, SP2001-7, pp.49--56, Apr. 2001.
{ Blind Source Separation, ICA, Array Signal Processing, Fast-Convergence Algorithm }

{SP2001-8} F. Asano,
``Blind signal separation in a reverberant sound field,''
IEICE, SP2001-8, pp.57--, Apr. 2001.
{ Blind source seoaration, Reflection, Frequency domain, Subspace method }

{SP2001-9} T. Inoue, M. Nishida, M. Fujimoto and Y. Ariki,
``Voice conversion using subspace method and Gaussian mixture model,''
IEICE, SP2001-9, pp.1--6, May 2001.
{ voice conversion, Gaussian nixture model, subspace method }

{SP2001-10} K. Masaru, A. Saori and A. Hitoshi,
``Tentatively about the achievement of various characters voice quality in the rule synthesis,''
IEICE, SP2001-10, pp.7--14, May 2001.
{ Rule voice synthesis, voice quality, reverse-filter, sound source, and swinging }

{SP2001-11} M. Tamura, T. Masuko, K. Tokuda and T. Kobayashi,
``Speaker adaptation of pitch and spectrum for HMM-based speech synthesis using MSD-MLLR,''
IEICE, SP2001-11, pp.15--20, May 2001.
{ speech synthesis, speaker adaptation, voice conversion, MLLR, average voice }

{SP2001-12} K. Masaru, M. Hiroko, A. Saori, A. Hitoshi and S. Yoshihiro,
``Attempt of identification of characterstic and individual of irregular change at pitch cycle,''
IEICE, SP2001-12, pp.21--28, May 2001.
{ pitch swinging, individual and speaker recognition }

{SP2001-13} D. Xu, H. Mori and H. Kasuya,
``Invariance of Relative F0 Change Field of Chinese Disyllabic Words,''
IEICE, SP2001-13, pp.29--34, May 2001.
{ prosody, fundamental frequency(F0), word-level F0 range, relative F0 change field, Mandarin Chinese }

{SP2001-14} T. Kaduragi 1)3), and M. Honda 2)3),
``Electromagnetic articulograph system based on a parametric representation of the magnetic field,''
IEICE, SP2001-14, pp.35--, May 2001.
{ Articulatory movement, Electromagnetic articulograph, Magnetic field representtion, Spline function }

{SP2001-15} M. Eto, N. Minematsu, K. Hirose and A. Sakurai,
``F0 Contour Generation Based on the Generation Process Model and Statistical Methods,''
IEICE, SP2001-15, pp.1--8, May 2001.
{ prosody, F0 contour generation process model, statistical methods }

{SP2001-16} H. Kawahara, M. Tsuzaki and H. Iwasawa,
``Source Information Representations for Synthetic Speech: Group Delay, Event and Harmonic Structures,''
IEICE, SP2001-16, pp.9--16, May 2001.
{ fundamental frequency, analysis and resynthesis, voicing, hearing, source characteristics }

{SP2001-17} Y. Morimoto, T. Nagai and A. Kurematsu,
``Detection of Phrase Baundaries Based on F0 Movement in Conversational Speech,''
IEICE, SP2001-17, pp.17--22, May 2001.
{ Spontaneous speech, Fundamental frequency, F0 rising, F0 emphasis, F0 flat, Phrase boundary }

{SP2001-18} G. Zhenglai, M. hiroki and K. Hideki,
``Stress accent placement of focused words in Mandarin Chinese,''
IEICE, SP2001-18, pp.23--28, May 2001.
{ Focus, accent, semantic structure, Mandarin Chinese, intonation }

{SP2001-19} M. Shigenaga,
``Characteristic Features of Emotionally Uttered Speech Revealed by Discriminant Analysis (VIII) Use of Neural Networks,''
IEICE, SP2001-19, pp.29--34, May 2001.
{ emotion, emotional speech, prosodic features, discrimination of emotions, neural network }

{SP2001-20} N. Nishizawa, N. Minematsu and K. Hirose,
``Formant speech sysnthesis partly using waveform concatenative synthesis ― Experimental study on VCV sounds ―,''
IEICE, SP2001-20, pp.35--42, May 2001.
{ formant sysnthesis, consonant, waveform concatenation, TD-PSOLA }

{SP2001-21} T. Ohtsuka and H. Kasuya,
``Rubust speech analysis-synthesis method based on the souce-filter model and its applications,''
IEICE, SP2001-21, pp.43--50, May 2001.
{ ARX model, Souce-filter model, Speech analysis-synthesis, Formant, Voice quality control }

{SP2001-22} H. Kawai, M. Tsuzaki, T. Masuda and H. Iwasawa,
``Perceptual Evaluation of Naturalness Degradation Due to Substitution of Phonetic Environment for Concatenative Speech Synthesis,''
IEICE, SP2001-22, pp.51--58, May 2001.
{ speech synthesis, waveform concatenation, unit selection, phonetic environment, naturalness, perception test }

{SP2001-23} P. Zolfaghari, H. Banno, H. Kawahara and F. Itakura,
``An Event Synchronous Sinusoidal Model based on Instantaneous Frequency,''
IEICE, SP2001-23, pp.59--66, May 2001.
{ sinusoidal model, glottal event, instantaneous frequency, time-frequency analysis }

{SP2001-24} H. Takemoto, K. Honda, S. Masaki, I. Shimada, I. Fujimoto, S. Takano and K. Takeo,
``Extraction of temporal pattern of vocal tract area function in a vowel sequence from a 3D MRI movie,''
IEICE, SP2001-24, pp.67--, May 2001.
{ MRI move, 3D motion imaging, vocal tract area function, articulate movement, visualization }

{SP2001-25} T. Fuda, Y. Nagata and M. Abe,
``Speech Recognition Under The Nonstationary Noise Useing Two Chanel Taget Speech Detection,''
IEICE, SP2001-25, pp.1--6, June 2001.
{ Two Channel Processing, Directive Microphone, Voice Dtection, Unsteady Noise, Adaptive Beamforming }

{SP2001-26} M. Fujimoto and Y. Ariki,
``Continuous Speech Recognition under Non-stationary Noisy Environments Using Kalman Filter and Iterative MLLR adaptation,''
IEICE, SP2001-26, pp.7--14, June 2001.
{ noisy speech recognition, non-stationary noise, Kalman filter, iterative unsupervised MLLR adaptation }

{SP2001-27} S. Taniguchi, Y. Obuchi, A. Amano and N. Hataoka,
``Robust Speech Recognition for Car Navigation System ― Voice/Unvoice judgment based on acoustic model ―,''
IEICE, SP2001-27, pp.15--20, June 2001.
{ Car Navigation System, Noise Robustness, Noise Rejection, Acoustic Model, Voice/Unvoice judgment }

{SP2001-28} S. Takeuchi, M. Yamashita, T. Uchida and M. Sugiyama,
``Optimization of Voice/Music Segmentation,''
IEICE, SP2001-28, pp.21--28, June 2001.
{ Speech feature, Segmentation, Discriminative training }

{SP2001-29} R. Oka, T. Nishimura, M. Ihara, J. Xin Zhang and T. Akasaka,
``Speech and Text Retrieval using Phoneme Sequence Representation and Continuous Dynamic Programming,''
IEICE, SP2001-29, pp.29--36, June 2001.
{ Speech retrieval, Text retrieval, Phoneme recognition, Continuous DP, Spotting }

{SP2001-30} H. Nakajima, I. Hirano and Y. Sagisaka,
``Pronunciation Variants Analysis using Speaking Style Parallel Corpus,''
IEICE, SP2001-30, pp.37--, June 2001.
{ speech recognition, pronunciation variation, pronunciation variants, speaking style, parallel corpus }

{SP2001-31} K. Onoe, H. Segi, T. Kobayakawa, S. Sato, T. Imai and A. Ando,
``Filter-Bank Subtraction for Recognition of Field Reporting Speech in Broadcast News Programs,''
IEICE, SP2001-31, pp.1--6, June 2001.
{ speech recognition, broadcast news, spectral subtraction, filter-bank }

{SP2001-32} H. Kokubo, S. Onishi, H. Yamamoto and Y. Sagisaka,
``Decoding with sub-word network models for out-of-vocabulary words recognition,''
IEICE, SP2001-32, pp.7--12, June 2001.
{ out of vocabulary, sub-word models, decoder, and speech recognition }

{SP2001-33} H. Yamamoto, S. Isogai and Y. Sagisaka,
``Multi-Class Composite N-gram Language Model Using Multiple Word Clusters and Word Successions,''
IEICE, SP2001-33, pp.13--18, June 2001.
{ N-gram language model, Class N-gram, Word clustering, Variable length N-gram }

{SP2001-34} Y. Kodama, T. Utsuro, H. Nishizaki and S. Nakagawa,
``Experimental Evaluation on Confidence of Agreement among Multiple Japanese LVCSR Models,''
IEICE, SP2001-34, pp.19--26, June 2001.
{ LVCSR models, confidence measures, combination of multiple models, acoustic models, recognition error detection, detection }

{SP2001-35} Y. Itoh,
``Performance Evaluation of Shift Continuous DP Algorithm for Detecting Similar Sections and its Application for Presentation Speech,''
IEICE, SP2001-35, pp.27--, June 2001.
{ similar section, spotting Continuous DP, labeling, presentation speech }

{SP2001-36} S. Kanno and T. Funada,
``A Study of Robustness in Noisy Speech Recognition using Weighted Variance Expansion of Word HMMs,''
IEICE, SP2001-36, pp.1--8, July 2001.
{ robustness, HMM, noisy environment, speech recognition, variance expansion }

{SP2001-37} T. Masuko, T. Kobayashi and K. Tokuda,
``A Study on Speaker Independent Phonetic Vocoder Using HMM,''
IEICE, SP2001-37, pp.9--16, July 2001.
{ speech coding, speech synthesis, speech recognition, HMM, speaker independent }

{SP2001-38} Y. Morita and T. Funada,
``Quality Improvement of Speech by Neural Network Vector Quantization Combined with Spectral Subtraction Method,''
IEICE, SP2001-38, pp.17--22, July 2001.
{ speech coding, neural network, vector quantization, LSP, noise suppression }

{SP2001-39} M. Tsuzaki,
``Effectiveness of Auditory Parameterization for Unit Selection in Concatenative Speech Synthesis: Comparison to a Physical Model and Perceptual Evaluation,''
IEICE, SP2001-, pp.--, July 2001.
{ concatenative speech synthesis, auditory model, unit selection, feature extraction, perceptual evaluation }

{SP2001-40} T. Irino, Roy D. Patterson and H. Kawahara,
``Signal resynthesis from Auditory Mellin Image using a high-quality VOCODER, STRAIGHT,''
IEICE, SP2001-40, pp.31--, July 2001.
{ Speech analysis/synthesis, Auditory model, Mellin transform, nonlinear multivariate analysis, DCT }

{SP2001-41} M. Iwaki, T. Ishizuka and T. Kiryu,
``Variations in Finger Plethysmograph and Respiration Under Tone-Search Task,''
IEICE, SP2001-41, pp.1--8, July 2001.
{ Tone-search task, Finger plethysmograph, Respiration, Mental stress, Principal Component Analysis }

{SP2001-42} Y. Arai, H. Kondo and K. Hikita,
``Intelligibility Test of a Directional Hearing Aid Using a Dummy Head in Simulated Noisy Surroundings,''
IEICE, SP2001-42, pp.9--16, July 2001.
{ intelligibility test, directional hearing aid, dummy head, simulated noisy surroundings }

{SP2001-43} Carlos T. Ishi, N. Minematsu and K. Hirose,
``Relationship between acoustically observed F0 and perceived pitch for Japanese accent and intonation,''
IEICE, SP2001-43, pp.17--22, July 2001.
{ F0 contour, F0mora, Pitch perception, Accent, Intonation }

{SP2001-44} K. Ishizuka and K. Aikawa,
``Comparison of vowel perception under a white noise and under a harmonic complex tone,''
IEICE, SP2001-44, pp.23--, July 2001.
{ vowel perception, harmonics, natural vowel, speech spectrum }

{SP2001-45} T. Seo, J. Murakami, T. Maeta and S. Ikehara,
``Word speech recognition using Mora Position and Mora Length,''
IEICE, SP2001-45, pp.1--6, Aug. 2001.
{ Word speech recognition, Mora position, Mora length }

{SP2001-46} N. Tamura, K. Sugahara, Y. Li and R. Konishi,
``On the Multiple Word Recognition System by Using Source Separation Technique,''
IEICE, SP2001-46, pp.7--14, Aug. 2001.
{ Independent Component Analysis, Word Recognition, Direction Of Arrival Estimation, Blind Deconvolution }

{SP2001-47} Y. Yamashita and H. Matsumoto,
``A Study on Acoustical Correlates to Adjective Ratings of Speaker Characteristics,''
IEICE, SP2001-47, pp.15--22, Aug. 2001.
{ Speaker characteristics, adjectives, formant loci, voice quality, voice conversion, semantic differential }

{SP2001-48} Carlos T. Ishi, N. Minematsu and K. Hirose,
``Identification of Japanese accent in continuous speech considering pitch perception,''
IEICE, SP2001-48, pp.23--30, Aug. 2001.
{ Accent types, Pitch perception, F0target }

{SP2001-49} M. Kimoto, T. Namiki, T. Shimizu, N. Isu and K. Sugata,
``Range of Phonemic Environmental Influence for Selection of VCV Sequences on VCV Speech Synthesis by Rule,''
IEICE, SP2001-49, pp.31--, Aug. 2001.
{ Speech Synthesis by Rule, VCV Instance, Selection Method, Phonemic Environment }

{SP2001-50} T. Imai, K. Yamamoto and H. Matsumoto,
``The SVD-based MLLR speaker adaption method,''
IEICE, SP2001-50, pp.1--8, Aug. 2001.
{ MLLR, SVD, regresion class, effective rank, dictation system }

{SP2001-51} E. Kunimitsu, T. Namiki, H. Yoshimura, T. Shimizu, N. Isu and K. Sugata,
``Analysis of Japanese vowel series by using sandglass type neural network,''
IEICE, SP2001-51, pp.9--16, Aug. 2001.
{ sand-glass type neural network, vowel, LSP analysis, formant, principal component analysis }

{SP2001-52} T. Namiki, Y. Furumoto, H. Yoshimura, N. Isu and K. Sugata,
``Application to a noise reduction filter for hearing aid of cascaded sandglass-type neural networks,''
IEICE, SP2001-52, pp.17--24, Aug. 2001.
{ Sandglass-type Neural Network, Speech Signal, NoiseReduction Filter, dynamics }

{SP2001-53} T. Maeta, J. Murakami and S. Ikehara,
``Segmentation using Mora position and Mora length,''
IEICE, SP2001-53, pp.25--, Aug. 2001.
{ Segmentation, mora position, mora length }

{SP2001-54} H. Takagi and T. Shimamura,
``Fundamental Frequency Extraction Method Based on the p-th Power of Amplitude Spectrum with Band Limitation,''
IEICE, SP2001-54, pp.1--8, Sep. 2001.
{ fundamental frequency, band limitation, exponentiated spectrum }

{SP2001-55} Z. Li, T. Kouyama and N. Shimizu,
``The proposal of the high-speed quantization circuit for MP3,''
IEICE, SP2001-55, pp.9--16, Sep. 2001.
{ Quantization circuit, Voice evaluation }

{SP2001-56} N. Kakita, K. Sugahara and R. Konishi,
``On the Lip Area Extraction Method based on the Color Information,''
IEICE, SP2001-56, pp.17--22, Sep. 2001.
{ Extraction of lip area, Fuzzy theory, The active contour model }

{SP2001-57} M. Washio, Y. Anno, K. Sugahara and R. Konishi,
``Hardware Realization of Active Contour Model for Real Time Video Image Processing,''
IEICE, SP2001-57, pp.23--30, Sep. 2001.
{ Contour extraction, Active contour model, FPGA }

{SP2001-58} T. Ozaki, Y. Sato and T. Okamoto,
``A personal use oriented geometrical transformer for moving pictures,''
IEICE, SP2001-58, pp.31--38, Sep. 2001.
{ digital moving picture, geometrical transformation, polynomial approximation, recursive formula }

{SP2001-59} A. Kawakami,
``A Note on the Multi-Input Multi-Output Separable-Denominator Two-Dimensional Systems,''
IEICE, SP2001-59, pp.39--44, Sep. 2001.
{ multi-input multi-output systems, separable-denominator form, two-dimensional systems, realization dimension, minimal dimension }

{SP2001-60} H. Fujisaki,
``Promoting Speech Research in Japan ― Forty Years in Retrospect and Expectation for the Future ―,''
IEICE, SP2001-60, pp.45--, Sep. 2001.
{ Speech research, Digital signal processing, Spoken language processing, Human-machine spoken dialogue, ICASSP, ICSLP }

{SP2001-61} Y. Yamada, H. Kiya and N. Kambayashi,
``A Tap-Position Estimation for Sparse-Tap Adaptive FIR Filter Based on the Multirate Technique,''
IEICE, SP2001-61, pp.1--8, Sep. 2001.
{ tap-position control, delay estimation, multirate, adaptive filter, VoIP }

{SP2001-62} T. Fumoto and S. Sasaki,
``A low bit rate speech codec using multi band excitation and LPC modeling of harmonic magnitudes for private mobile radio,''
IEICE, SP2001-62, pp.9--16, Sep. 2001.
{ low bit rate, speech codec, multi-band excitation, MBE, linear predictive coding, error correction }

{SP2001-63} T. Yoshimura, K. Tokuda, T. Masuko, T. Kobayashi and T. Kitamura,
``Introduction of Mixed Excitation Model and Postfilter to HMM-based Speech Synthesis,''
IEICE, SP2001-63, pp.17--22, Sep. 2001.
{ hidden Markov model, text-to-speech synthesis, mixed excitation model, postfilter }

{SP2001-64} H. Banno, K. Takeda and F. Itakura,
``Perceptual Distance Measure of Phase Based on Group Delay,''
IEICE, SP2001-64, pp.23--30, Sep. 2001.
{ the human auditory system, distance measure, phase spectrum, group delay, ellipsoidal mapping }

{SP2001-65} H. Iwasawa, M. Tsuzaki, H. Kawai and H. Kawahara,
``Optimizing phase dispersion for excitation source in speech synthesis with STRAIGHT: Psychoacoustical evaluation and optimization of control parameters,''
IEICE, SP2001-65, pp.31--38, Sep. 2001.
{ analysis/synthesis, psychoacoustical evaluation, voicing source, phase, group delay, voice quality variation }

{SP2001-66} S. Nakagawa,
``A Survey on Automatic Speech Recognition,''
IEICE, SP2001-66, pp.39--40, Sep. 2001.
{ speech recognition, acoustic model, HMM, language model, ngram }

{SP2001-67} T. Masuko,
``Multi-Space Probability Distribution HMM,''
IEICE, SP2001-67, pp.41--42, Sep. 2001.
{ multi-space probability distribution, HMM, pitch }

{SP2001-68} H. Saruwatari, S. Kajita, K. Takeda and F. Itakura,
``Speech enhancement using nonlinear microphone array based on noise adaptive complementary beamforming,''
IEICE, SP2001-68, pp.43--44, Sep. 2001.
{ , , , ,  }

{SP2001-69} A. Sasou and K. Tanaka,
``Separation of concurrent sounds based on the excitation sources using the time-domain ― Gain-adapted AR-HMM decomposition method ―,''
IEICE, SP2001-69, pp.45--52, Sep. 2001.
{ monaural concurrent sounds, sound-separation, noise reduction, source filter, AR-HMM }

{SP2001-70} H. Zen, K. Tokuda, T. Masuko, T. Kobayashi and T. Kitamura,
``A Pitch Pattern Modeling Technique using Dynamic Features on the Border of Voiced and Unvoiced Segments,''
IEICE, SP2001-70, pp.53--58, Sep. 2001.
{ pitch pattern generation, multi-space probability distribution, hidden Markov model, speech synthesis }

{SP2001-71} T. Ito, K. Takeda and F. Itakura,
``Acoustic Analysis and Recognition of Whispered Speech,''
IEICE, SP2001-71, pp.59--64, Sep. 2001.
{ whispered speech, acoustic analysis, speech recognition }

{SP2001-72} A. Sawabe, K. Shichiri, T. Yoshimura, K. Tokuda, T. Masuko, T. Kobayashi and T. Kitamura,
``Application of eigenvoice technique to spectrum and pitch pattern modeling in HMM-based speech synthesis,''
IEICE, SP2001-72, pp.65--, Sep. 2001.
{ Eigenvoice, MSD-HMM, Speech Synthesis, Clustering }

{SP2001-73} M. Fujiyoshi, T. Ishida, H. Sawazaki and N. Ohtake,
``Development of the second edition of Unified Japanese Braille Code,''
IEICE, SP2001-73, pp.1--8, Oct. 2001.
{ Braille, Visual Disability, Unicode, Unified Japanese Braille Code, Unified English Braille Code }

{SP2001-74} K. Yamaguchi and F. Kawane,
``On the Prototype System of Aloud-Reading in Japanese for Documents Including Mathematical Expressions,''
IEICE, SP2001-74, pp.9--16, Oct. 2001.
{ Visual Disabilities, Accessibility, Mathematical Expression, Aloud Reading LaTeX, Japanese }

{SP2001-75} T. Saito, Y. Horiuchi and A. Ichikawa,
``On an Interactive Method for Web Accessibility,''
IEICE, SP2001-75, pp.17--22, Oct. 2001.
{ Web Accessibility, visually impaired, Interactive, World Wide Web, text-to-speech }

{SP2001-76} N. Hasegawa, H. Kiyoto and W. Takahasi,
``Aphasia with Agrammatism and the Elementary Verbal Perceptions in Japanese Language,''
IEICE, SP2001-76, pp.23--30, Oct. 2001.
{ agrammatism, aphasia, verbal perception, elementary verbal perception, universal interface, Japanese language }

{SP2001-77} M. Sekiguchi, N. Minematsu and K. Hirose,
``Identification of Aged Speakers with Speaker Recognition Techniques and Its Improvement,''
IEICE, SP2001-77, pp.31--38, Oct. 2001.
{ aged people, spoken dialogue system, speaker recognition, speech rate, local power perturbation }

{SP2001-78} T. Konashi, M. Suzuki and S. Makino,
``Spoken dialogue system with an autonomous care robot,''
IEICE, SP2001-78, pp.39--44, Oct. 2001.
{ spoken dialogue system, filler model, an autonomous care robot }

{SP2001-79} T. Satoh, T. Masuko, T. Kobayashi and K. Tokuda,
``A Study on Discrimination between Synthetic and Natural Speech for Speaker Verification Sysems,''
IEICE, SP2001-79, pp.45--50, Oct. 2001.
{ Speaker verification, imposture, synthetic speech discrimination }

{SP2001-80} M. Yoshioka, M. Tamura, T. Masuko, T. Kobayashi and K. Tokuda,
``On the Effect of Contextual Factors on HMM-Based Speech Synthesis,''
IEICE, SP2001-80, pp.51--56, Oct. 2001.
{ HMM-based speech synthesis, contextual factors, prosody }

{SP2001-81} Y. Ishimatsu, K. Tokuda, T. Masuko, T. Kobayashi and T. Kitamura,
``Investigation of State Duration Model based on Gamma distribution for HMM-based Speech Synthesis,''
IEICE, SP2001-81, pp.57--62, Oct. 2001.
{ Gamma distribution, state duration model, context clustering, speech synthesis }

{SP2001-82} K. Iwata, T. Takahashi, T. Nakanishi, K. Tokuda and T. Kitamura,
``Simultaneous Determination of LPC Analysis Order and Frame Partitioning Based on MDL Criterion,''
IEICE, SP2001-82, pp.63--68, Oct. 2001.
{ LPC analysis, variable analysis order, variable frame lenrth, MDL criterion, DP }

{SP2001-83} K. Shinoda, D. Tran and K. Iso,
``Efficient Reduction of Gaussian Components Using MDL Criterion for Speech Recognition,''
IEICE, SP2001-83, pp.69--, Dec. 2001.
{ speech recognition, HMM, MDL criterion, tree structure, Gaussian distribution }

{SP2001-84} T. S. Kobayakawa, H. Segi, A. Matsui, K. Onoe, S. Sato, I. Toru and A. Ando,
``A method for improving the accuracy of dynamic features calculation in speech recognition,''
IEICE, SP2001-84, pp.1--6, Dec. 2001.
{ Speech Recognition, Dynamic Features, Regression Coefficient }

{SP2001-85} T. Fukuda and T. Nitta,
``The Method of Extracting Peripheral Features that Achieve High Performance in both Tasks of LVCSR and ISWR,''
IEICE, SP2001-85, pp.7--12, Dec. 2001.
{ speech recognition, feature extraction, CMN, peripheral feature, local feature, HMM }

{SP2001-86} N. Takahashi and S. Nakagawa,
``Syllable recognition using syllable-segmental statistics and syllable-based HMM,''
IEICE, SP2001-86, pp.13--18, Dec. 2001.
{ syllable recognition, segment model, HMM }

{SP2001-87} J. Rokui, N. Mitsuru, H. Shimodaira and S. Sagayama,
``Vocal Tract Length Normalization Using Linear Transformation based on Maximum Likelihood Estimation,''
IEICE, SP2001-87, pp.19--25, Dec. 2001.
{ Vocal Tract Length Normalization, Linear Transformation, Maximum Likelihood Estimation, Speaker Adaptation, Speaker Normalization }

{SP2001-88} N. Watanabe, T. Yamada, N. Kitawaki and F. Asano,
``Voice Activity Detection for Sentence Utterances Using Environment Sound Models and HMM Composition,''
IEICE, SP2001-88, pp.25--30, Dec. 2001.
{ Voice activity detection, Viterbi alignment, environment sound models, HMM composition }

{SP2001-89} M. Kobayashi, T. Kitanura and S. Kitazawa,
``A study of speaker segmentation of dialogue speech with speech overlapped section,''
IEICE, SP2001-89, pp.31--36, Dec. 2001.
{ multispeaker, speaker recognition, speaker segmentation, dialogue speech }

{SP2001-90} K. Markov, T. Matsui, R. Gruhn, J. Zhang and S. Nakamura,
``ATR System for Robust Speech Recognition in Real World Noisy and Channel Environments,''
IEICE, SP2001-90, pp.37--44, Dec. 2001.
{ noise robustness, online adaptation, hypothesis combination, robust features }

{SP2001-91} K. Yao, J. Chen, K. Paliwal and S. Nakamura,
``Noise Speech Recognition based on Robust Features and A Model-Based Noise Compensation evaluated on Aurora-2 Task,''
IEICE, SP2001-91, pp.45--50, Dec. 2001.
{ Speech recognition, Noise, Robust speech recognition }

{SP2001-92} M. Ida and S. Nakamura,
``Rapid Model Adaptation with a Prior Noise GMM and Multi-SNR Models for Noisy Speech Recognition,''
IEICE, SP2001-92, pp.51--56, Dec. 2001.
{ HMM composition, noise model, nonstationary noise, multipath model }

{SP2001-93} M. Fujimoto and Y. Ariki,
``Noise Robust Speech Recognition by Integration of MLLR Adaptation and Feature Extraction for Noise Reduced Speech,''
IEICE, SP2001-93, pp.57--62, Dec. 2001.
{ noise robust speech recognition, non-stationary noise, Kalman filter, unsupervised MLLR adaptation, root cepstral coefficient }

{SP2001-94} K. Aikawa and K. Ishizuka,
``An Effect of Pitch Fluctuation on Noise-Robust Speech Recognition,''
IEICE, SP2001-94, pp.63--68, Dec. 2001.
{ noise-robust, speech recognition, pitch, synchronous, spectral analysis }

{SP2001-95} H. G. Okuno and K. Nakadai,
``Research Issues and Current Status of Robot Audition,''
IEICE, SP2001-95, pp.69--74, Dec. 2001.
{ robot audition, computational auditory scene analysis, auditory epipolar geometry, active audition, sensor fusion }

{SP2001-96} N. Iwahashi,
``Language Acquisition by Robots,''
IEICE, SP2001-96, pp.75--80, Dec. 2001.
{ robot, language, mutual belief, communication, learning }

{SP2001-97} Y. Yamakata, T. Kawahara and H. G. Okuno,
``Spoken Dialogue System for Robot with Computer Vision,''
IEICE, SP2001-97, pp.81--86, Dec. 2001.
{ spoken dialogue system, spoken language understanding, user model, belief network }

{SP2001-98} A. Miyata, N. Iwahashi and A. Kurematsu,
``Mutual Belief Forming by Robots based on the Process of Utterance Comprehension,''
IEICE, SP2001-98, pp.87--92, Dec. 2001.
{ mutual belief, utterance comprehension, robot, communication, learning }

{SP2001-99} R. Nisimura, T. Uchida, A. Lee, H. Saruwatari and K. Shikano,
``Development of Julius-based Speech Dialogue System for Campus Receptionist Robot,''
IEICE, SP2001-99, pp.93--98, Dec. 2001.
{ Speech dialogue robot, LVCSR engine Julius, Keyword search, N-gram language model }

{SP2001-100} S. Kobashikawa, N. Minematsu, K. Hirose and D. Erickson,
``Modeling of Stressed Syllables for their Detection in English Sentences to Develop an English Rhythm Learning System,''
IEICE, SP2001-100, pp.99--104, Dec. 2001.
{ English rhythm, stressed syllable, HMM, English CAI }

{SP2001-101} M. Terao, N. Minematsu and K. Hirose,
``Improvement of N-gram Language Models Using Accent Phrase Boundaries,''
IEICE, SP2001-101, pp.105--, Dec. 2001.
{ language model, prosody, accent phrase boundary, transition of part-of-speech, continuous speech recognition }

{SP2001-102} T. Shinozaki and S. Furui,
``A statistical analysis of individual differences in spontaneous speech recognition performance,''
IEICE, SP2001-102, pp.1--6, Dec. 2001.
{ spontaneous speech recognition, Corpus of Spontaneous Japanese, individual differences, unsupervised speaker adaptation }

{SP2001-103} H. Namjo and T. Kawahara,
``Speaking-Rate Dependent Decoding and Adaptation for Spontaneous Lecture Speech Recognition,''
IEICE, SP2001-103, pp.7--12, Dec. 2001.
{ automatic speech recognition, spontaneous speech, lecture speech, speaking rate, acoustic model, speaker adaptation }

{SP2001-104} K. Okuda, T. Kawahara and S. Nakamura,
``Lecture speech recognition considering the speaking rate variation,''
IEICE, SP2001-104, pp.13--18, Dec. 2001.
{ automatic speech recognition, lecture speech, speaking rate, frame period, frame length }

{SP2001-105} J. Ogata and Y. Ariki,
``Unsupervised Adaptation of an Acoustic Model Using Confidence Measures Based on Phoneme Posterior Probabilities,''
IEICE, SP2001-105, pp.19--24, Dec. 2001.
{ unsupervised adaptation, confidence measures, word error minimization, phoneme error minimization }

{SP2001-106} K. Maekawa, H. Kikuchi and Y. Igarashi,
``X-JToBI:An Intonation Labeling Scheme for Spontaneous Japanese,''
IEICE, SP2001-106, pp.25--30, Dec. 2001.
{   }

{SP2001-107} N. Kawaguchi, S. Matsubara, K. Takeda, F. Itakura and Y. Inagaki,
``IN-Car Spoken Dialogue Database,''
IEICE, SP2001-107, pp.31--36, Dec. 2001.
{ speech database, in-car spoken dialogue, robust speech recognition, spoken language processing }

{SP2001-108} S. Kurohashi,
``Automatic Question Answering based on Large Text Knowledge Base,''
IEICE, SP2001-108, pp.37--42, Dec. 2001.
{ Dialogue System, Information Retrieval, Question Answering, Text Knowledge Base }

{SP2001-109} C. Hori and S. Furui,
``Automatic Speech Summarization for English Broadcast News Speech,''
IEICE, SP2001-109, pp.43--48, Dec. 2001.
{ speech summarization, English broadcast news speech, word significance measure, linguistic likelihood, confidence measure, stochastic dependency context free grammar, two-level dynamic programming }

{SP2001-110} Y. Okimoto, H. Yamamoto, E. Sumita and G. Kikui,
``Correcting Mis-recognitions Using Basic Travel Expression Corpus,''
IEICE, SP2001-110, pp.49--54, Dec. 2001.
{ Error correction, Expression Corpus, Travel expression }

{SP2001-111} K. Itou, A. Fujii and T. Ishikawa,
``A Lecture-On-demand System using Spoken Document Retrival,''
IEICE, SP2001-111, pp.55--60, Dec. 2001.
{ Spoken Document Retrival, Speech Recognition, Information Retrival, Lecure-on-demand System, e-education }

{SP2001-112} H. Nishizaki and S. Nakagawa,
``News Spoken Document Retrieval by Considering Out-of-Vocabulary Keywords,''
IEICE, SP2001-112, pp.61--66, Dec. 2001.
{ spoken document retrieval, OOV detection processing, DP matching, word spotting }

{SP2001-113} K. Komatani, T. Kawahara, Y. Kiyota, S. Kurohashi and P. Fung,
``Restaurant Search System with Speech Interface using Flexible Language Model and Matching,''
IEICE, SP2001-113, pp.67--, Dec. 2001.
{ speech recognition, search system, domain adaptation, matching method, syntactic structure, statistical language model }

{SP2001-114} S. Toyoda, T. Moriyama and H. Ogawa,
``The speech synthesis system using DMP reflected speech of characteristic,''
IEICE, SP2001-114, pp.1--8, Jan. 2002.
{ DMP, speech synthesis, speech of characteristic }

{SP2001-115} K. Sasaki and T. Yoshida,
``Compound Accentuation Rules for Compound Nouns,''
IEICE, SP2001-115, pp.9--16, Jan. 2002.
{ compound accentuation lule, compound noun, accent phrase, morphological analysis }

{SP2001-116} M. Eto, K. Hirose and N. Minematsu,
``F0 Contour Generation Based on the Generation Process Model and Statistical Methods Considering Syntactic Structures,''
IEICE, SP2001-116, pp.17--22, Jan. 2002.
{ prosody, F0 contour generation process model, statistical methods, syntactic structure }

{SP2001-117} T. Kitamura, T. Itoh, K. Mochizuki and S. Kitazawa,
``Prosodic phrase labeling based on prosodic features for developing prosodic database,''
IEICE, SP2001-117, pp.23--30, Jan. 2002.
{ Prosody, Prosodic phrase, Database, Phoneme boundary, Segmentation }

{SP2001-118} D. Nagahata and M. Yanagida,
``Improvement of Phoneme Identifiability by Fluctuation of Fundamental Frequency,''
IEICE, SP2001-118, pp.31--38, Jan. 2002.
{ phoneme identifiability, fundamental frequency jitter, formant frequency, singing voice, synthetic vowel }

{SP2001-119} T. Hirai, S. Tenpaku and K. Shikano,
``Using start/end timings of spectral transitions between phonemes in concatenative speech synthesis,''
IEICE, SP2001-119, pp.39--44, Jan. 2002.
{ Speech synthesis, Spectral transitions between phonemes, Label, Concatenative speech synthesis system }

{SP2001-120} T. Toda, H. Kawai, M. Tsuzaki and K. Shikano,
``A Unit Selection Algorithm for Japanese Speech Synthesis Based on Both Phoneme Unit and Diphone Unit,''
IEICE, SP2001-120, pp.45--52, Jan. 2002.
{ Japanese Text-to-Speech, unit selection, vowel sequence, diphone unit, cost function }

{SP2001-121} B. Nettre, K. Hirose and N. Minematsu,
``An Experimental Study on Concatenative Speech Synthesis Using a Fusion Technique and VCV/VV Units,''
IEICE, SP2001-121, pp.53--60, Jan. 2002.
{ Concatenative Speech Synthesis, Unit Fusion, Sinusoidal Modeling, Spectral Modifications, TD-PSOLA }

{SP2001-122} T. Masuda, T. Toda, H. Kawanami, H. Saruwatari and K. Shikano,
``A Study on the Speech Synthesis Method by Using Database with Variety of Speech-Rate,''
IEICE, SP2001-122, pp.61--68, Jan. 2002.
{ speech synthesis, STRAIGHT, speech-rate, database, phoneme duration }

{SP2001-123} J. Ni and H. Kawai,
``Design of a Speech Corpus for Mandarin Speech Synthesis Taking into Account Phrasal Prosody,''
IEICE, SP2001-123, pp.69--76, Jan. 2002.
{ Speech corpus, Mandarin prosody, Corpus-based speech synthesis }

{SP2001-124} K. Fujihata, H. Matsui and H. Kawahara,
``Emotional speech database design for highly natural synthesis of emotional speech,''
IEICE, SP2001-124, pp.77--, Jan. 2002.
{ emotional speech, emotional speech database, speech synthesis, subjectivity appraisal method }

{SP2001-125} A. Hirano, N. Babaguchi and T. Kitahashi,
``Understanding Auditory Environment in Broadcasted Video Streams,''
IEICE, SP2001-125, pp.1--8, Jan. 2002.
{ Audio Based Video Indexing, Broadcasted Sports Video, Auditory Environment, Voiced Sound, Ambient Sound }

{SP2001-126} Y. Hioka and N. Hamada,
``Voice Activity Detection Using Microphone Array Combining with Wavelet Analysis,''
IEICE, SP2001-126, pp.9--16, Jan. 2002.
{ Microphone Array, Wavelet Packet, Voice Activity Detection, Eigenspace Analysis }

{SP2001-127} I. Katsuse and Y. Sugano,
``PHONOBEST: a noise-robust automatic speech recognition system with phonetic estimation process based on "expectation",''
IEICE, SP2001-127, pp.17--24, Jan. 2002.
{ non-stationary noise, automatic speech recognition, phonetic knowledge, top-down process }

{SP2001-128} T. Utsuro, H. Nishizaki, T. Harada, Y. Kodama and S. Nakagawa,
``Performance Analysis of Confidence of Agreement among Multiple LVCSR Models,''
IEICE, SP2001-128, pp.25--32, Jan. 2002.
{ LVCSR models, confidence measures, combination of multiple models, acoustic models, recognition error detection, dictation }

{SP2001-129} G. Nagino, D. Willett, Y. Minami, E. Mcdermott, A. Nakamura, N. Miyazaki and K. Shikano,
``Evaluation of a decoder based on finite state transducer using real dialogue speech data,''
IEICE, SP2001-129, pp.33--40, Jan. 2002.
{ Finite State Transducer, Speech Dialogue System, Dialogue Speech Recognition }

{SP2001-130} T. Yamamoto, J. Ogata and Y. Ariki,
``Effectiveness of An Expanded Dictionary in Information Retrieval System by Keyword Spotting,''
IEICE, SP2001-130, pp.41--46, Jan. 2002.
{ web news, N-gram language model, LSI method, expanded dictionary, keyword spotting }

{SP2001-131} A. Inoue, T. Mikami and Y. Yamashita,
``Use of F0 model for the extraction of important sentences in speech summarization,''
IEICE, SP2001-131, pp.47--54, Jan. 2002.
{ Summarization, F0 model, Speech, Prosodic, Important sentence }

{SP2001-132} M. Takayama, T. Nishimoto, M. Araki and Y. Niimi,
``The Role of Audio Effects in Interactive Voice Response Systems,''
IEICE, SP2001-132, pp.55--62, Jan. 2002.
{ Interactive voice response system, Spoken dialogue, Audio effect, Speech recognition, Human-interface }

{SP2001-133} K. Imoto, M. Dantsuji and T. Kawahara,
``Computer-Assisted English Prosody Learning System based on Automatic Detection of Sentence Stress and Stress-Timing,''
IEICE, SP2001-133, pp.63--70, Jan. 2002.
{ prosody, CALL, stress rhythm, HMM, a multi-stage discrimination }

{SP2001-134} A. Raux and T. Lawahara,
``Intelligibility Assessment and Pronunciation Error Diagnosis for a CALL System,''
IEICE, SP2001-134, pp.71--78, Jan. 2002.
{ CALL, Intelligibility Assessment, Pronunciation Errors, Error Diagnosis, ASR }

{SP2001-135} N. Maeda and Y. Yamashita,
``A method of pronunciation evaluation for English words using Japanese and English phonemic models,''
IEICE, SP2001-135, pp.79--, Jan. 2002.
{ CALL, Language learning, Pronunciation evaluation, Phonemic model }

{SP2001-136} K. Shirai,
``Towards Spoken Dialogue System From Speech Recognition,''
IEICE, SP2001-136, pp.1--8, Mar. 2002.
{ Speech recognition, Spoken dialogue system, Articulatory motion estimation, Conversation robot, Feature extraction }

{SP2001-137} S. Hashimoto,
``Characteristics of Japanese Accents based on Linguistic, Auditory and Physical Attributes,''
IEICE, SP2001-137, pp.9--20, Mar. 2002.
{ Accent, Pitch, Rhythm }

{SP2001-138} T. Nakatani and T. Irino,
``Fundamental Frequency Estimation Based on Dominance Spectrum,''
IEICE, SP2001-138, pp.21--28, Mar. 2002.
{ degree of dominance, fundamental frequency estimation, instantaneous frequency, background noise, spectral distortion, multi-talker noise, SRAEN filter }

{SP2001-139} K. Okada,
``Methodology of Speech Recognition using Island Analysis,''
IEICE, SP2001-139, pp.29--36, Mar. 2002.
{ HMM(Hidden Markov Model), Vocal Parameter }

{SP2001-140} I. Kinoshita, T. Nishimoto, M. Araki and Y. Niimi,
``Accent Type Recognition Using HMMs,''
IEICE, SP2001-140, pp.37--42, Mar. 2002.
{ accent phrase, accent type, Hidden Markov Model }

{SP2001-141} I. Hiradate and M. Akagi,
``Analyses of acoustic features of "anger" emotional speech,''
IEICE, SP2001-141, pp.43--50, Mar. 2002.
{ emotional speech, acoustic feature, accent portion, Neutral, Cold Anger, Hot Anger }

{SP2001-142} T. Miwa and S. Nakagawa,
``Analysis and comparison of the prosodic features for Japanese English and native English,''
IEICE, SP2001-142, pp.51--58, Mar. 2002.
{ Japanese English, prosody, stress rhythm, pronunciation proficiency, linear regerssion model }

{SP2001-143} R. Tako and M. Akagi,
``Dynamical structures contributing to perception of central vowel in /VVV/ concatenated vowels,''
IEICE, SP2001-143, pp.59--66, Mar. 2002.
{ time expansion, time contraction, noise replacement, SFTR(Spectral Feature Transition Rate) }

{SP2001-144} J. Lu and H. Kawai,
``Perceptual Evaluation of Naturalness due to Substitution of Chinese Syllable for Concatenative Speech Synthesis,''
IEICE, SP2001-144, pp.67--, Mar. 2002.
{ speech synthesis, perception test, naturalness, waveform concatenation, unit selection, tone }

{SP2001-145} K. Kanamori, N. Suto and S. Sumi,
``Influences of Duration and Interval on the Cross-Dimensional Interference Effects on Pitch And Timbre,''
IEICE, SP2001-145, pp.1--8, Mar. 2002.
{ Pitch, Timbre, Interference effects, Duration }

{SP2001-146} K. Ono, N. Sutou and S. Sumi,
``The Effect of Cue Validity on the Recognition of Tones: A Comparison between the Recognition Task and the Priming Task :,''
IEICE, SP2001-146, pp.9--16, Mar. 2002.
{ Tone, Attention, Template Model, Endogenous, and Facilitation Effect }

{SP2001-147} A. Tanaka, K. Mori and Y. Takano,
``Evidence for Dissociation of Working Memory for Suprasegmental Phonology and Other Suprasegmental Features,''
IEICE, SP2001-147, pp.17--24, Mar. 2002.
{ working memory, prosody, pitch accent, dialect, L2 }

{SP2001-148} Y. Sato, K. Mori and Y. Fukushima,
``Temporal Characteristics of Fundamental Frequency Control by Auditory Feedback and its Application to Stuttering Evaluation,''
IEICE, SP2001-148, pp.25--30, Mar. 2002.
{ TAF(Transformed Auditory Feedback), F0(Fundamental frequency), Stuttering, Disfluency }

{SP2001-149} N. Ono, A. Saito and S. Ando,
``Theory and Experiment of Sound Source Localization Sensor by Gradient-Detection with Mimicking Ormia Ochracea,''
IEICE, SP2001-149, pp.31--36, Mar. 2002.
{ sound source localization, micro sensor, gradient detection, gimbal structure, ormia ochracea }

{SP2001-150} T. Kaburagi, K. Wakamiya, K. Sawada and M. Honda,
``Three-dimensional electromagnetic articulography using a spline representation of the magnetic field,''
IEICE, SP2001-150, pp.37--44, Mar. 2002.
{ Speech production, Articulatory motion, Electromagnetic position sensor, Magnetic field representation, spline function }

{SP2001-151} T. Koizumi, N. Hirata, K. Mori, Y. Sato, R. Hayashi, Y. Minagawa, H. Hosoi and Y. Nakajima,
``Responses in the Auditory Cortex to Moving Sound Images Measured with fMRI,''
IEICE, SP2001-151, pp.45--52, Mar. 2002.
{ functional MRI(fMRI), sound localization, moving sound image, auditory cortex, Heschl's gyrus, planum temporale }

{SP2001-152} Y. Tamekawa, H. Itoh, T. Sasaki, H. Nakahara, T. Deguchi and S. Imaizumi,
``Articulatory interaction between native and non-native phonemes: Palatographic analyses of English /r/, l/, and Japanese consonants,''
IEICE, SP2001-152, pp.53--58, Mar. 2002.
{ non-native phonetic contrasts, palatograph, multi-dimensional scaling, variation speed in the contact patters }

{SP2001-153} A. Kaoru,
``Linearity of transducer in hearing aid,''
IEICE, SP2001-153, pp.59--, Mar. 2002.
{ hearing aid, non-linearity distortion, HATS, acoustical gain }