Professor
(School of Human Sciences)
Faculty of Human Sciences(Graduate School of Human Sciences)
Faculty of Science and Engineering(Graduate School of Fundamental Science and Engineering)
Research Council (Research Organization)/Affiliated organization(Global Education Center)
研究所員 2010-2010
研究員 2004-2005
研究員 2005-2009
兼任研究員 2010-2010
研究員 2006-2010
兼任研究員 2008-2010
研究所員 2010-
研究所員 2011-2013
研究所員 2010-2015
研究所員 2015-2019
研究所員 2016-
研究所員 2013-2017
研究所員 2013-2018
研究所員 2013-2018
研究所員 2018-
研究所員 2014-2019
研究所員 2019-
Informatics / Human informatics / Intelligent informatics
藤江真也, 江尻康, 菊池英明, 小林哲則
電子情報通信学会論文誌 J88-D-II(2) p.489 - 4982005/03-
八木大三, 藤江真也, 菊池英明, 小林哲則
日本音響学会2005年春季研究発表会講演論文集 p.65 - 662005/03-
小林季実子, 菊池英明, 白井克彦
日本音響学会2004年秋季研究発表会講演論文集 p.281 - 2822004/09-
大久保崇, 菊池英明, 白井克彦
日本音響学会2004年秋季研究発表会講演論文集 p.133 - 1342004/09-
向後千春, 松居辰則, 西村昭治, 浅田匡, 菊池英明, 金群, 野嶋栄一郎
第11回日本教育メディ ア学会年次大会発表論文集 p.45 - 482004/10-
Kikuo Maekawa, Hideaki Kikuchi, and Wataru Tsukahara
International Symposium on Large-scale Knowledge Resources (LKR2004) p.19 - 242004/03-
FUJIE Shinya, YAGI Daizo, MATSUSAKA Yosuke, KIKUCHI Hideaki, KOBAYASHI Tetsunori
proc. of SP2004(International Conference Speech Prosody,2004) p.387 - 3902004/03-
早稲田大学語学教育研究所 2004/03-
八木大三, 藤江真也, 菊池英明, 小林哲則
日本音響学会2004年春季研究発表会講演論文集 p.141 - 1422004/03-
国立国語研究所 2004/03-
FUJIE Shinya, EJIRI Yasushi, MATSUSAKA Yosuke, KIKUCHI Hideaki, KOBAYASHI Tetsunori
proc. of IEEE ASRU2003(Automatic Speech Recognition and Understanding Workshop) p.231 - 2362003/12-
大久保崇,菊池英明,白井克彦
電子情報通信学会技術報告 103(519) p.235 - 2402003/12-
菊池英明, 前川喜久雄, 五十嵐陽介, 米山聖子, 藤本雅子
音声研究 7(3) p.16 - 262003/12-
藤江真也, 江尻康, 菊池英明, 小林哲則
情報処理学会音声言語情報処理研究会 2003-SLP-48p.13 - 202003/10-
藤江真也, 八木大三, 菊池英明, 小林哲則
日本音響学会2003年秋季研究発表会講演論文集 p.39 - 402003/09-
KIKUCHI Hideaki, MAEKAWA Kikuo
proc. of 15th Int'l Congress of Phonetic Sciences p.579 - 5822003/08-
MAEKAWA Kikuo, KOISO Hanae, KIKUCHI Hideaki, YONEYAMA Kiyoko
proc. of 15th Int'l Congress of Phonetic Sciences p.643 - 6442003/08-
籠宮隆之,五十嵐陽介,菊池英明,米山聖子,前川喜久雄
日本音響学会2003年春季研究発表会講演論文集 p.317 - 3182003/03-
Kikuchi, H. and K. Maekawa
proc. of the ISCA & IEEE Workshop on Spontaneous Speech Processing and Recognition (SSPR2003) p.191 - 1942003/02-
菊池英明, 塚原渉, 前川喜久雄
第3回話し言葉の科学と工学ワークショップ講演予稿集 p.21 - 262003/02-
塚原渉, 菊池英明, 前川喜久雄
第3回話し言葉の科学と工学ワークショップ講演予稿集 p.15 - 202003/02-
菊池英明, 前川喜久雄
人工知能学会言語・音声理解と対話処理研究会 SIG-SLUD-A202-06p.33 - 372002/11-
Maekawa, K., H. Kikuchi, Y. Igarashi, and J. Venditti
proc. 7th International Congress on Spoken Language Processing (ICSLP2002) p.1545 - 15482002/10-
菊池英明, 前川喜久雄
日本音響学会2002年春季研究発表会講演論文集 p.259 - 2602002/09-
菊池英明, 白井克彦
情報処理学会論文誌 Vol.43, No.7p.2130 - 21372002/07-
籠宮隆之,小磯花絵,小椋秀樹,山口昌也,菊池英明,間淵洋子,土屋菜穂子,斎藤美紀,西川賢哉,前川喜久雄
国語学会2002年度春季大会要旨集 p.225 - 2322002/05-
菊池英明, 前川喜久雄
日本音響学会2001年春季研究発表会講演論文集 p.97 - 982002/03-
前川喜久雄, 菊池英明, 五十嵐陽介
日本音響学会2002年春季研究発表会講演論文集 p.313 - 3142002/03-
菊池英明, 前川喜久雄
第2回話し言葉の科学と工学ワークショップ講演予稿集 p.53 - 582002/02-
菊池英明, 前川喜久雄, 五十嵐陽介
第2回話し言葉の科学と工学ワークショップ講演予稿集 p.19 - 262002/02-
前川喜久雄, 菊池英明, 五十嵐陽介
情報処理学会音声言語情報処理研究会 SLP-39-23p.135 - 1402001/12-
籠宮隆之, 槙洋一, 菊池英明, 前川喜久雄
日本音響学会2001年秋季研究発表会講演論文集 p.381 - 3822001/09-
鈴木堅悟, 青山一美, 菊池英明, 白井克彦
情報処理学会音声言語情報処理研究会 2001-SLP037p.13 - 182001/06-
小磯花絵, 土屋菜緒子, 間淵洋子, 斉藤美紀, 籠宮隆之, 菊池英明, 前川喜久雄
日本語科学 Vol.9p.43 - 582001/04-
菊池英明, 籠宮隆之, 前川喜久雄, 竹内京子
日本音響学会2001年春季研究発表会講演論文集 p.383 - 3842001/03-
菊池英明, 阿部賢司, 桐山伸也, 大野澄雄, 河原達也, 板橋秀一, 広瀬啓吉, 中川聖一, 堂下修二, 白井克彦, 藤崎博也
情報処理学会音声言語情報処理研究会 2001-SLP-35p.85 - 902001/02-
菊池英明
人工知能学会研究会資料 SIG-SLUD-A003-4p.21 - 242001/02-
前川喜久雄, 菊池英明, 籠宮隆之, 山口昌也, 小磯花絵, 小椋秀
日本語学, 明治書院 p.61 - 792001-
小磯花絵, 土屋菜穂子, 間淵洋子, 斉藤美紀, 籠宮隆之, 菊池英明, 前川喜久雄
電子情報通信学会技術報告 NLC2000-56, SP2000-104p.55 - 602000/12-
proc. of IEEE Int'l. Conference on Systems, Man and Cybernetics 2000/10-
proc. of Int'l. Conference on Spoken Language Processing 2000/10-
proc. of Int'l. Conference on Spoken Language Processing 2000/10-
proc. of IEEE Int'l. Conference on Systems, Man and Cybernetics 2000/10-
proc. of Int'l. Conference on Spoken Language Processing 2000/10-
第14回日本音声学会全国大会予稿集 2000/10-
日本音響学会2000年秋季研究発表会講演論文集 2000/09-
音声研究 4月2日2000/08-
人工知能学会全国大会(第14回) 6月8日2000/06-
ヒューマンインタフェース学会誌 Vol.2, No.22000/05-
情報処理学会音声言語情報処理研究会 2000-SLP-302000/02-
proc. of Eurospeech 1999/09-
proc. of Eurospeech 1999/09-
人工知能学会誌 Vol.14, No.51999/09-
proc. of ESCA Workshop Interactive Dialogue in Multi-modal Systems 1999/05-
日本音響学会講演論文集 pp.109-1101999/03-
人工知能学会 言語・音声理解と対話処理研究会予稿集 pp.1-61999/02-
情報処理学会論文誌 Vol.40, No.21999/02-
Proc. Of International Conference on Spoken Language Processing (ICSLP) pp.2351 - 23541998/12-
日本音響学会誌 54, No.111998/11-
Proc. Of International Conference on Intelligent Robots and Systems (IROS) 1, pp.255-2601998/10-
人工知能学会全国大会論文集 37-07, pp.677-6781998/06-
人工知能学会全国大会論文集 15-14, pp.242-2451998/06-
Proc. Of International Wireless and Telecommunications Symposium (IWTS) pp.322-3251998/05-
電子情報通信学会論文誌 Vol.J81-D-II, No.51998/05-
情報処理学会音声言語情報処理研究会資料 21-7, pp.69-741998/05-
日本音響学会講演論文集 3-6-16, pp.113-1141998/03-
電子情報通信学会誌 80;10, pp.1031-10351997/10-
電子情報通信学会技術研究報告 SP97-55, pp.31-361997/10-
日本音響学会講演論文集 3-1-13, pp.111-1121997/09-
International Conference on Speech Processing (ICSP) '97 Proceedings 2, pp.539-5441997/08-
人工知能学会全国大会論文集 21-06, pp.433-4361997/06-
日本音響学会誌 52巻8号1996/08-
proc. of ICMCS(International Conference on Multimedia Conputing and Systems) 1996/06-
proc. of Seventh ACM Conf. on Hypertext 1996/03-
情報処理学会全国大会講演論文集 1996/03-
情報処理学会全国大会講演論文集 1996/03-
1996/03-
proc. of IWHIT95 (International Workshop on Human Interface Technology 95) 1995/10-
proc. of HCI(Human Conputer Interaction) International '95 1995/07-
情報処理学会音声言語情報処理研究会 SLP-7-181995/07-
電子情報通信学会春季大会講演論文集 1995/03-
第10回ヒューマンインタフェースシンポジウム論文集 1994/10-
第10回ヒューマンインタフェースシンポジウム論文集 1994/10-
電子情報通信学会秋季大会講演論文集 1994/09-
電子情報通信学会論文誌 Vol.J77-D-II, No.81994/08-
電子情報通信学会春季大会講演論文集 Vol.D-1081993/03-
IEICE Trans. Inf. & Sys. Vol.E76-D, No.11993/01-
人工知能学会言語・音声理解と対話処理研究会 SIG-SLUD-9203-31992/10-
Human interface 12(4) p.335 - 3432010/11-2010/11
ISSN:13447262
Japan journal of educational technology 34(4) p.429 - 4382011/03-2011/03
ISSN:13498290
Outline:In 2002, Maekawa reported the identification of categories of paralinguistic information by subjects. Using Maekawa' s methods, Ariga extracted categories of paralinguistic information from teachers' speech during their teaching behaviors (2008). However, whereas both of the above studies used recordings of spoken examples of paralinguistic categories, speech is continuous in spoken communications. In the present study, subjects were divided into three groups under different conditions and asked to listen for categories of paralinguistic information in continuous speech to determine how such information is recognized by subjects. This revealed clear differences in the recognition of categories of paralinguistic information among the three groups.
Journal of the Phonetic Society of Japan 14(2) p.25 - 342010/08-2010/08
ISSN:13428675
Outline:An experiment on impression rating of paralinguistic information was conducted using categories of paralinguistic information expressed by teachers during teaching as speech stimuli. The aim of this study was to determine whether there are differences in the impression ratings of listeners subjected to paralinguistic information conveyed in the spontaneous speech of teachers. The experimental design involved preparing six categories of paralinguistic information as speech stimuli, and recording subject responses to 45 pairs of assessment words. Variability among subjects with regard to their impressions of categories of paralinguistic information was determined to exist.
The Transactions of the Institute of Electronics, Information and Communication Engineers. A 95(1) p.27 - 362012/01-2012/01
ISSN:09135707
The Transactions of the Institute of Electronics, Information and Communication Engineers. A 95(1) p.146 - 1562012/01-2012/01
ISSN:09135707
IEICE technical report. Speech 111(471) p.19 - 242012/03-2012/03
ISSN:0913-5685
Outline:We propose the phoneme acquisition model that simulates human auditory system that perceives the temporal variations of the speech. Our model is designed to learn values of the acoustic characteristic of a continuous speech and to estimate the number and boundaries of the phoneme categories without using explicit instructions. A purpose of this study is to clarify the role of an Infant-Directed Speech (IDS), that has the high variability. We performed a learning experiment of an Adult Directed Speech (ADS) and IDS using our model and compared the results. As a result, in the accuracy rate of the voiceless stops, IDS is significantly higher than ADS. Accordingly, it is possible that IDS emphasized the acoustic features of unsteady phonemes, and to assist the acquisition.
IEICE technical report. Speech 111(471) p.89 - 942012/03-2012/03
ISSN:0913-5685
Outline:In this research, we paid attention to impressions of a speaker which is informed by prosody that his or her voice has and make communication smoothly. We aimed to build up a 3-layer model in imitation of the process how we perceive and recognize impressions of a speaker such as feelings and physical conditions. Through this research, we expected to understand the structure of the process indirectly. In this paper, we confirmed the two effects through the model. One result showed relation of information from voices were clearly observable, the second showed expressive voice could be caught.
Technical report of IEICE. HCS 111(464) p.3 - 72012/02-2012/02
Outline:Tactile interaction of three 4-month-old infants and their mothers in play situation was observed at home. Playful repetitive arm/leg movements were observed in all mother-infant pairs and their sequential structure was analyzed. Tempo of the repetitive movement decreased with expanding moving arm/leg area of infant. Mother and infants intended to keep and enjoy their play and their interaction through infants' active physical contact and utterance was observed. Sequential features of the repetitive arm/leg movements obtained from this situation imply the importance of the tactile interaction for the early development of communication.
24(60) p.714 - 7152012/08-2012/08
IPSJ SIG technical reports 2014(22) p.1 - 62014/03-2014/03
ISSN:09196072
Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.
IPSJ SIG Notes 2014(22) p.1 - 62014/03-2014/03
ISSN:09196072
Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.
IPSJ SIG technical reports 2014(22) p.1 - 62014/03-2014/03
ISSN:09196072
Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.
IPSJ SIG technical reports 2014(22) p.1 - 62014/03-2014/03
ISSN:09196072
Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.
NAKAZATO Shu;OSHIRO Yuji;KIKUCHI Hideaki
Technical report of IEICE. HIP 112(483) p.109 - 1142013/03-2013/03
ISSN:0913-5685
Outline:In this study, we analyzed the change of manner of speech with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, we extracted four significant factors: "Liveliness", "Pleasantness", "Fluency" and "Speed". Comparing the factor scores for the low intimacy dialogues with the high intimacy dialogues, we found similar results for different partners in the low intimacy dialogues, butdifferent factor scores for different partners in the high intimacy dialogues. In particular, the fluency score increased in the dialogues after six months.
KANATO Ai;KIKUCHI Hideaki
IEICE technical report. Speech 112(422) p.49 - 542013/01-2013/01
ISSN:0913-5685
Outline:In this research, we tried to construct an evaluation scale for singing voice of popular music. In this paper, we considered the effectiveness of the scale in amateur singing voice and the factor of evaluation for singing voice. As a result, we constructed the scale with 12 words and confirmed its reliability. And, we found a characteristic factor in singing voice which differs from speaking voice.
OGAWA Yoshito;HARADA Kaho;KIKUCHI Hideaki
IEICE technical report. Speech 112(450) p.35 - 402013/02-2013/02
ISSN:0913-5685
Outline:In this study, we consider effects of an agent's grasping user features on the user's attachment to the agent. Recently, some previous studies of HAI have researched for about strategies that make users continue to use spoken dialogue systems long period of time. In this research, we suggest a system estimate user's degree of activeness from prosody and accumulate that with a correct label decided from user response as training data for following estimation. Our results show our system performs more stable estimate, and higher estimation accuracy makes users conscious more intense attachment.
MIYAJIMA Takahiro;KIKUCHI Hideaki;SHIRAI Katsuhiko;OKAWA Shigeki
Journal of the Phonetic Society of Japan 17(3) p.10 - 232013/12-2013/12
ISSN:13428675
Outline:This paper explains the procedure to enhance the expressiveness in acted speech. We designed our own "format of acting script" referring to the theory of drama and created 280 acting scripts. We presented these acting scripts as acting directions to three actresses and collected 840 speech data. For comparison, using typical emotional words as acting directions, we also collected 160 speech data from each actress. Then, we compared tendencies of various features of each data type and each speaker and found that our acting scripts are effective on the enhancement of expressiveness in acted speech psychologically/acoustically.
Takeya Yuki;Ogawa Yoshito;Kikuchi Hideaki
12(3) p.507 - 5082013/08-2013/08
Aoki Yuki;Miyajima Takahiro;Kikuchi Hideaki;Shiomi Kakuichi
12(2) p.269 - 2702013/08-2013/08
KIKUCHI Hideaki;MIYAZAWA Kouki;OGAWA Yoshito;FUJIKURA Shouhei
IEICE technical report. Speech 113(220) p.21 - 262013/09-2013/09
ISSN:0913-5685
Outline:We aimed at improvement of desire of continuing interaction with a spoken dialogue system through the three cases of construction of spoken dialogue system. System utterances with humor, control of speech rate of system utterances and estimation of user's personality based on user's utterances are effective for improvement of desire of continuing interaction.
FUJIKURA Shohei;OGAWA Yoshito;KIKUCHI Hideaki
IEICE technical report. Natural language understanding and models of communication 113(338) p.29 - 322013/11-2013/11
ISSN:0913-5685
Outline:In this study, we propose humor generate method for Non-task-oriented Dialogue System using Twitter. We have been aiming to establish the design of dialogue systems with desire of continuing interaction by analyzing the factors to feel, "want to chat with next time". We confirmed dealing humor is valid for desire of continuing interaction. In this paper, we proposes a method which dialogue system can automatically generate humor with knowledge, extract from Twitter as Modifier-Noun pair and Value-Predicate clauses pair. And in Evaluation experiment, We confirmed proposal method can generate humor.
MIYAZAWA Kouki;KAGETANI Takuya;SHEN Raymond;KIKUCHI Hideaki;OGAWA Yoshito;HATA Chihiro;OHTA Katsumi;HOZUMI Hideaki;MITAMURA Takeshi
Transactions of the Japanese Society for Artificial Intelligence 25(6) p.723 - 7322010-2010
ISSN:1346-0714
Outline:In this study, we aim at clarification of the factor that promotes an user's acceptance of suggestion from an interactive agent in driving environment. Our aim is to figure out how human beings accept the encouragement from interaction objects, and also which kinds of dialogues or action controls are necessary for the design of car navigation system which makes suggestion and requests to drivers. Firstly, we had an experiment for collecting dialogue between humans in driving simulation environment, then we analyzed the drivers' acceptance and evaluation for the navigators. As the results, we found that the presence and reliability of the navigator highly relate to the acceptance of suggestion from the navigator. When navigators were next to drivers, the rate of drivers' suggestion acceptance rose. However, the stress of drivers increased. In addition, based on the linguistic and acoustic analysis of the navigators' utterances, we found out some points of designing system utterance of suggestion to promote user's acceptance. We found that expressing the grounds of suggestions, showing the exact numbers, and the wide pitch ranges, all highly relate to the acceptance of suggestions.
MIYAZAWA Kouki;SHIROSE Ayako;MAZUKA Reiko;KIKUCHI Hideaki
J. SOFT 26(1) p.510 - 5202014-2014
ISSN:1347-7986
Outline:We assume that SOM is adequate as a language acquisition model of the native phonetic system. However, many studies don't consider the quantitative features (the appearance frequency and the number of frames of each phoneme) of the input data. Our model is designed to learn values of the acoustic characteristic of a natural continuous speech and to estimate the number and boundaries of the vowel categories without using explicit instructions. In the simulation trial, we investigate the relationship between the quantity of learning and the accuracy for the vowels in a single Japanese speaker's natural speech. As a result, it is found that the recognition accuracy rate (of our model) are 5% (/u/)-92% (/s/).
Shen Raymond;Kikuchi Hideaki
Journal of Natural Language Processing 21(3) p.445 - 4642014-2014
ISSN:1340-7619
Outline:Recent developments in computer technology have allowed the construction and widespread application of large-scale speech corpora. To enable users of speech corpora to easier data retrieval, we attempt to characterise the speaking style of speakers recorded in the corpora. We first introduce the three scales for measuring speaking style which were proposed by Eskenazi in 1993. We then use morphological features extracted from speech transcriptions that have proven effective in discriminating between styles and identifying authors in the field of natural language processing to construct an estimation model of speaking style. More specifically, we randomly choose transcriptions from various speech corpora as text stimuli with which to conduct a rating experiment on speaking style perception. Then, using the features extracted from these stimuli and rating results, we construct an estimation model of speaking style, using a multi-regression analysis. After cross-validation (leave-1-out), the results show that among the three scales of speaking style, the ratings of two scales can be estimated with high accuracy, which proves the effectiveness of our method in the estimation of speaking style.
Ichikawa Akira;Oohashi Hiroki;Naka Makiko;Kikuchi Hideaki;Horiuchi Yasuo;Kuroiwa Shingo
Studies in Science and Technology 5(1) p.113 - 1222016-2016
ISSN:2186-4942
Nakamura, Ryuji; Miyazawa, Kouki; Ishihara, Hisashi; Nishikawa, Ken'ya; Kikuchi, Hideaki; Asada, Minoru; Mazuka, Reiko
HAI 2015 - Proceedings of the 3rd International Conference on Human-Agent Interaction p.167 - 1692015/10-2015/10
Outline:© 2015 ACM.The characteristics of the spoken language used to address infants have been eagerly studied as a part of the language acquisition research. Because of the uncontrollability factor with regard to the infants, the features and roles of infantdirected speech were tried to be revealed by the comparison of speech directed toward infants and that toward other listeners. However, they share few characteristics with infants, while infants have many characteristics which may derive the features of IDS. In this study, to solve this problem, we will introduce a new approach that replaces the infant with an infant-like robot which is designed to control its motions and to imitate its appearance very similar to a real infant. We have now recorded both infant-and infantlike robot-directed speech and are constructing both corpora. Analysis of these corpora is expected to contribute to the studies of infant-directed speech. In this paper, we discuss the contents of this approach and the outline of the corpora.
Fujikura, Shohei; Ogawa, Yoshito; Kikuchi, Hideaki
HAI 2015 - Proceedings of the 3rd International Conference on Human-Agent Interaction p.171 - 1732015/10-2015/10
Outline:© 2015 ACM.We propose a humor utterance generation method that is compatible with dialogue systems, to increase "desire of continuing dialogue". A dialogue system retrieves leading-item: noun pairs from Twitter as knowledge and attempts to select the most humorous reply using word similarity, which reveals that incongruity can be explained by the incongruity-resolution model. We consider the differences among individuals, and confirm the validity of the proposed method. Ex-perimental results indicate that high-incongruity replies are significantly effective against low-incongruity replies with a limited condition.
Ogawa, Yoshito; Kikuchi, Hideaki
New Generation Computing 35(2) p.181 - 2092017/04-2017/04
ISSN:02883635
Outline:© 2017, Ohmsha, Ltd. and Springer Japan. A method to assign a personality to a spoken dialogue agent is proposed and evaluated. The proposed method assigns a personality using agent reporting about behavior independent of interaction with a user. The proposed method attempts to assigning complex personalities. For this purpose, we have defined a behavior report dialogue and designed a personality assigning method using behavior reporting. The proposed method consists of three steps: collecting stereotypes between a personality and behavior through a questionnaire, designing the behavior report dialogue from the collected stereotypes, and agent reports about behavior at the start of interactions with a user. Experimental results show that the proposed method can assign a personality by repeating the behavior report dialogue, (the assigned personality is equivalent to the personality determined by the collected stereotypes) and that reporting behavior influences the assigned personality. In addition, we verified that the proposed method can assign “kind”, “judicious” and the five basic personalities defined in the Tokyo University Egogram Second Edition.
Miyazawa, Kouki; Miyazawa, Kouki; Shinya, Takahito; Martin, Andrew; Martin, Andrew; Kikuchi, Hideaki; Mazuka, Reiko; Mazuka, Reiko
Cognition 166p.84 - 932017/09-2017/09
ISSN:00100277
Outline:© 2017 Elsevier B.V. Infant-directed speech (IDS) is known to differ from adult-directed speech (ADS) in a number of ways, and it has often been argued that some of these IDS properties facilitate infants’ acquisition of language. An influential study in support of this view is Kuhl et al. (1997), which found that vowels in IDS are produced with expanded first and second formants (F1/F2) on average, indicating that the vowels are acoustically further apart in IDS than in ADS. These results have been interpreted to mean that the way vowels are produced in IDS makes infants’ task of learning vowel categories easier. The present paper revisits this interpretation by means of a thorough analysis of IDS vowels using a large-scale corpus of Japanese natural utterances. We will show that the expansion of F1/F2 values does occur in spontaneous IDS even when the vowels’ prosodic position, lexical pitch accent, and lexical bias are accounted for. When IDS vowels are compared to carefully read speech (CS) by the same mothers, however, larger variability among IDS vowel tokens means that the acoustic distances among vowels are farther apart only in CS, but not in IDS when compared to ADS. Finally, we will show that IDS vowels are significantly more breathy than ADS or CS vowels. Taken together, our results demonstrate that even though expansion of formant values occurs in spontaneous IDS, this expansion cannot be interpreted as an indication that the acoustic distances among vowels are farther apart, as is the case in CS. Instead, we found that IDS vowels are characterized by breathy voice, which has been associated with the communication of emotional affect.
HATAOKA Nobuo, ANDO Haru, and KIKUCHI Hideaki
Ohmsha/IOS Press2005-
ISBN:4-274-90637-x
MAEKAWA Kikuo and KIKUCHI Hideaki
Mouton de Gruyter, Berlin and New York2006-
KIKUCHI Hideaki and SHIRAI Katsuhiko
丸善2006-
ISBN:4-621-07674-4
KIKUCHI Hideaki
オーム社2009-
ISBN:4274207327
KIKUCHI Hideaki
Americas Group Publications,U.S.2010-
ISBN:0935047727
KIKUCHI Hideaki
コロナ社2010-
ISBN:4339008104
KIKUCHI Hideaki
CMC出版2019-
ISBN:978-4-7813-1430-3
Reference Number:19
連続音声認識装置および方法(日本)白井 克彦, 城崎 康夫, 菊池 英明
H11-281437、2001-109491
Reference Number:481
音声認識装置及び音声認識用プログラム(日本)白井 克彦, 菊池 英明, 大久保 崇
2005- 80732、2006-259641
Reference Number:756
検索語クラスタリング装置、検索語クラスタリング方法、検索語クラスタリングプログラム及び記録媒体(日本)白井 克彦, 菊池 英明, 新関 一馬
2007-193700、2009- 31931
Research Classification:
Fundamental Study for Conversion between Spoken and Written Japanese Considering Influence of Interactivity2013/-0-2016/-0
Allocation Class:¥4680000
Research Classification:
On the physical factors which makes the mother tongue dialogues smoothly - through the comparison with the non-mother tongue2012/-0-2015/-0
Allocation Class:¥3900000
Research Classification:
Analysis of infant-directed speech by acoustic and computational modeling methods.Allocation Class:¥4810000
Research Classification:
A Study on a framework of spontaneous communication depending on dialogue situationAllocation Class:¥15360000
Research Classification:
Quantitative Analysis of Linguistic Variation Using a Spontaneous Speech CorpusAllocation Class:¥14800000
Research Classification:
Analysis of effective communication process by using body and artifact for faculty DevelopmentAllocation Class:¥3700000
Research Classification:
Development of new prototypes and models of higher education utilizing broadband networks.Allocation Class:¥8300000
Research Classification:
XML documentation of complex annotation on spontaneous speech dataAllocation Class:¥3900000
Research Classification:
Elaboration of articulatory phonetics by means of realtime-MRI and WAVE data2017/-0-2020/-0
Allocation Class:¥17810000
Research Classification:
Estimation of User's Impression Space for Improving Desire of Interaction with Spoken Dialogue System2014/-0-2017/-0
Allocation Class:¥3900000
Research Classification:
Preparation of Database and Analysis of Dialectial Difference in Phonetics, Phonology, Grammar and Vocabulary among Dialects of Ainu2012/-0-2015/-0
Allocation Class:¥4940000
Research Classification:
Development of protocol of effective social work interview for older adults with dementia2019/-0-2022/-0
Allocation Class:¥6890000
Research Classification:
Real-time MRI database of articulatory movements of Japanese2020/-0-2024/-0
Allocation Class:¥17420000
2013
Research Results Outline:音声対話システムのシステム発話を制御することによってユーザに与える印象を変化させる技術の開発を目指している。本特定課題研究では、擬人化したシステムの自音声対話システムのシステム発話を制御することによってユーザに与える印象を変化させる技術の開発を目指している。本特定課題研究では、擬人化したシステムの自己開示によってパーソナリティを付与する手法に関する基礎研究を中心に進めた。実験を通じて、自己開示量...音声対話システムのシステム発話を制御することによってユーザに与える印象を変化させる技術の開発を目指している。本特定課題研究では、擬人化したシステムの自己開示によってパーソナリティを付与する手法に関する基礎研究を中心に進めた。実験を通じて、自己開示量と内容によって特定のパーソナリティを付与できることを確認した。この成果は、ヒューマンインタフェース学会論文誌に査読論文として掲載された(「自己開示による音声対話エージェントへのパーソナリティ付与」)。他にも、マイクロブログからユーモア発話を自動生成する技術(「非タスク指向対話システムにおけるマイクロブログを用いたユーモア発話の自動生成」)、発話速度あるいは無音区間長を制御する手法(「ロボット発話の話速・無音区間長の制御によるパーソナリティ認知と対話継続欲求の向上」)を検討し、それぞれによってユーザがシステムに抱く印象がどのように変化するかを実験により調査した。いずれも国内学会にて成果を発表した。ユーザ発話における音声のプロソディを解析することによってユーザの心的状態を推定して、それに応じてシステムの振舞を変えることによってユーザが抱く愛着感を変化させられることを実験により確認し、この成果を国際会議にて発表した(「Effects of an Agent Feature Comprehension on the Emotional Attachment of Users」)。いずれの研究においても、ユーザが抱く印象を変化させるこれらの手法によって、音声対話システムに対するインタラクション継続欲求が向上することを確認している。このことを一旦整理して国内学会にて発表した(「音声対話システムに対するインタラクション継続欲求」)。こうした一連の研究成果を体系化してさらに幅広く応用可能な技術を開発するために、インタラクション継続欲求とユーザ印象空間の関係を明確にする必要性が生じている。そこで最終的に、本特定課題研究を経て、2014年度科学研究費の基盤(C)に「音声対話システムに対するインタラクション欲求向上のためのユーザ印象空間の推定」というテーマを申請して採択された。
2006
Research Results Outline: 本研究では、音声からの心的状態の推定において話者の心的状況をより高い精度で推定するため、生体情報を教師信号としたモデル学習を行う“生理心理学的アプロ 本研究では、音声からの心的状態の推定において話者の心的状況をより高い精度で推定するため、生体情報を教師信号としたモデル学習を行う“生理心理学的アプローチの導入”を提案する。 従来の感情推定は、モデル学習の際に実験者の判断による評定結果が教師信号と... 本研究では、音声からの心的状態の推定において話者の心的状況をより高い精度で推定するため、生体情報を教師信号としたモデル学習を行う“生理心理学的アプローチの導入”を提案する。 従来の感情推定は、モデル学習の際に実験者の判断による評定結果が教師信号として用いられるため主観的方法であることが否めない。また推定の対象も基本的な感情にのみ重点が置かれてきた。 生体情報は、意図的な操作が入らず継時的な変化を捉えられることができるとされている。そのため、推定を行う際に実験の第一段階として生体信号を利用することで、多様で連続的な心情の変化を対象とすることができるようになり、またより客観的で精度の高い判断が可能になると思われる。 難度の異なる音読課題を2つ用意し、課題間における生体信号の反応の違いが音声の違いにも現れるのかを観察した。実験者の主観的評価によってストレス状態と判断された被験者の音声と、それらのうち生体信号の変化からもストレス状態にあると判断できた被験者の音声の比較を行う。 生体信号には、心的状態の推定へ利用できると思われた容積脈派(BVP)、心電図(EKG)、皮膚温(TEMP)、皮膚コンダクタンス(SC)を用いた。 音声の比較には、各音声からF0とパワーそれぞれの最大値、最小値、振幅、平均値、それに発話速度を加えた9つの特徴量を抽出し、これらを決定木学習に利用した。決定木学習には、C4.5アルゴリズムを使用し、交差検定を用いて評価を行う。 全データ(実験者の主観的評価のみによってストレス状態を判断した)で学習モデルを生成した場合平均63.9%であった判別率が、選別データ(主観的評価に加え、生体信号の変化からもストレス状態を判断した)で学習モデルを生成した場合には平均77.8%まで精度が向上した。 生体信号がストレス状態を判断するうえで一つの指標となり得ることを示唆する結果となった。本実験の結果、音声からの心的状態の推定を行う際に生体情報を利用することの有益性が実証された。
2009/03-2010/03
Affiliation: オハイオ州立大学(アメリカ)、北京大学(中国)
Course Title | School | Year | Term |
---|---|---|---|
Research Method for Measurement and Modeling of Human 01 | School of Human Sciences | 2020 | fall quarter |
Research Method for Measurement and Modeling of Human 02 | School of Human Sciences | 2020 | fall quarter |
Research Method for Measurement and Modeling of Human 03 | School of Human Sciences | 2020 | fall quarter |
Seminar II (Language and Information Science) | School of Human Sciences | 2020 | fall semester |
Language and Information Sciences | School of Human Sciences (Online Degree Program) | 2020 | fall semester |
Language and Information Sciences A | Graduate School of Human Sciences | 2020 | spring semester |
Language and Information Sciences B | Graduate School of Human Sciences | 2020 | fall semester |
Language and Information Sciences(1) A | Graduate School of Human Sciences | 2020 | spring semester |
Language and Information Sciences(1) B | Graduate School of Human Sciences | 2020 | fall semester |
Language and Information Sciences(1) C | Graduate School of Human Sciences | 2020 | spring semester |
Language and Information Sciences(1) D | Graduate School of Human Sciences | 2020 | fall semester |
Language and Information Sciences(2) A | Graduate School of Human Sciences | 2020 | spring semester |
Language and Information Sciences(2) B | Graduate School of Human Sciences | 2020 | fall semester |
Language and Information Sciences(2) C | Graduate School of Human Sciences | 2020 | spring semester |
Language and Information Sciences(2) D | Graduate School of Human Sciences | 2020 | fall semester |
Language and Information Sciences | Graduate School of Human Sciences | 2020 | fall quarter |
Language and Information Sciences(D) A | Graduate School of Human Sciences | 2020 | spring semester |
Language and Information Sciences(D) B | Graduate School of Human Sciences | 2020 | fall semester |