Name

KIKUCHI, Hideaki

Official Title

Professor

Affiliation

(School of Human Sciences)

Contact Information

Mail Address

Mail Address
kikuchi@waseda.jp

URL

Web Page URL

https://sites.google.com/view/kikuchihideaki/home

http://www.f.waseda.jp/kikuchi/

Grant-in-aids for Scientific Researcher Number
70308261

Sub-affiliation

Sub-affiliation

Faculty of Human Sciences(Graduate School of Human Sciences)

Faculty of Science and Engineering(Graduate School of Fundamental Science and Engineering)

Research Council (Research Organization)/Affiliated organization(Global Education Center)

Affiliated Institutes

知覚情報システム研究所

研究所員 2010-2010

空間科学研究所

研究員 2004-2005

音声_認識_基盤技術研究所

研究員 2005-2009

メディアネットワークセンター

兼任研究員 2010-2010

コミュニケーション科学研究所

研究員 2006-2010

メディアネットワークセンター

兼任研究員 2008-2010

応用脳科学研究所

研究所員 2010-

波動コミュニケーション研究所

研究所員 2011-2013

知覚情報システム研究所

研究所員 2010-2015

知覚情報システム研究所

研究所員 2015-2019

波動場・コミュニケーション科学研究所

研究所員 2016-

メディア研究所

研究所員 2013-2017

移民・エスニック文化研究所

研究所員 2013-2018

ことばの科学研究所

研究所員 2013-2018

ことばの科学研究所

研究所員 2018-

応用脳科学研究所

研究所員 2014-2019

応用脳科学研究所

研究所員 2019-

Research Field

Keywords

Speech Science, Spoken Dialogue, Human Agent Interaction

Grants-in-Aid for Scientific Research classification

Informatics / Human informatics / Intelligent informatics

Paper

肯定的/否定的発話態度の認識とその音声対話システムへの応用

藤江真也, 江尻康, 菊池英明, 小林哲則

電子情報通信学会論文誌 J88-D-II(2) p.489 - 4982005/03-

韻律情報を用いた発話態度認識とその対話システムへの応用

八木大三, 藤江真也, 菊池英明, 小林哲則

日本音響学会2005年春季研究発表会講演論文集 p.65 - 662005/03-

音声対話における発話の感情判別

小林季実子, 菊池英明, 白井克彦

日本音響学会2004年秋季研究発表会講演論文集 p.281 - 2822004/09-

韻律情報を利用した文章入力システムのための韻律制御モデル

大久保崇, 菊池英明, 白井克彦

日本音響学会2004年秋季研究発表会講演論文集 p.133 - 1342004/09-

早稲田大学eスクールの実践:大学教育におけるeラーニングの展望

向後千春, 松居辰則, 西村昭治, 浅田匡, 菊池英明, 金群, 野嶋栄一郎

第11回日本教育メディ ア学会年次大会発表論文集 p.45 - 482004/10-

Corpus of Spontaneous Japanese: Design, Annotation and XML Representation

Kikuo Maekawa, Hideaki Kikuchi, and Wataru Tsukahara

International Symposium on Large-scale Knowledge Resources (LKR2004) p.19 - 242004/03-

Spoken Dialogue System Using Prosody As Para-Linguistic Information

FUJIE Shinya, YAGI Daizo, MATSUSAKA Yosuke, KIKUCHI Hideaki, KOBAYASHI Tetsunori

proc. of SP2004(International Conference Speech Prosody,2004) p.387 - 3902004/03-

アイヌ語音声データベース

早稲田大学語学教育研究所 2004/03-

韻律情報を用いた肯定的/否定的態度の認識

八木大三, 藤江真也, 菊池英明, 小林哲則

日本音響学会2004年春季研究発表会講演論文集 p.141 - 1422004/03-

日本語話し言葉コーパス

国立国語研究所 2004/03-

Recognition of Para-Linguistic Information and Its Application to Spoken Dialogue System

FUJIE Shinya, EJIRI Yasushi, MATSUSAKA Yosuke, KIKUCHI Hideaki, KOBAYASHI Tetsunori

proc. of IEEE ASRU2003(Automatic Speech Recognition and Understanding Workshop) p.231 - 2362003/12-

音声対話における韻律を用いた話題境界検出

大久保崇,菊池英明,白井克彦

電子情報通信学会技術報告 103(519) p.235 - 2402003/12-

日本語話し言葉コーパスの音声ラベリング

菊池英明, 前川喜久雄, 五十嵐陽介, 米山聖子, 藤本雅子

音声研究 7(3) p.16 - 262003/12-

パラ言語の理解能力を有する対話ロボット

藤江真也, 江尻康, 菊池英明, 小林哲則

情報処理学会音声言語情報処理研究会 2003-SLP-48p.13 - 202003/10-

パラ言語情報を用いた音声対話システム

藤江真也, 八木大三, 菊池英明, 小林哲則

日本音響学会2003年秋季研究発表会講演論文集 p.39 - 402003/09-

Evaluation of the effectiveness of "X-JToBI": A new prosodic labeling scheme for spontaneous Japanese speech

KIKUCHI Hideaki, MAEKAWA Kikuo

proc. of 15th Int'l Congress of Phonetic Sciences p.579 - 5822003/08-

Use of a large-scale spontaneous speech corpus in the study of linguistic variation

MAEKAWA Kikuo, KOISO Hanae, KIKUCHI Hideaki, YONEYAMA Kiyoko

proc. of 15th Int'l Congress of Phonetic Sciences p.643 - 6442003/08-

自発音声コーパスにおけるF0下降開始位置の分析

籠宮隆之,五十嵐陽介,菊池英明,米山聖子,前川喜久雄

日本音響学会2003年春季研究発表会講演論文集 p.317 - 3182003/03-

Performance of segmental and prosodic labeling of spontaneous speech

Kikuchi, H. and K. Maekawa

proc. of the ISCA & IEEE Workshop on Spontaneous Speech Processing and Recognition (SSPR2003) p.191 - 1942003/02-

XMLを利用した『日本語話し言葉コーパス』(CSJ)の整合性検証

菊池英明, 塚原渉, 前川喜久雄

第3回話し言葉の科学と工学ワークショップ講演予稿集 p.21 - 262003/02-

『日本語話し言葉コーパス』(CSJ)のXML検索環境

塚原渉, 菊池英明, 前川喜久雄

第3回話し言葉の科学と工学ワークショップ講演予稿集 p.15 - 202003/02-

日本語自発音声韻律ラベリングスキームX-JToBIの能力検証

菊池英明, 前川喜久雄

人工知能学会言語・音声理解と対話処理研究会 SIG-SLUD-A202-06p.33 - 372002/11-

X-JToBI: An extended J_ToBI for spontaneous speech

Maekawa, K., H. Kikuchi, Y. Igarashi, and J. Venditti

proc. 7th International Congress on Spoken Language Processing (ICSLP2002) p.1545 - 15482002/10-

自発音声韻律ラベリングスキームX-JToBIによるラベリング精度の検証

菊池英明, 前川喜久雄

日本音響学会2002年春季研究発表会講演論文集 p.259 - 2602002/09-

音声対話における心的状態変化の予測をともなうメタ発話生成機構

菊池英明, 白井克彦

情報処理学会論文誌 Vol.43, No.7p.2130 - 21372002/07-

大規模自発音声コーパス『日本語話し言葉コーパス』の仕様と作成

籠宮隆之,小磯花絵,小椋秀樹,山口昌也,菊池英明,間淵洋子,土屋菜穂子,斎藤美紀,西川賢哉,前川喜久雄

国語学会2002年度春季大会要旨集 p.225 - 2322002/05-

自発音声に対する音素自動ラベリング精度の検証

菊池英明, 前川喜久雄

日本音響学会2001年春季研究発表会講演論文集 p.97 - 982002/03-

日本語自発音声の韻律ラベリング体系: X-JToBI

前川喜久雄, 菊池英明, 五十嵐陽介

日本音響学会2002年春季研究発表会講演論文集 p.313 - 3142002/03-

自発音声に対する音素自動ラベリング精度の検証

菊池英明, 前川喜久雄

第2回話し言葉の科学と工学ワークショップ講演予稿集 p.53 - 582002/02-

日本語自発音声の韻律ラベリングスキーム: X-JToBI

菊池英明, 前川喜久雄, 五十嵐陽介

第2回話し言葉の科学と工学ワークショップ講演予稿集 p.19 - 262002/02-

X-JToBI: 自発音声の韻律ラベリングスキーム

前川喜久雄, 菊池英明, 五十嵐陽介

情報処理学会音声言語情報処理研究会 SLP-39-23p.135 - 1402001/12-

自発音声コーパスにおける印象評定とその要因

籠宮隆之, 槙洋一, 菊池英明, 前川喜久雄

日本音響学会2001年秋季研究発表会講演論文集 p.381 - 3822001/09-

多次元心的状態を扱う音声対話システムの構築

鈴木堅悟, 青山一美, 菊池英明, 白井克彦

情報処理学会音声言語情報処理研究会 2001-SLP037p.13 - 182001/06-

「日本語話し言葉コーパス」における書き起こしの方法とその基準について

小磯花絵, 土屋菜緒子, 間淵洋子, 斉藤美紀, 籠宮隆之, 菊池英明, 前川喜久雄

日本語科学 Vol.9p.43 - 582001/04-

自発音声に対するJ_ToBIラベリングの問題点検討

菊池英明, 籠宮隆之, 前川喜久雄, 竹内京子

日本音響学会2001年春季研究発表会講演論文集 p.383 - 3842001/03-

音声対話に基づく知的情報検索システム

菊池英明, 阿部賢司, 桐山伸也, 大野澄雄, 河原達也, 板橋秀一, 広瀬啓吉, 中川聖一, 堂下修二, 白井克彦, 藤崎博也

情報処理学会音声言語情報処理研究会 2001-SLP-35p.85 - 902001/02-

日本語音声への韻律ラベリング

菊池英明

人工知能学会研究会資料 SIG-SLUD-A003-4p.21 - 242001/02-

『日本語話し言葉コーパス』の構築における計算機利用

前川喜久雄, 菊池英明, 籠宮隆之, 山口昌也, 小磯花絵, 小椋秀

日本語学, 明治書院 p.61 - 792001-

『日本語話し言葉コーパス』の書き起こし基準について

小磯花絵, 土屋菜穂子, 間淵洋子, 斉藤美紀, 籠宮隆之, 菊池英明, 前川喜久雄

電子情報通信学会技術報告 NLC2000-56, SP2000-104p.55 - 602000/12-

Controlling Non-verbal Information in Speaker-change for Spoken Dialogue

proc. of IEEE Int'l. Conference on Systems, Man and Cybernetics 2000/10-

Designing a Domain Independent Platform of Spoken Dialogue System

proc. of Int'l. Conference on Spoken Language Processing 2000/10-

Improvement of Dialogue Efficiency by Dialogue Control Model According to Performance of Processes

proc. of Int'l. Conference on Spoken Language Processing 2000/10-

Modeling of Spoken Dialogue Control for Improvement of Dialogue Efficiency

proc. of IEEE Int'l. Conference on Systems, Man and Cybernetics 2000/10-

Overview of an Intelligent System for Information Retrieval Based on Human-Machine Dialogue through Spoken Language

proc. of Int'l. Conference on Spoken Language Processing 2000/10-

モノローグを対象とした自発音声コーパス:その設計について

第14回日本音声学会全国大会予稿集 2000/10-

大規模話し言葉コーパスにおける発話スタイルの諸相---書き起こしテキストの分析から---

日本音響学会2000年秋季研究発表会講演論文集 2000/09-

日本語話し言葉コーパスの設計

音声研究 4月2日2000/08-

音声対話システム汎用プラットフォームにおける行動管理部の構築

人工知能学会全国大会(第14回) 6月8日2000/06-

対話効率の向上を目的とした音声対話制御のモデル化

ヒューマンインタフェース学会誌 Vol.2, No.22000/05-

音声対話システム汎用プラットフォームの検討

情報処理学会音声言語情報処理研究会 2000-SLP-302000/02-

A Post-Processing of Speech for Hearing Impaired Integrate into Standard Digital Audio Decorders

proc. of Eurospeech 1999/09-

Improving Recognition Correct Rate of Important Words in Large Vocabulary Speech Recognition

proc. of Eurospeech 1999/09-

課題遂行対話における対話潤滑語の認定

人工知能学会誌 Vol.14, No.51999/09-

Controlling Dialogue Strategy According to Performance of Processes

proc. of ESCA Workshop Interactive Dialogue in Multi-modal Systems 1999/05-

音声対話システムにおける処理性能と対話戦略の関係についての一考察

日本音響学会講演論文集 pp.109-1101999/03-

システムの処理性能を考慮した対話制御方法の検討

人工知能学会 言語・音声理解と対話処理研究会予稿集 pp.1-61999/02-

人間型ロボットの対話インタフェースにおける発話交替時の非言語情報の制御

情報処理学会論文誌 Vol.40, No.21999/02-

Use of Nonverbal Information in Communication between Human and Robot

Proc. Of International Conference on Spoken Language Processing (ICSLP) pp.2351 - 23541998/12-

非言語的現象の分析と対話処理 -電子メール討論

日本音響学会誌 54, No.111998/11-

Controlling Gaze of Humanoid in Communication with Human

Proc. Of International Conference on Intelligent Robots and Systems (IROS) 1, pp.255-2601998/10-

時間的制約を考慮した対話制御方法の実現方法

人工知能学会全国大会論文集 37-07, pp.677-6781998/06-

人間型対話インタフェースにおけるまばたき制御の検討

人工知能学会全国大会論文集 15-14, pp.242-2451998/06-

Multimodal Communication Between Human and Robot

Proc. Of International Wireless and Telecommunications Symposium (IWTS) pp.322-3251998/05-

情報のら旋成長を支援するコミュニケーション形電子図書館

電子情報通信学会論文誌 Vol.J81-D-II, No.51998/05-

人間とロボットのコミュニケーションにおける非言語情報の利用

情報処理学会音声言語情報処理研究会資料 21-7, pp.69-741998/05-

時間的制約を考慮した対話制御方法の検討

日本音響学会講演論文集 3-6-16, pp.113-1141998/03-

音声を利用したマルチモーダルインタフェース

電子情報通信学会誌 80;10, pp.1031-10351997/10-

自由会話における時間的制約の影響の分析

電子情報通信学会技術研究報告 SP97-55, pp.31-361997/10-

複数ユーザとロボットの対話における非言語情報の役割

日本音響学会講演論文集 3-1-13, pp.111-1121997/09-

The Role of Non-Verbal Information in Spoken Dialogue between a Man and a Robot

International Conference on Speech Processing (ICSP) '97 Proceedings 2, pp.539-5441997/08-

ロボットとの対話における非言語情報の役割

人工知能学会全国大会論文集 21-06, pp.433-4361997/06-

音響学会員のためのインターネット概説

日本音響学会誌 52巻8号1996/08-

User Interface for a Digital Library to Support Construction of a "Virtual Personal Library"

proc. of ICMCS(International Conference on Multimedia Conputing and Systems) 1996/06-

Extensions of World-wide Aiming at the construction of a "Virtual Personal Library"

proc. of Seventh ACM Conf. on Hypertext 1996/03-

ハイパーメディア共有アーキテクチャ

情報処理学会全国大会講演論文集 1996/03-

ハイパーメディア共有アーキテクチャにおけるバージョン管理方式

情報処理学会全国大会講演論文集 1996/03-

情報処理学会第53回全国大会大会優秀賞

1996/03-

Multimodal Interface Using Speech and Pointing Gestures, and Its Applications for Interior Design and PC Window Manipulation

proc. of IWHIT95 (International Workshop on Human Interface Technology 95) 1995/10-

Agent-typed Multimodal Interface Using Speech, Pointing Gestures, and CG

proc. of HCI(Human Conputer Interaction) International '95 1995/07-

音声とペンを入力手段とするマルチモーダルインタフェースの構築

情報処理学会音声言語情報処理研究会 SLP-7-181995/07-

仮想個人図書館の構築を支援するユーザインタフェースの開発

電子情報通信学会春季大会講演論文集 1995/03-

マルチモーダルウインドウシステムの構築

第10回ヒューマンインタフェースシンポジウム論文集 1994/10-

音声・ポインティング・CGによるエージェント型ユーザインタフェースの試作と評価

第10回ヒューマンインタフェースシンポジウム論文集 1994/10-

音声・ポインティング・CGによるエージェント型ユーザインタフェースシステム

電子情報通信学会秋季大会講演論文集 1994/09-

音声対話インタフェースにおける発話権管理による割り込みへの対処

電子情報通信学会論文誌 Vol.J77-D-II, No.81994/08-

音声対話システムにおける発話権の制御

電子情報通信学会春季大会講演論文集 Vol.D-1081993/03-

Three Different LR Parsing Algorithm for Phoneme-Context-Dependent HMM-Based Continuous Speech Recognition

IEICE Trans. Inf. & Sys. Vol.E76-D, No.11993/01-

ナビゲーションシステムにおける音声対話インタフェースの構成

人工知能学会言語・音声理解と対話処理研究会 SIG-SLUD-9203-31992/10-

A Method of Estimation of Stress for Evaluation of Speech Interface

Human interface 12(4) p.335 - 3432010/11-2010/11

CiNii

Detail

ISSN:13447262

Speech Perception of Paralinguistic Information Expressed in Teacher's Continuous Speech

Japan journal of educational technology 34(4) p.429 - 4382011/03-2011/03

CiNii

Detail

ISSN:13498290

Outline:In 2002, Maekawa reported the identification of categories of paralinguistic information by subjects. Using Maekawa' s methods, Ariga extracted categories of paralinguistic information from teachers' speech during their teaching behaviors (2008). However, whereas both of the above studies used recordings of spoken examples of paralinguistic categories, speech is continuous in spoken communications. In the present study, subjects were divided into three groups under different conditions and asked to listen for categories of paralinguistic information in continuous speech to determine how such information is recognized by subjects. This revealed clear differences in the recognition of categories of paralinguistic information among the three groups.

Impression Rating of Categories of Paralinguistic Information in Spontaneous Speech through the Use of Assessment Words

Journal of the Phonetic Society of Japan 14(2) p.25 - 342010/08-2010/08

CiNii

Detail

ISSN:13428675

Outline:An experiment on impression rating of paralinguistic information was conducted using categories of paralinguistic information expressed by teachers during teaching as speech stimuli. The aim of this study was to determine whether there are differences in the impression ratings of listeners subjected to paralinguistic information conveyed in the spontaneous speech of teachers. The experimental design involved preparing six categories of paralinguistic information as speech stimuli, and recording subject responses to 45 pairs of assessment words. Variability among subjects with regard to their impressions of categories of paralinguistic information was determined to exist.

9-215 Current status of e-Learning at Waseda University

23(59) p.436 - 4372011/08-2011/08

CiNii

Factors of Interaction in the Spoken Dialogue System with High Desire of Sustainability

The Transactions of the Institute of Electronics, Information and Communication Engineers. A 95(1) p.27 - 362012/01-2012/01

CiNii

Detail

ISSN:09135707

Comparison between Spoken and Written Utterances Describing the Same Object

The Transactions of the Institute of Electronics, Information and Communication Engineers. A 95(1) p.146 - 1562012/01-2012/01

CiNii

Detail

ISSN:09135707

Acoustic Analysis of Infant Directed Speech Using the Multi-Timescale Phoneme Acquisition Model

IEICE technical report. Speech 111(471) p.19 - 242012/03-2012/03

CiNii

Detail

ISSN:0913-5685

Outline:We propose the phoneme acquisition model that simulates human auditory system that perceives the temporal variations of the speech. Our model is designed to learn values of the acoustic characteristic of a continuous speech and to estimate the number and boundaries of the phoneme categories without using explicit instructions. A purpose of this study is to clarify the role of an Infant-Directed Speech (IDS), that has the high variability. We performed a learning experiment of an Adult Directed Speech (ADS) and IDS using our model and compared the results. As a result, in the accuracy rate of the voiceless stops, IDS is significantly higher than ADS. Accordingly, it is possible that IDS emphasized the acoustic features of unsteady phonemes, and to assist the acquisition.

Modeling of perception and cognition process of impressions of speaker

IEICE technical report. Speech 111(471) p.89 - 942012/03-2012/03

CiNii

Detail

ISSN:0913-5685

Outline:In this research, we paid attention to impressions of a speaker which is informed by prosody that his or her voice has and make communication smoothly. We aimed to build up a 3-layer model in imitation of the process how we perceive and recognize impressions of a speaker such as feelings and physical conditions. Through this research, we expected to understand the structure of the process indirectly. In this paper, we confirmed the two effects through the model. One result showed relation of information from voices were clearly observable, the second showed expressive voice could be caught.

Sequential Structure Analysis of Repetitive Movements and Utterances in Mother-infant Play Interaction

Technical report of IEICE. HCS 111(464) p.3 - 72012/02-2012/02

CiNii

Detail

Outline:Tactile interaction of three 4-month-old infants and their mothers in play situation was observed at home. Playful repetitive arm/leg movements were observed in all mother-infant pairs and their sequential structure was analyzed. Tempo of the repetitive movement decreased with expanding moving arm/leg area of infant. Mother and infants intended to keep and enjoy their play and their interaction through infants' active physical contact and utterance was observed. Sequential features of the repetitive arm/leg movements obtained from this situation imply the importance of the tactile interaction for the early development of communication.

8-329 Challenges for opening course contents at Waseda University : Focusing on issues toward "Waseda Vision 150"

24(60) p.714 - 7152012/08-2012/08

CiNii

The Relationship between the Level of Intimacy and the Vocabulary Used in Dialogues

IPSJ SIG technical reports 2014(22) p.1 - 62014/03-2014/03

CiNii

Detail

ISSN:09196072

Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.

The Relationship between the Level of Intimacy and the Vocabulary Used in Dialogues

IPSJ SIG Notes 2014(22) p.1 - 62014/03-2014/03

CiNii

Detail

ISSN:09196072

Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.

The Relationship between the Level of Intimacy and the Vocabulary Used in Dialogues

IPSJ SIG technical reports 2014(22) p.1 - 62014/03-2014/03

CiNii

Detail

ISSN:09196072

Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.

The Relationship between the Level of Intimacy and the Vocabulary Used in Dialogues

IPSJ SIG technical reports 2014(22) p.1 - 62014/03-2014/03

CiNii

Detail

ISSN:09196072

Outline:In this study, we analyzed the change in spoken vocabulary with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, for 23 factors we found differences for 7 factors: "Polite", "Powerful", "Cheerful", "Considerate", "Skillful", "Good-volume" and "Clear". Comparing the frequency of words used in low and high intimacy dialogue, we found a high frequency appearance of "desu", "masu" and "hai" forms in the low intimacy dialogues and the use of "un", "kedo", "maji" and "ore" in the high intimacy dialogues.

The Relationship between the Level of Intimacy and Manner of Speech

NAKAZATO Shu;OSHIRO Yuji;KIKUCHI Hideaki

Technical report of IEICE. HIP 112(483) p.109 - 1142013/03-2013/03

CiNii

Detail

ISSN:0913-5685

Outline:In this study, we analyzed the change of manner of speech with the level of intimacy. For our experiments, we collected two kinds of dialogue data: the initial meeting dialogue data we called "low intimacy" and that after six months we called "high intimacy". In our experiments, subjects listened to the dialogue of three pairs of speakers, and evaluated their impressions of the manner of speech through a questionnaire. Analyzing the results, we extracted four significant factors: "Liveliness", "Pleasantness", "Fluency" and "Speed". Comparing the factor scores for the low intimacy dialogues with the high intimacy dialogues, we found similar results for different partners in the low intimacy dialogues, butdifferent factor scores for different partners in the high intimacy dialogues. In particular, the fluency score increased in the dialogues after six months.

The construction of an evaluation scale for singing voice of popular music : in Amateur singing voice

KANATO Ai;KIKUCHI Hideaki

IEICE technical report. Speech 112(422) p.49 - 542013/01-2013/01

CiNii

Detail

ISSN:0913-5685

Outline:In this research, we tried to construct an evaluation scale for singing voice of popular music. In this paper, we considered the effectiveness of the scale in amateur singing voice and the factor of evaluation for singing voice. As a result, we constructed the scale with 12 words and confirmed its reliability. And, we found a characteristic factor in singing voice which differs from speaking voice.

Effects of an Agent's Feature Grasping on an User's Attachment

OGAWA Yoshito;HARADA Kaho;KIKUCHI Hideaki

IEICE technical report. Speech 112(450) p.35 - 402013/02-2013/02

CiNii

Detail

ISSN:0913-5685

Outline:In this study, we consider effects of an agent's grasping user features on the user's attachment to the agent. Recently, some previous studies of HAI have researched for about strategies that make users continue to use spoken dialogue systems long period of time. In this research, we suggest a system estimate user's degree of activeness from prosody and accumulate that with a correct label decided from user response as training data for following estimation. Our results show our system performs more stable estimate, and higher estimation accuracy makes users conscious more intense attachment.

Effect of Schemed Acting Directions on Speech Expressions : Toward the Achievement of Expressive Acted Speech

MIYAJIMA Takahiro;KIKUCHI Hideaki;SHIRAI Katsuhiko;OKAWA Shigeki

Journal of the Phonetic Society of Japan 17(3) p.10 - 232013/12-2013/12

CiNii

Detail

ISSN:13428675

Outline:This paper explains the procedure to enhance the expressiveness in acted speech. We designed our own "format of acting script" referring to the theory of drama and created 280 acting scripts. We presented these acting scripts as acting directions to three actresses and collected 840 speech data. For comparison, using typical emotional words as acting directions, we also collected 160 speech data from each actress. Then, we compared tendencies of various features of each data type and each speaker and found that our acting scripts are effective on the enhancement of expressiveness in acted speech psychologically/acoustically.

J-054 Personality Recognition and Improvement of Dialogue Continuance Desire Using Control of Speech Rate and Speech Interval Length

Takeya Yuki;Ogawa Yoshito;Kikuchi Hideaki

12(3) p.507 - 5082013/08-2013/08

CiNii

E-037 Development of Fatigue Degree Estimation System for Smartphone

Aoki Yuki;Miyajima Takahiro;Kikuchi Hideaki;Shiomi Kakuichi

12(2) p.269 - 2702013/08-2013/08

CiNii

Desire of Continuing Interaction with Spoken Dialogue System

KIKUCHI Hideaki;MIYAZAWA Kouki;OGAWA Yoshito;FUJIKURA Shouhei

IEICE technical report. Speech 113(220) p.21 - 262013/09-2013/09

CiNii

Detail

ISSN:0913-5685

Outline:We aimed at improvement of desire of continuing interaction with a spoken dialogue system through the three cases of construction of spoken dialogue system. System utterances with humor, control of speech rate of system utterances and estimation of user's personality based on user's utterances are effective for improvement of desire of continuing interaction.

Humor Utterance Generation Method for Non-task-oriented Dialogue System

FUJIKURA Shohei;OGAWA Yoshito;KIKUCHI Hideaki

IEICE technical report. Natural language understanding and models of communication 113(338) p.29 - 322013/11-2013/11

CiNii

Detail

ISSN:0913-5685

Outline:In this study, we propose humor generate method for Non-task-oriented Dialogue System using Twitter. We have been aiming to establish the design of dialogue systems with desire of continuing interaction by analyzing the factors to feel, "want to chat with next time". We confirmed dealing humor is valid for desire of continuing interaction. In this paper, we proposes a method which dialogue system can automatically generate humor with knowledge, extract from Twitter as Modifier-Noun pair and Value-Predicate clauses pair. And in Evaluation experiment, We confirmed proposal method can generate humor.

Study on System Utterance of Suggestion to Promote User's Accepatance in Driving Environment

MIYAZAWA Kouki;KAGETANI Takuya;SHEN Raymond;KIKUCHI Hideaki;OGAWA Yoshito;HATA Chihiro;OHTA Katsumi;HOZUMI Hideaki;MITAMURA Takeshi

Transactions of the Japanese Society for Artificial Intelligence 25(6) p.723 - 7322010-2010

CiNii

Detail

ISSN:1346-0714

Outline:In this study, we aim at clarification of the factor that promotes an user's acceptance of suggestion from an interactive agent in driving environment. Our aim is to figure out how human beings accept the encouragement from interaction objects, and also which kinds of dialogues or action controls are necessary for the design of car navigation system which makes suggestion and requests to drivers. Firstly, we had an experiment for collecting dialogue between humans in driving simulation environment, then we analyzed the drivers' acceptance and evaluation for the navigators. As the results, we found that the presence and reliability of the navigator highly relate to the acceptance of suggestion from the navigator. When navigators were next to drivers, the rate of drivers' suggestion acceptance rose. However, the stress of drivers increased. In addition, based on the linguistic and acoustic analysis of the navigators' utterances, we found out some points of designing system utterance of suggestion to promote user's acceptance. We found that expressing the grounds of suggestions, showing the exact numbers, and the wide pitch ranges, all highly relate to the acceptance of suggestions.

Self Organizing Maps as the Perceptual Acquisition Model:-Unsupervised Phoneme Learning from Continuous Speech-

MIYAZAWA Kouki;SHIROSE Ayako;MAZUKA Reiko;KIKUCHI Hideaki

J. SOFT 26(1) p.510 - 5202014-2014

CiNii

Detail

ISSN:1347-7986

Outline:We assume that SOM is adequate as a language acquisition model of the native phonetic system. However, many studies don't consider the quantitative features (the appearance frequency and the number of frames of each phoneme) of the input data. Our model is designed to learn values of the acoustic characteristic of a natural continuous speech and to estimate the number and boundaries of the vowel categories without using explicit instructions. In the simulation trial, we investigate the relationship between the quantity of learning and the accuracy for the vowels in a single Japanese speaker's natural speech. As a result, it is found that the recognition accuracy rate (of our model) are 5% (/u/)-92% (/s/).

Automatic Estimation of Speaking Style in Speech Corpora Focusing on Speech Transcriptions

Shen Raymond;Kikuchi Hideaki

Journal of Natural Language Processing 21(3) p.445 - 4642014-2014

CiNii

Detail

ISSN:1340-7619

Outline:Recent developments in computer technology have allowed the construction and widespread application of large-scale speech corpora. To enable users of speech corpora to easier data retrieval, we attempt to characterise the speaking style of speakers recorded in the corpora. We first introduce the three scales for measuring speaking style which were proposed by Eskenazi in 1993. We then use morphological features extracted from speech transcriptions that have proven effective in discriminating between styles and identifying authors in the field of natural language processing to construct an estimation model of speaking style. More specifically, we randomly choose transcriptions from various speech corpora as text stimuli with which to conduct a rating experiment on speaking style perception. Then, using the features extracted from these stimuli and rating results, we construct an estimation model of speaking style, using a multi-regression analysis. After cross-validation (leave-1-out), the results show that among the three scales of speaking style, the ratings of two scales can be estimated with high accuracy, which proves the effectiveness of our method in the estimation of speaking style.

Turn-taking timing of mother tongue

Ichikawa Akira;Oohashi Hiroki;Naka Makiko;Kikuchi Hideaki;Horiuchi Yasuo;Kuroiwa Shingo

Studies in Science and Technology 5(1) p.113 - 1222016-2016

CiNii

Detail

ISSN:2186-4942

Constructing the corpus of infant-directed speech and infant-like robot-directed speech

Nakamura, Ryuji; Miyazawa, Kouki; Ishihara, Hisashi; Nishikawa, Ken'ya; Kikuchi, Hideaki; Asada, Minoru; Mazuka, Reiko

HAI 2015 - Proceedings of the 3rd International Conference on Human-Agent Interaction p.167 - 1692015/10-2015/10

DOIScopus

Detail

Outline:© 2015 ACM.The characteristics of the spoken language used to address infants have been eagerly studied as a part of the language acquisition research. Because of the uncontrollability factor with regard to the infants, the features and roles of infantdirected speech were tried to be revealed by the comparison of speech directed toward infants and that toward other listeners. However, they share few characteristics with infants, while infants have many characteristics which may derive the features of IDS. In this study, to solve this problem, we will introduce a new approach that replaces the infant with an infant-like robot which is designed to control its motions and to imitate its appearance very similar to a real infant. We have now recorded both infant-and infantlike robot-directed speech and are constructing both corpora. Analysis of these corpora is expected to contribute to the studies of infant-directed speech. In this paper, we discuss the contents of this approach and the outline of the corpora.

Humor utterance generation for non-task-oriented dialogue systems

Fujikura, Shohei; Ogawa, Yoshito; Kikuchi, Hideaki

HAI 2015 - Proceedings of the 3rd International Conference on Human-Agent Interaction p.171 - 1732015/10-2015/10

DOIScopus

Detail

Outline:© 2015 ACM.We propose a humor utterance generation method that is compatible with dialogue systems, to increase "desire of continuing dialogue". A dialogue system retrieves leading-item: noun pairs from Twitter as knowledge and attempts to select the most humorous reply using word similarity, which reveals that incongruity can be explained by the incongruity-resolution model. We consider the differences among individuals, and confirm the validity of the proposed method. Ex-perimental results indicate that high-incongruity replies are significantly effective against low-incongruity replies with a limited condition.

Assigning a Personality to a Spoken Dialogue Agent by Behavior Reporting

Ogawa, Yoshito; Kikuchi, Hideaki

New Generation Computing 35(2) p.181 - 2092017/04-2017/04

DOIScopus

Detail

ISSN:02883635

Outline:© 2017, Ohmsha, Ltd. and Springer Japan. A method to assign a personality to a spoken dialogue agent is proposed and evaluated. The proposed method assigns a personality using agent reporting about behavior independent of interaction with a user. The proposed method attempts to assigning complex personalities. For this purpose, we have defined a behavior report dialogue and designed a personality assigning method using behavior reporting. The proposed method consists of three steps: collecting stereotypes between a personality and behavior through a questionnaire, designing the behavior report dialogue from the collected stereotypes, and agent reports about behavior at the start of interactions with a user. Experimental results show that the proposed method can assign a personality by repeating the behavior report dialogue, (the assigned personality is equivalent to the personality determined by the collected stereotypes) and that reporting behavior influences the assigned personality. In addition, we verified that the proposed method can assign “kind”, “judicious” and the five basic personalities defined in the Tokyo University Egogram Second Edition.

Vowels in infant-directed speech: More breathy and more variable, but not clearer

Miyazawa, Kouki; Miyazawa, Kouki; Shinya, Takahito; Martin, Andrew; Martin, Andrew; Kikuchi, Hideaki; Mazuka, Reiko; Mazuka, Reiko

Cognition 166p.84 - 932017/09-2017/09

DOIScopus

Detail

ISSN:00100277

Outline:© 2017 Elsevier B.V. Infant-directed speech (IDS) is known to differ from adult-directed speech (ADS) in a number of ways, and it has often been argued that some of these IDS properties facilitate infants’ acquisition of language. An influential study in support of this view is Kuhl et al. (1997), which found that vowels in IDS are produced with expanded first and second formants (F1/F2) on average, indicating that the vowels are acoustically further apart in IDS than in ADS. These results have been interpreted to mean that the way vowels are produced in IDS makes infants’ task of learning vowel categories easier. The present paper revisits this interpretation by means of a thorough analysis of IDS vowels using a large-scale corpus of Japanese natural utterances. We will show that the expansion of F1/F2 values does occur in spontaneous IDS even when the vowels’ prosodic position, lexical pitch accent, and lexical bias are accounted for. When IDS vowels are compared to carefully read speech (CS) by the same mothers, however, larger variability among IDS vowel tokens means that the acoustic distances among vowels are farther apart only in CS, but not in IDS when compared to ADS. Finally, we will show that IDS vowels are significantly more breathy than ADS or CS vowels. Taken together, our results demonstrate that even though expansion of formant values occurs in spontaneous IDS, this expansion cannot be interpreted as an indication that the acoustic distances among vowels are farther apart, as is the case in CS. Instead, we found that IDS vowels are characterized by breathy voice, which has been associated with the communication of emotional affect.

Books And Publication

"Spoken Language Systems", S. Nakagawa et al. (Eds.)

HATAOKA Nobuo, ANDO Haru, and KIKUCHI Hideaki

Ohmsha/IOS Press2005-

Detali

ISBN:4-274-90637-x

"Voicing in Japanese," Van de Weijer, Nanjo, Nishihara (Eds.)

MAEKAWA Kikuo and KIKUCHI Hideaki

Mouton de Gruyter, Berlin and New York2006-

"韻律を利用した対話状態の推定", 広瀬啓吉 編「韻律と音声言語情報処理」

KIKUCHI Hideaki and SHIRAI Katsuhiko

丸善2006-

Detali

ISBN:4-621-07674-4

"音声コミュニケーションの分析単位 -ToBI-", 坊農真弓, 高梨克也編,「多人数インタラクションの分析手法」

KIKUCHI Hideaki

オーム社2009-

Detali

ISBN:4274207327

“Outline of J_ToBI”, in "COMPUTER PROCESSING OF ASIAN SPOKEN LANGUAGES"

KIKUCHI Hideaki

Americas Group Publications,U.S.2010-

Detali

ISBN:0935047727

"音声対話システム", 白井克彦編著「音声言語処理の潮流」

KIKUCHI Hideaki

コロナ社2010-

Detali

ISBN:4339008104

"感情の音声表出", 石井克典監修「IoHを指向する感情・思考センシング技術」

KIKUCHI Hideaki

CMC出版2019-

Detali

ISBN:978-4-7813-1430-3

Patent

Reference Number:19

連続音声認識装置および方法(日本)

白井 克彦, 城崎 康夫, 菊池 英明

H11-281437、2001-109491

Reference Number:481

音声認識装置及び音声認識用プログラム(日本)

白井 克彦, 菊池 英明, 大久保 崇

2005- 80732、2006-259641

Reference Number:756

検索語クラスタリング装置、検索語クラスタリング方法、検索語クラスタリングプログラム及び記録媒体(日本)

白井 克彦, 菊池 英明, 新関 一馬

2007-193700、2009- 31931

Research Grants & Projects

Grant-in-aids for Scientific Research Adoption Situation

Research Classification:

Fundamental Study for Conversion between Spoken and Written Japanese Considering Influence of Interactivity

2013/-0-2016/-0

Allocation Class:¥4680000

Research Classification:

On the physical factors which makes the mother tongue dialogues smoothly - through the comparison with the non-mother tongue

2012/-0-2015/-0

Allocation Class:¥3900000

Research Classification:

Analysis of infant-directed speech by acoustic and computational modeling methods.

Allocation Class:¥4810000

Research Classification:

A Study on a framework of spontaneous communication depending on dialogue situation

Allocation Class:¥15360000

Research Classification:

Quantitative Analysis of Linguistic Variation Using a Spontaneous Speech Corpus

Allocation Class:¥14800000

Research Classification:

Analysis of effective communication process by using body and artifact for faculty Development

Allocation Class:¥3700000

Research Classification:

Development of new prototypes and models of higher education utilizing broadband networks.

Allocation Class:¥8300000

Research Classification:

XML documentation of complex annotation on spontaneous speech data

Allocation Class:¥3900000

Research Classification:

Elaboration of articulatory phonetics by means of realtime-MRI and WAVE data

2017/-0-2020/-0

Allocation Class:¥17810000

Research Classification:

Estimation of User's Impression Space for Improving Desire of Interaction with Spoken Dialogue System

2014/-0-2017/-0

Allocation Class:¥3900000

Research Classification:

Preparation of Database and Analysis of Dialectial Difference in Phonetics, Phonology, Grammar and Vocabulary among Dialects of Ainu

2012/-0-2015/-0

Allocation Class:¥4940000

Research Classification:

Development of protocol of effective social work interview for older adults with dementia

2019/-0-2022/-0

Allocation Class:¥6890000

Research Classification:

Real-time MRI database of articulatory movements of Japanese

2020/-0-2024/-0

Allocation Class:¥17420000

On-campus Research System

Special Research Project

音声対話システム発話の音声言語的特徴制御によるインタラクション欲求向上

2013

Research Results Outline:音声対話システムのシステム発話を制御することによってユーザに与える印象を変化させる技術の開発を目指している。本特定課題研究では、擬人化したシステムの自音声対話システムのシステム発話を制御することによってユーザに与える印象を変化させる技術の開発を目指している。本特定課題研究では、擬人化したシステムの自己開示によってパーソナリティを付与する手法に関する基礎研究を中心に進めた。実験を通じて、自己開示量...音声対話システムのシステム発話を制御することによってユーザに与える印象を変化させる技術の開発を目指している。本特定課題研究では、擬人化したシステムの自己開示によってパーソナリティを付与する手法に関する基礎研究を中心に進めた。実験を通じて、自己開示量と内容によって特定のパーソナリティを付与できることを確認した。この成果は、ヒューマンインタフェース学会論文誌に査読論文として掲載された(「自己開示による音声対話エージェントへのパーソナリティ付与」)。他にも、マイクロブログからユーモア発話を自動生成する技術(「非タスク指向対話システムにおけるマイクロブログを用いたユーモア発話の自動生成」)、発話速度あるいは無音区間長を制御する手法(「ロボット発話の話速・無音区間長の制御によるパーソナリティ認知と対話継続欲求の向上」)を検討し、それぞれによってユーザがシステムに抱く印象がどのように変化するかを実験により調査した。いずれも国内学会にて成果を発表した。ユーザ発話における音声のプロソディを解析することによってユーザの心的状態を推定して、それに応じてシステムの振舞を変えることによってユーザが抱く愛着感を変化させられることを実験により確認し、この成果を国際会議にて発表した(「Effects of an Agent Feature Comprehension on the Emotional Attachment of Users」)。いずれの研究においても、ユーザが抱く印象を変化させるこれらの手法によって、音声対話システムに対するインタラクション継続欲求が向上することを確認している。このことを一旦整理して国内学会にて発表した(「音声対話システムに対するインタラクション継続欲求」)。こうした一連の研究成果を体系化してさらに幅広く応用可能な技術を開発するために、インタラクション継続欲求とユーザ印象空間の関係を明確にする必要性が生じている。そこで最終的に、本特定課題研究を経て、2014年度科学研究費の基盤(C)に「音声対話システムに対するインタラクション欲求向上のためのユーザ印象空間の推定」というテーマを申請して採択された。

生体情報を教師信号としたモデル学習による感情推定技術の高度化

2006

Research Results Outline: 本研究では、音声からの心的状態の推定において話者の心的状況をより高い精度で推定するため、生体情報を教師信号としたモデル学習を行う“生理心理学的アプロ 本研究では、音声からの心的状態の推定において話者の心的状況をより高い精度で推定するため、生体情報を教師信号としたモデル学習を行う“生理心理学的アプローチの導入”を提案する。 従来の感情推定は、モデル学習の際に実験者の判断による評定結果が教師信号と... 本研究では、音声からの心的状態の推定において話者の心的状況をより高い精度で推定するため、生体情報を教師信号としたモデル学習を行う“生理心理学的アプローチの導入”を提案する。 従来の感情推定は、モデル学習の際に実験者の判断による評定結果が教師信号として用いられるため主観的方法であることが否めない。また推定の対象も基本的な感情にのみ重点が置かれてきた。 生体情報は、意図的な操作が入らず継時的な変化を捉えられることができるとされている。そのため、推定を行う際に実験の第一段階として生体信号を利用することで、多様で連続的な心情の変化を対象とすることができるようになり、またより客観的で精度の高い判断が可能になると思われる。 難度の異なる音読課題を2つ用意し、課題間における生体信号の反応の違いが音声の違いにも現れるのかを観察した。実験者の主観的評価によってストレス状態と判断された被験者の音声と、それらのうち生体信号の変化からもストレス状態にあると判断できた被験者の音声の比較を行う。 生体信号には、心的状態の推定へ利用できると思われた容積脈派(BVP)、心電図(EKG)、皮膚温(TEMP)、皮膚コンダクタンス(SC)を用いた。 音声の比較には、各音声からF0とパワーそれぞれの最大値、最小値、振幅、平均値、それに発話速度を加えた9つの特徴量を抽出し、これらを決定木学習に利用した。決定木学習には、C4.5アルゴリズムを使用し、交差検定を用いて評価を行う。 全データ(実験者の主観的評価のみによってストレス状態を判断した)で学習モデルを生成した場合平均63.9%であった判別率が、選別データ(主観的評価に加え、生体信号の変化からもストレス状態を判断した)で学習モデルを生成した場合には平均77.8%まで精度が向上した。 生体信号がストレス状態を判断するうえで一つの指標となり得ることを示唆する結果となった。本実験の結果、音声からの心的状態の推定を行う際に生体情報を利用することの有益性が実証された。

Foreign Countries Research Activity

Research Project Title: 音声言語における感情・評価・態度の解析技術高度化

2009/03-2010/03

Affiliation: オハイオ州立大学(アメリカ)、北京大学(中国)

Lecture Course

Course TitleSchoolYearTerm
Research Method for Measurement and Modeling of Human 01School of Human Sciences2020fall quarter
Research Method for Measurement and Modeling of Human 02School of Human Sciences2020fall quarter
Research Method for Measurement and Modeling of Human 03School of Human Sciences2020fall quarter
Seminar II (Language and Information Science)School of Human Sciences2020fall semester
Language and Information SciencesSchool of Human Sciences (Online Degree Program)2020fall semester
Language and Information Sciences AGraduate School of Human Sciences2020spring semester
Language and Information Sciences BGraduate School of Human Sciences2020fall semester
Language and Information Sciences(1) AGraduate School of Human Sciences2020spring semester
Language and Information Sciences(1) BGraduate School of Human Sciences2020fall semester
Language and Information Sciences(1) CGraduate School of Human Sciences2020spring semester
Language and Information Sciences(1) DGraduate School of Human Sciences2020fall semester
Language and Information Sciences(2) AGraduate School of Human Sciences2020spring semester
Language and Information Sciences(2) BGraduate School of Human Sciences2020fall semester
Language and Information Sciences(2) CGraduate School of Human Sciences2020spring semester
Language and Information Sciences(2) DGraduate School of Human Sciences2020fall semester
Language and Information SciencesGraduate School of Human Sciences2020fall quarter
Language and Information Sciences(D) AGraduate School of Human Sciences2020spring semester
Language and Information Sciences(D) BGraduate School of Human Sciences2020fall semester