sequence to sequence, B3TB2006, i

Similar documents
Haiku Generation Based on Motif Images Using Deep Learning Koki Yoneda 1 Soichiro Yokoyama 2 Tomohisa Yamashita 2 Hidenori Kawamura Scho

一般社団法人電子情報通信学会 THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGIN

_314I01BM浅谷2.indd

Computational Semantics 1 category specificity Warrington (1975); Warrington & Shallice (1979, 1984) 2 basic level superiority 3 super-ordinate catego

A Japanese Word Dependency Corpus ÆüËܸì¤Îñ¸ì·¸¤ê¼õ¤±¥³¡¼¥Ñ¥¹

自然言語処理24_705

_AAMT/Japio特許翻訳研究会.key

No. 3 Oct The person to the left of the stool carried the traffic-cone towards the trash-can. α α β α α β α α β α Track2 Track3 Track1 Track0 1

Machine Learning for NLP

Outline ACL 2017 ACL ACL 2017 Chairs/Presidents

¥ì¥·¥Ô¤Î¸À¸ì½èÍý¤Î¸½¾õ

Convolutional Neural Network A Graduation Thesis of College of Engineering, Chubu University Investigation of feature extraction by Convolution

( )

main.dvi

gengo.dvi

SICE東北支部研究集会資料(2017年)

( : A9TB2096)

21 Pitman-Yor Pitman- Yor [7] n -gram W w n-gram G Pitman-Yor P Y (d, θ, G 0 ) (1) G P Y (d, θ, G 0 ) (1) Pitman-Yor d, θ, G 0 d 0 d 1 θ Pitman-Yor G

johnny-paper2nd.dvi

IPSJ-TOD

WHITE PAPER RNN

& Vol.5 No (Oct. 2015) TV 1,2,a) , Augmented TV TV AR Augmented Reality 3DCG TV Estimation of TV Screen Position and Ro

ニュラールネットに基づく機械翻訳 ニューラルネットに 基づく機械翻訳 Graham Neubig 奈良先端科学技術大学院大学 (NAIST)

IPSJ SIG Technical Report Vol.2009-DPS-141 No.20 Vol.2009-GN-73 No.20 Vol.2009-EIP-46 No /11/27 1. MIERUKEN 1 2 MIERUKEN MIERUKEN MIERUKEN: Spe

[1], B0TB2053, i

2

IPSJ SIG Technical Report Vol.2014-CG-155 No /6/28 1,a) 1,2,3 1 3,4 CG An Interpolation Method of Different Flow Fields using Polar Inter

IPSJ SIG Technical Report Vol.2011-MUS-91 No /7/ , 3 1 Design and Implementation on a System for Learning Songs by Presenting Musical St

: ( 1) () 1. ( 1) 2. ( 1) 3. ( 2)

(a) (8.87) (b) (5.38) (c) (.3) Annotated English Gigaword Corpus Ru

1 4 4 [3] SNS 5 SNS , ,000 [2] c 2013 Information Processing Society of Japan


独立行政法人情報通信研究機構 Development of the Information Analysis System WISDOM KIDAWARA Yutaka NICT Knowledge Clustered Group researched and developed the infor

<> <name> </name> <body> <></> <> <title> </title> <item> </item> <item> 11 </item> </>... </body> </> 1 XML Web XML HTML 1 name item 2 item item HTML

1 Kinect for Windows M = [X Y Z] T M = [X Y Z ] T f (u,v) w 3.2 [11] [7] u = f X +u Z 0 δ u (X,Y,Z ) (5) v = f Y Z +v 0 δ v (X,Y,Z ) (6) w = Z +

Vol.55 No (Jan. 2014) saccess 6 saccess 7 saccess 2. [3] p.33 * B (A) (B) (C) (D) (E) (F) *1 [3], [4] Web PDF a m

MDD PBL ET 9) 2) ET ET 2.2 2), 1 2 5) MDD PBL PBL MDD MDD MDD 10) MDD Executable UML 11) Executable UML MDD Executable UML

Core1 FabScalar VerilogHDL Cache Cache FabScalar 1 CoreConnect[2] Wishbone[3] AMBA[4] AMBA 1 AMBA ARM L2 AMBA2.0 AMBA2.0 FabScalar AHB APB AHB AMBA2.0

[2][3][4][5] 4 ( 1 ) ( 2 ) ( 3 ) ( 4 ) 2. Shiratori [2] Shiratori [3] [4] GP [5] [6] [7] [8][9] Kinect Choi [10] 3. 1 c 2016 Information Processing So

,,, Twitter,,, ( ), 2. [1],,, ( ),,.,, Sungho Jeon [2], Twitter 4 URL, SVM,, , , URL F., SVM,, 4 SVM, F,.,,,,, [3], 1 [2] Step Entered

Input image Initialize variables Loop for period of oscillation Update height map Make shade image Change property of image Output image Change time L

IPSJ SIG Technical Report Vol.2010-GN-74 No /1/ , 3 Disaster Training Supporting System Based on Electronic Triage HIROAKI KOJIMA, 1 KU

untitled

[4], [5] [6] [7] [7], [8] [9] 70 [3] 85 40% [10] Snowdon 50 [5] Kemper [3] 2.2 [11], [12], [13] [14] [15] [16]

ipod touch 1 2 Apple ipod touch ipod touch 3 ( ) ipod touch ( 1 ) Apple ( 2 ) Web 1),2) 3. ipod touch 1 2 ipod touch x y z i

A Study on Throw Simulation for Baseball Pitching Machine with Rollers and Its Optimization Shinobu SAKAI*5, Yuichiro KITAGAWA, Ryo KANAI and Juhachi

2014/1 Vol. J97 D No. 1 2 [2] [3] 1 (a) paper (a) (b) (c) 1 Fig. 1 Issues in coordinating translation services. (b) feast feast feast (c) Kran

2 ( ) i

IPSJ SIG Technical Report Pitman-Yor 1 1 Pitman-Yor n-gram A proposal of the melody generation method using hierarchical pitman-yor language model Aki

3807 (3)(2) ,267 1 Fig. 1 Advertisement to the author of a blog. 3 (1) (2) (3) (2) (1) TV 2-0 Adsense (2) Web ) 6) 3

IPSJ SIG Technical Report Vol.2012-CG-148 No /8/29 3DCG 1,a) On rigid body animation taking into account the 3D computer graphics came

1 1 tf-idf tf-idf i

Vol.54 No (July 2013) [9] [10] [11] [12], [13] 1 Fig. 1 Flowchart of the proposed system. c 2013 Information

自然言語処理21_249

DPA,, ShareLog 3) 4) 2.2 Strino Strino STRain-based user Interface with tacticle of elastic Natural ObjectsStrino 1 Strino ) PC Log-Log (2007 6)

Fig. 2 28th Ryuou Tournament, Match 5, 59th move. The last move is Black s Rx5f. 1 Tic-Tac-Toe Fig. 1 AsearchtreeofTic-Tac-Toe. [2] [3], [4]

Abstract

DL4NL-tsuboi intro

..,,,, , ( ) 3.,., 3.,., 500, 233.,, 3,,.,, i

IPSJ SIG Technical Report Vol.2011-EC-19 No /3/ ,.,., Peg-Scope Viewer,,.,,,,. Utilization of Watching Logs for Support of Multi-

203 図 2,re re, [Nivre 08]. y, 1 y i. ŷ = arg max y Y * J j=1 P r(y j y j 1 1,x) 2,, Pr y j y 1 j 1, x.,. ŷ = arg max y Y * 図 1 J j=1 exp(w o φ(y j,y j

,,,,., C Java,,.,,.,., ,,.,, i

kut-paper-template2.dvi

IPSJ SIG Technical Report Vol.2014-HCI-158 No /5/22 1,a) 2 2 3,b) Development of visualization technique expressing rainfall changing conditions

(2008) JUMAN *1 (, 2000) google MeCab *2 KH coder TinyTextMiner KNP(, 2000) google cabocha(, 2001) JUMAN MeCab *1 *2 h

IPSJ SIG Technical Report Vol.2013-CVIM-187 No /5/30 1,a) 1,b), 1,,,,,,, (DNN),,,, 2 (CNN),, 1.,,,,,,,,,,,,,,,,,, [1], [6], [7], [12], [13]., [

24 Region-Based Image Retrieval using Fuzzy Clustering

DEIM Forum 2009 B4-6, Str

IPSJ SIG Technical Report Vol.2017-SLP-115 No /2/18 1,a) 1 1,2 Sakriani Sakti [1][2] [3][4] [5][6][7] [8] [9] 1 Nara Institute of Scie

% 95% 2002, 2004, Dunkel 1986, p.100 1

17 Proposal of an Algorithm of Image Extraction and Research on Improvement of a Man-machine Interface of Food Intake Measuring System

HASC2012corpus HASC Challenge 2010,2011 HASC2011corpus( 116, 4898), HASC2012corpus( 136, 7668) HASC2012corpus HASC2012corpus

( ) [1] [4] ( ) 2. [5] [6] Piano Tutor[7] [1], [2], [8], [9] Radiobaton[10] Two Finger Piano[11] Coloring-in Piano[12] ism[13] MIDI MIDI 1 Fig. 1 Syst

Study on Throw Accuracy for Baseball Pitching Machine with Roller (Study of Seam of Ball and Roller) Shinobu SAKAI*5, Juhachi ODA, Kengo KAWATA and Yu

Vol.53 No (Mar. 2012) 1, 1,a) 1, 2 1 1, , Musical Interaction System Based on Stage Metaphor Seiko Myojin 1, 1,a

( : A8TB2163)

2015 9

IPSJ SIG Technical Report Vol.2017-MUS-116 No /8/24 MachineDancing: 1,a) 1,b) 3 MachineDancing MachineDancing MachineDancing 1 MachineDan

Run-Based Trieから構成される 決定木の枝刈り法

i

1: NTT Yahoo! , , , , CIAIR CIAIR- ICSD CIAIR-ICSD 6 dlwz lg 3 30 dl wz 2 60 dl O D O:

28 Horizontal angle correction using straight line detection in an equirectangular image

1 1 CodeDrummer CodeMusician CodeDrummer Fig. 1 Overview of proposal system c

音響モデル triphone 入力音声 音声分析 デコーダ 言語モデル N-gram bigram HMM の状態確率として利用 出力層 triphone: 3003 ノード リスコア trigram 隠れ層 2048 ノード X7 層 1 Structure of recognition syst

PowerPoint プレゼンテーション

DEIM Forum 2014 B Twitter Twitter Twitter 2006 Twitter 201

Mimehand II[1] [2] 1 Suzuki [3] [3] [4] (1) (2) 1 [5] (3) 50 (4) 指文字, 3% (25 個 ) 漢字手話 + 指文字, 10% (80 個 ) 漢字手話, 43% (357 個 ) 地名 漢字手話 + 指文字, 21

DEIM Forum 2009 C8-4 QA NTT QA QA QA 2 QA Abstract Questions Recomme

IPSJ SIG Technical Report Vol.2009-CVIM-167 No /6/10 Real AdaBoost HOG 1 1 1, 2 1 Real AdaBoost HOG HOG Real AdaBoost HOG A Method for Reducing

レビューテキストの書き の評価視点に対する評価点の推定 29 3

[12] Qui [6][7] Google N-gram[11] Web ( 4travel 5, 6 ) ( 7 ) ( All About 8 ) (1) (2) (3) 3 3 (1) (2) (3) (a) ( (b) (c) (d) (e) (1

知能と情報, Vol.30, No.5, pp

3.1 Thalmic Lab Myo * Bluetooth PC Myo 8 RMS RMS t RMS(t) i (i = 1, 2,, 8) 8 SVM libsvm *2 ν-svm 1 Myo 2 8 RMS 3.2 Myo (Root

29 Short-time prediction of time series data for binary option trade

単語の分散表現と構成性の計算モデルの発展

Honda 3) Fujii 4) 5) Agrawala 6) Osaragi 7) Grabler 8) Web Web c 2010 Information Processing Society of Japan

DEIM Forum 2018 C ARIMA Long Short-Term Memory LSTM

3 2 2 (1) (2) (3) (4) 4 4 AdaBoost 2. [11] Onishi&Yoda [8] Iwashita&Stoica [5] 4 [3] 3. 3 (1) (2) (3)

14 2 5

2017 (413812)

Transcription:

B3TB2006 2017 3 31

sequence to sequence, B3TB2006, 2017 3 31. i

A Study on a Style Control for Dialogue Response Generation Reina Akama Abstract We propose a new dialogue response generation model combining sequence to sequence and metastasis learning for controling a STYLE, which impresses a character of a specific speaker. To generate consistent style responses, most of priviouse works for collected text pairs by handmade rules or human annotation. However, thier costs were too expensive to learn response generation models with a consistent style. In our methods, after learning pre-training model with a largescale copora without restriction styles, we learn a response generation model with a small-scale copora with a style restriction and the pre-training model. Compared with responses between our proposed model and baseline models without metastasis learning, our model generated consistent style and appropriate responses to input texts. Keywords: response generation, dialog, style control, newral network, transfer learning Graduation Thesis, Department of Information and Intelligent Systems, Tohoku University, B3TB2006, March 31, 2017. ii

Contents 1 1 2 3 2.1 Sequence to sequence..................... 3 2.2.................... 4 2.3................................ 5 3 6 3.1............... 6 3.1.1 Transfer........................ 6 3.2........... 7 3.2.1 Transfer+freq..................... 7 3.2.2 Transfer+sim..................... 7 4 9 4.1........................... 9 4.2........................... 9 5 11 5.1.............................. 11 5.1.1........................ 11 5.1.2..................... 11 5.1.3............... 13 5.2.................................. 13 5.2.1........................ 13 5.2.2............. 14 6 16 17 iii

List of Figures 1 seq2seq.................... 3 2.............................. 6 3 5..................... 12 iv

List of Tables 1........... 1 2....................... 9 3.................... 10 4................... 14 5.................. 14 6................. 15 7............... 15 v

1 Twitter (SNS) 1 Recurrent Neural Network (RNN) sequence to sequence (seq2seq) [18] SNS seq2seq 1 2) 4) 6) 1 [19, 20] [10] Table 1: 1) User: 2) System: 3) User: 4) System: 5) User: 6) System: 1

seq2seq SNS 2

2 2.1 Sequence to sequence 1 seq2seq [4, 18] Sutskever seq2seq 1 / / / / / / Figure 1: seq2seq seq2seq RNN RNN (x 1,..., x T ) t (1) h t (2) y t T (y 1,..., y T ) W hx W hh W yh h t = sigm(w hx x t + W hh h t 1 ) (1) y t = W yh h t (2) RNN T RNN v () RNN T ( ) Seq2seq RNN Long Short-Term Memory (LSTM) [7] Sutskever seq2seq 2 LSTM 1 LSTM (3) 3

(x 1,..., x T ) (y 1,..., y T ) p(y 1,..., y T x 1,..., x T ) p(y 1,..., y T x 1,..., x T ) = T t=1 p(y t v, y 1,..., y t 1 ) (3) v LSTM 1 y 1 LSTM v seq2seq [5, 18] [14] [15] seq2seq seq2seq 100 2.2 Walker [19, 11, 12, 13] Walker [20] 4

seq2seq [10] Twitter 2.3 1 [1] [2] [3] [9, 6] [16] GloVe[17]. 5

3 3.1?? Sutskever seq2seq [18] seq2seq RNN 3.1.1 Transfer ( ) N s Transfer seq2seq seq2seq 事前学習コーパス コーパス内の単語 学習で扱う語彙 スタイルコーパス コーパス内の単語 Figure 2: 6

3.2 Seq2seq ( unk ) 2 2 Transfer+freq Transfer+sim 2 3.2.1 Transfer+freq Transfer+freq N s N p N s 2 N p = 25, 000, N s = 1, 000 25,000 24,000 1,000 3.2.2 Transfer+sim Transfer+sim Glove [17] 2 7

8

4 Twitter TV 2 2 95% 5% Table 2: - - - 3,688,162 591,880 12,564 12,102 1,476 2,137 4.1 Twitter Twitter 2015 1 12 - - URL 370 4.2 TV 2015 9 2016 5 2 2-3 - 9

0.3% 0.04% Table 3: (a) (b)??! 10

5 5.1 5.1.1 3 Transfer 3(c) Transfer+freq 3(d) Transfer+sim 3(e) 25,000 5 Transfer Transfer+freq 1,000 500 Transfer+sim Twitter GloVe 128 0.6 10 5.1.2 seq2seq Base 3(a), Mixed 3(b) 2 Base 25,000 Base 3 Mixed Mixed 24,000 1,000 24,500 500 5 11

seq2seq seq2seq seq2seq seq2seq 事前学習コーパス コーパス内の単語 学習で扱う語彙 スタイルコーパス コーパス内の単語 事前学習コーパス + スタイルコーパス コーパス内の単語 学習で扱う単語 スタイルコーパス コーパス内の単語 (a) Base (b) Mixed seq2seq seq2seq seq2seq seq2seq 事前学習コーパス コーパス内の単語 学習で扱う語彙 スタイルコーパス コーパス内の単語 事前学習コーパス コーパス内の単語 学習で扱う語彙 スタイルコーパス コーパス内の単語 (c) Transfer (d) Transfer+freq seq2seq seq2seq 事前学習コーパス コーパス内の単語 学習で扱う語彙 スタイルコーパス コーパス内の単語 (e) Transfer+sim Figure 3: 5 12

5.1.3 64 Seq2seq 1024 LSTM 2 2048 dropout rate 0.2 Adam [8] Adam 5.2 5.2.1 1. 2. 2 1) / 2 2) / 2. Twitter 50 10 4 5 4 3 Base 80% 3 80% Transfer+freq Mixed Mixed 5 Transfer Transfer+freq 90% 13

Table 4: 1. 2. Base 39 (78%) 18 (36%) Mixed 39 (78%) 23 (46%) Transfer 41 (82%) 39 (78%) Transfer+freq 38 (76%) 39 (78%) Transfer+sim 39 (78%) 38 (76%) * 4 Table 5: 1. 2. Base 43 (86%) 0 (0%) Mixed 40 (80%) 16 (32%) Transfer 29 (58%) 45 (90%) Transfer+freq 31 (62%) 47 (94%) Transfer+sim 32 (64%) 44 (88%) * 2 3 60% Base Transfer+freq 7 2) 5.2.2 2 Transfer+freq 6 7 1 14

3 1 1 Base 3 1,400 - Table 6: 1) User: 2) System: 3) User: 4) System: 5) User: 6) System: Table 7: 1) User: 2) System: 3) User: 4) System: 5) User: 6) System: 15

6 seq2seq TV 16

Preferred Networks TV 17

References [1] Nips 2005 workshop inductive transfer: 10 years later. 2005. http://iitrl.acadiau.ca/itws05/. [2] Andrew Arnold, Ramesh Nallapati, and William W Cohen. A comparative study of methods for transductive transfer learning. In Data Mining Workshops, 2007. ICDM Workshops 2007. Seventh IEEE International Conference on, pp. 77 82. IEEE, 2007. [3] John Blitzer, Mark Dredze, Fernando Pereira, et al. Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In ACL, Vol. 7, pp. 440 447, 2007. [4] Kyunghyun Cho, Bart van Merrienboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. Learning phrase representations using rnn encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Association for Computational Linguistics, 2014. http://www.aclweb.org/anthology/d14-1179. [5] Bahdanau Dzmitry, Cho Kyunghyun, and Bengio Yoshua. Neural machine translation by jointly learning to align and translate. In The International Conference on Learning Representations (ICLR), 2015. [6] Almut Silja Hildebrand, Matthias Eck, Stephan Vogel, and Alex Waibel. Adaptation of the translation model for statistical machine translation based on information retrieval. In Proceedings of EAMT, Vol. 2005, pp. 133 142, 2005. [7] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, pp. 1735 1780, 1997. [8] Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In The International Conference on Learning Representations (ICLR), 2015. 18

[9] Philipp Koehn and Josh Schroeder. Experiments in domain adaptation for statistical machine translation. In Proceedings of the second workshop on statistical machine translation, pp. 224 227. Association for Computational Linguistics, 2007. [10] Jiwei Li, Michel Galley, Chris Brockett, Georgios Spithourakis, Jianfeng Gao, and Bill Dolan. A persona-based neural conversation model. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, pp. 994 1003, 2016. [11] Grace I Lin and Marilyn A Walker. All the world s a stage: Learning character models from film. In AIIDE, 2011. [12] François Mairesse and Marilyn A Walker. Towards personality-based user adaptation: psychologically informed stylistic language generation. User Modeling and User-Adapted Interaction, pp. 227 278, 2010. [13] François Mairesse and Marilyn A Walker. Controlling user perceptions of linguistic style: Trainable generation of personality traits. Computational Linguistics, pp. 455 488, 2011. [14] Ramesh Nallapati, Bowen Zhou, Cicero dos Santos, Caglar Gulcehre, and Bing Xiang. Abstractive text summarization using sequence-to-sequence rnns and beyond. In Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, pp. 280 290. Association for Computational Linguistics, 2016. [15] Vinyals Oriol and Le Quoc. A neural conversational model. In International Conference on Machine Learning(ICML) Deep Learning Workshop 2015, 2015. [16] Sinno Jialin Pan and Qiang Yang. A survey on transfer learning. IEEE Transactions on knowledge and data engineering, Vol. 22, No. 10, pp. 1345 1359, 2010. [17] Jeffrey Pennington, Richard Socher, and Christopher D. Manning. Glove: Global vectors for word representation. In Empirical Meth- 19

ods in Natural Language Processing (EMNLP), pp. 1532 1543, 2014. http://www.aclweb.org/anthology/d14-1162. [18] Ilya Sutskever, Oriol Vinyals, and Quoc V Le. Sequence to sequence learning with neural networks. In Advances in neural information processing systems, pp. 3104 3112, 2014. [19] Marilyn A Walker, Grace I Lin, and Jennifer Sawyer. An annotated corpus of film dialogue for learning and characterizing character style. In LREC, pp. 1373 1378, 2012. [20],,,,,.., Vol. 31, No. 1, pp. DSF E 1, 2016. 20