: ) B 2.7) A B 2.7) 3) 4) 5) substring) subsequence) A LCStr, LCS s = s 1,..., s m, t = t 1,..., t m character) mo

Similar documents
A Japanese Word Dependency Corpus ÆüËܸì¤Îñ¸ì·¸¤ê¼õ¤±¥³¡¼¥Ñ¥¹

Modal Phrase MP because but 2 IP Inflection Phrase IP as long as if IP 3 VP Verb Phrase VP while before [ MP MP [ IP IP [ VP VP ]]] [ MP [ IP [ VP ]]]

IPSJ-TOD

¥ì¥·¥Ô¤Î¸À¸ì½èÍý¤Î¸½¾õ

自然言語処理25_527

Computational Semantics 1 category specificity Warrington (1975); Warrington & Shallice (1979, 1984) 2 basic level superiority 3 super-ordinate catego

@08470030ヨコ/篠塚・窪田 221号

[4], [5] [6] [7] [7], [8] [9] 70 [3] 85 40% [10] Snowdon 50 [5] Kemper [3] 2.2 [11], [12], [13] [14] [15] [16]

21 Pitman-Yor Pitman- Yor [7] n -gram W w n-gram G Pitman-Yor P Y (d, θ, G 0 ) (1) G P Y (d, θ, G 0 ) (1) Pitman-Yor d, θ, G 0 d 0 d 1 θ Pitman-Yor G

2

BLEU Kishore Papineni, Salim Roukos, Todd Ward and Wei-Jing Zhu. (2002) BLEU: a method for Automatic Evaluation of Machine Translation. ACL. MT ( ) MT

( )

IPSJ SIG Technical Report Vol.2010-NL-199 No /11/ treebank ( ) KWIC /MeCab / Morphological and Dependency Structure Annotated Corp

自然言語処理24_705

I

I A A441 : April 15, 2013 Version : 1.1 I Kawahira, Tomoki TA (Shigehiro, Yoshida )

数学の基礎訓練I

医系の統計入門第 2 版 サンプルページ この本の定価 判型などは, 以下の URL からご覧いただけます. このサンプルページの内容は, 第 2 版 1 刷発行時のものです.

untitled


II (Percolation) ( 3-4 ) 1. [ ],,,,,,,. 2. [ ],.. 3. [ ],. 4. [ ] [ ] G. Grimmett Percolation Springer-Verlag New-York [ ] 3

(MIRU2008) HOG Histograms of Oriented Gradients (HOG)

main.dvi

Part () () Γ Part ,

July 28, H H 0 H int = H H 0 H int = H int (x)d 3 x Schrödinger Picture Ψ(t) S =e iht Ψ H O S Heisenberg Picture Ψ H O H (t) =e iht O S e i

1 1 tf-idf tf-idf i

a) Extraction of Similarities and Differences in Human Behavior Using Singular Value Decomposition Kenichi MISHIMA, Sayaka KANATA, Hiroaki NAKANISHI a


gengo.dvi

IPSJ SIG Technical Report Vol.2016-CE-137 No /12/ e β /α α β β / α A judgment method of difficulty of task for a learner using simple

2 2.1 NPCMJ ( (Santorini, 2010) (NPCMJ, 2016) (1) (, 2016) (1) (2) (1) ( (IP-MAT (CONJ ) (PP (NP (D ) (N )) (P )) (NP-SBJ *

No. 3 Oct The person to the left of the stool carried the traffic-cone towards the trash-can. α α β α α β α α β α Track2 Track3 Track1 Track0 1

1 7.35% 74.0% linefeed point c 200 Information Processing Society of Japan

NINJAL Project Review Vol.3 No.3

Studies of Foot Form for Footwear Design (Part 9) : Characteristics of the Foot Form of Young and Elder Women Based on their Sizes of Ball Joint Girth

149 (Newell [5]) Newell [5], [1], [1], [11] Li,Ryu, and Song [2], [11] Li,Ryu, and Song [2], [1] 1) 2) ( ) ( ) 3) T : 2 a : 3 a 1 :

JFE.dvi


自然言語処理21_249


28 Horizontal angle correction using straight line detection in an equirectangular image

4. C i k = 2 k-means C 1 i, C 2 i 5. C i x i p [ f(θ i ; x) = (2π) p 2 Vi 1 2 exp (x µ ] i) t V 1 i (x µ i ) 2 BIC BIC = 2 log L( ˆθ i ; x i C i ) + q

研究シリーズ第40号

Z: Q: R: C: sin 6 5 ζ a, b

matsuda.dvi

IPSJ SIG Technical Report Vol.2017-SLP-115 No /2/18 1,a) 1 1,2 Sakriani Sakti [1][2] [3][4] [5][6][7] [8] [9] 1 Nara Institute of Scie

ii 3.,. 4. F. ( ), ,,. 8.,. 1. (75% ) (25% ) =7 24, =7 25, =7 26 (. ). 1.,, ( ). 3.,...,.,.,.,.,. ( ) (1 2 )., ( ), 0., 1., 0,.

浜松医科大学紀要

[1], B0TB2053, i

Table 1 Means and standard deviations of topic familiarity for the topics used in the study Note. standard deviations are in parenthesis.




2000年度『数学展望 I』講義録

TOP URL 1

24 I ( ) 1. R 3 (i) C : x 2 + y 2 1 = 0 (ii) C : y = ± 1 x 2 ( 1 x 1) (iii) C : x = cos t, y = sin t (0 t 2π) 1.1. γ : [a, b] R n ; t γ(t) = (x

Hanbury-Brown Twiss (ver. 2.0) van Cittert - Zernike mutual coherence

ii 3.,. 4. F. (), ,,. 8.,. 1. (75%) (25%) =7 20, =7 21 (. ). 1.,, (). 3.,. 1. ().,.,.,.,.,. () (12 )., (), 0. 2., 1., 0,.

IPSJ SIG Technical Report 1, Instrument Separation in Reverberant Environments Using Crystal Microphone Arrays Nobutaka ITO, 1, 2 Yu KITANO, 1

第5章 偏微分方程式の境界値問題

Abstract This paper concerns with a method of dynamic image cognition. Our image cognition method has two distinguished features. One is that the imag

ohpmain.dvi

Isogai, T., Building a dynamic correlation network for fat-tailed financial asset returns, Applied Network Science (7):-24, 206,

Input image Initialize variables Loop for period of oscillation Update height map Make shade image Change property of image Output image Change time L

IS1-09 第 回画像センシングシンポジウム, 横浜,14 年 6 月 2 Hough Forest Hough Forest[6] Random Forest( [5]) Random Forest Hough Forest Hough Forest 2.1 Hough Forest 1 2.2


III 1 (X, d) d U d X (X, d). 1. (X, d).. (i) d(x, y) d(z, y) d(x, z) (ii) d(x, y) d(z, w) d(x, z) + d(y, w) 2. (X, d). F X.. (1), X F, (2) F 1, F 2 F

‚åŁÎ“·„´Šš‡ðŠp‡¢‡½‹âfi`fiI…A…‰…S…−…Y…•‡ÌMarkovŸA“½fiI›ð’Í

corpus.indd

カルマンフィルターによるベータ推定( )

x T = (x 1,, x M ) x T x M K C 1,, C K 22 x w y 1: 2 2

kawa (Spin-Orbit Tomography: Kawahara and Fujii 21,Kawahara and Fujii 211,Fujii & Kawahara submitted) 2 van Cittert-Zernike Appendix A V 2

Mantel-Haenszelの方法

December 28, 2018

main.dvi

N cos s s cos ψ e e e e 3 3 e e 3 e 3 e

untitled

³ÎΨÏÀ

untitled

微分積分 サンプルページ この本の定価 判型などは, 以下の URL からご覧いただけます. このサンプルページの内容は, 初版 1 刷発行時のものです.


,.,. NP,., ,.,,.,.,,, (PCA)...,,. Tipping and Bishop (1999) PCA. (PPCA)., (Ilin and Raiko, 2010). PPCA EM., , tatsukaw

日本看護管理学会誌15-2

IPSJ SIG Technical Report Pitman-Yor 1 1 Pitman-Yor n-gram A proposal of the melody generation method using hierarchical pitman-yor language model Aki


Optical Flow t t + δt 1 Motion Field 3 3 1) 2) 3) Lucas-Kanade 4) 1 t (x, y) I(x, y, t)

14 2 5

授受補助動詞の使用制限に与える敬語化の影響について : 「くださる」「いただく」を用いた感謝表現を中心に

CVaR

橡表紙参照.PDF

.2 ρ dv dt = ρk grad p + 3 η grad (divv) + η 2 v.3 divh = 0, rote + c H t = 0 dive = ρ, H = 0, E = ρ, roth c E t = c ρv E + H c t = 0 H c E t = c ρv T

DEIM Forum 2009 C8-4 QA NTT QA QA QA 2 QA Abstract Questions Recomme

独立行政法人情報通信研究機構 Development of the Information Analysis System WISDOM KIDAWARA Yutaka NICT Knowledge Clustered Group researched and developed the infor

1 (Contents) (1) Beginning of the Universe, Dark Energy and Dark Matter Noboru NAKANISHI 2 2. Problem of Heat Exchanger (1) Kenji

18 I ( ) (1) I-1,I-2,I-3 (2) (3) I-1 ( ) (100 ) θ ϕ θ ϕ m m l l θ ϕ θ ϕ 2 g (1) (2) 0 (3) θ ϕ (4) (3) θ(t) = A 1 cos(ω 1 t + α 1 ) + A 2 cos(ω 2 t + α

7 π L int = gψ(x)ψ(x)φ(x) + (7.4) [ ] p ψ N = n (7.5) π (π +,π 0,π ) ψ (σ, σ, σ )ψ ( A) σ τ ( L int = gψψφ g N τ ) N π * ) (7.6) π π = (π, π, π ) π ±

‰gficŒõ/’ÓŠ¹


I. (CREMONA ) : Cremona [C],., modular form f E f. 1., modular X H 1 (X, Q). modular symbol M-symbol, ( ) modular symbol., notation. H = { z = x

The copyright of this material is retained by the Information Processing Society of Japan (IPSJ). The material has been made available on the website

Transcription:

1,a) 1 2 1. [1] BCCWJ) A B C 3 A B B C ) 1 NINJAL, Tachikawa, Tokyo 190 8561, Japan 2 MEXT, Chiyoda, Tokyo 100 8959, Japan a) masayu-a@ninjal.ac.jp ) A, B, C Information Structure) Information Status) Topic) Focus) [2] old) new) accessible) c 2015 Information Processing Society of Japan 1

: 2.1-2.5) B 2.7) A B 2.7) 3) 4) 5) 2. 2.1 substring) subsequence) 2.5 1 4 3 A.1 2.2 LCStr, LCS 2.2.1 s = s 1,..., s m, t = t 1,..., t m character) morpheme) string) character) character-based string) morpheme) morpheme-based) substring) n n-gram s i n-gram s i,...,i n+1 subsequence) p p-mer s p-mer i = i 1,..., i p 1 i 1 < i 2 < < i p s ) s[ i] 2.2.2 Longest Common String: LCStr) Longest Common String) abbreviation LCS 2.2.2 Longest Common Subsequence) LCS LCStr, LCS s, t : LCStrs, t) = arg max n s i,...,i n+1 j,s i,...,i n+1=t j,...,j n+1 s, t LCStr ) : c 2015 Information Processing Society of Japan 2

LCStr s, t) = max n i, j,s i,...,i n+1=t j,...,j n+1 [0,1] : Score LCStr s, t) = 2 LCStr s + t 2.2.3 Longest Common Subsequence: LCS) Levenshtein s, t Longest Common Subsequence: LCS) : LCSs, t) = arg max i s[ i] j,s[ i]=t[ j] s, t LCS ) : LCSs, t) = max i i, j:s[ i]=t[ j] [0,1] : Score LCS s, t) = 2 LCS s + t 1 1 2 ) Levenshtein ) LCS : d Levenshtein s, t) = s + t 2 LCS LCS 2.4.2.2 Ulam 2.2.4 LCS LCStr LCS LCS LCS LCSC, R) R ) LCSC, R) C ) LCSC, R) R = LCSC, R) C = arg max i j j j1) i, j,c[ i]=r[ j] arg max i i i i1) i, j,c[ i]=r[ j] R WLCS C, R) P WLCS C, R) R WLCS C, R) = α LCS RC,R) LCS LCS R P WLCS C, R) = α LCS CC,R) LCS LCS s Score γ) WLCS C, R) = 1 + γ2 )R WLCS C, R)P WLCS C, R) R WLCS C, R) + γ 2 P WLCS C, R) 2.3 / 2.3.1 2.3.1.1 ROUGE-L [3] ROUGE-L [3] LCS) Score γ) ROUGE-L C, R) = 1 + γ2 ) R LCS C, R) P LCS C, R) R LCS C, R) + γ 2 P LCS C, R) R LCS C, R) P LCS C, R) : R LCS C, R) = LCSC,R) R P LCS C, R) = LCSC,R) C c i C r j R LCS 2.3.1.2 ROUGE-W [3] LCS LCSC, R) R + LCSC, R) C fx) : fx + y) > fx) + fy), x > 0, y > 0, x N, y N N ) LCS ROUGE-W fx) = x α Score γ) C, R) WLCS 2.3.1.3 ROUGE-N [3], [4] ROUGE-N [3], [4] n-gram c 2015 Information Processing Society of Japan 3

Score R) C, R) = ROUGE-N e n-gram clip C,R) e e n-gramr) e e n-gramc) C n-gram n-gramr) R n-gram n-gram clip C, R) n-gram e n-gramc) e n-gramr) : n-gram clip C, R) { n-gramc) if n-gramc) n-gramr) = n-gramr) otherwise 2.3.1.4 ROUGE-SU) [3], [5] ROUGE-S 2-mer Score γ) ROUGE-S C, R) = 1 + γ2 )P s C, R)R s C, R) R S C, R) + γ 2 P S C, R) P S C, R) R S C, R) : e 2-mer clip C,R) P S C, R) = e e 2-merC) e 2-mer clip C,R) R S C, R) = e e 2-merR) p-merc): p-mer p-merr): p-mer p-mer clip C, R) p-mer e p-merc) p-mer e p-merr) : p-mer clip C, R) { p-merc) if p-merc) p-merr) = p-merr) otherwise ROUGE-SU ROUGE-S p = 2 p 2 2.3.1.5 ESK [6] ESK [6] p-mer e e e Score p-mer ESK = C, R) u p-merc) v p-merr) λ e p) u u ) + u,u p-merc) λ e p δu, v) u v v,v p-merr) λ e p) v v ) [6] 2-mer 2.3.2 2.3.2.1 BLEU[7] BLEU [7] n n-gram ω n ) P n-gram C, R) = BLEU e n-gram clip C,R) e e n-gramc) N Score BLEU C, R) = BP C, R) exp ω n log P n-gram C, R)) BLEU n=1 N n ω n = 1 Brevity Penalty BP) BPC, R) = { 2.3.2.2 IMPACT [8] e 1 if C > R exp1 r c ) if C R IMPACT[8] LCS LCStr RN α r e β )) 1 r=0 e LCStrC R IP C, R) = r),r r) ) β R β RN α i e β )) 1 r=0 e LCStrC P IP C, R) = r),r r) ) β C β α rr RN) c 2015 Information Processing Society of Japan 4

α < 1.0) β LCStr β > 1.0) C 1) = C R 1) = R C r) = C r 1) \ {LCStrC r 1), R r 1) )} R r) = R r 1) \ {LCStrC r 1), R r 1) )} Score IP = 1 + γ2 )R IP P IP R IP + γ 2 P IP 2.4.1.1 LCStr LCStr RN 2.3.2.3 RIBES [9] RIBES [9] Score RIBES = d Kendall 1-gram align C, R)) P RIBES C, R)) α ) BPC, R) ) β d Kendall µ, ν) 2.4.2.2 µ, ν Kendall 1-gram align µ, ν) [9] wonder 2 1-gram ) P RIBES C, R) = 1-gram align C,R) C 1-gram align µ, ν) wonder ) α β BLEU BP P RIBES C, R) : Score Kendall LRscore C, R) = α BP C, R) d Kendall Ĉ, ˆR)+ 2.4 1 α)score BLEU 2.4.1 ) [11] [0,1] ) : Score K s, t) = K s, t) K s, s) K t, t) - γ : Score γ) 1 + γ 2 )K s, t) K s, t) = K s, s)) 2 + γ 2 K t, t)) 2 2.4.1.1 All String Kernel or Exact Matching Kernel) n u F all str P RIBES C, R) = Score P ) C, R) ROUGE-1 e e 1-gram clip C,R) = e 2.3.2.4 LRscore [10] e 1-gramR) LRscore [10] Hamming Kendall Score Hamming LRscore C, R) = α BP C, R) d Hamming Ĉ, ˆR)+ 1 α)score BLEU Φ str : σ F all str R σ Φ str = ϕ us)) u σ Kn-grams, t) = Φ str s), Φ str t) F all str = u σ ϕ us)ϕ t s) ϕ us) = {i s i... = u} : K all seq s, t) = min s, t ) n=1 s n+1 i=1 t n+1 j=1 δs i...i+n 1, t i...i+n 1 ) 2002 c 2015 Information Processing Society of Japan 5

n-gram Length Weighted All String Kernel or Length Weighted Exact Matching Kernel) K all seq s, t) = min s, t ) n=1 s n+1 i=1 t n+1 j=1 ω s δs i...i+n 1, t i...i+n 1 ) ω n n 2.3.2.2 IMPACT n- Suffix Tree 2.4.1.2 n- n-gram Spectrum Kernel) n n-gram) n u Fn-gram Ψ seq : σ F all seq R σ Ψ seqs) = ψ vs)) v σ ψ vs) = { i s[ i] = v} K all seq s, t) = Ψ seqs), Ψ seqt) F all seq = v σ ψ vs) ψ vt) ψ vs) = { i s[ i] = v} K all seq s, t) O s t ) ϵ K all seq s, ϵ) = K all seq t, ϵ) = 1 K all seq s, t) K all seq s a, t) = K all seq s, t) + 1 i t,j:t j=a K all seq s, t i...j 1) K all seq s a, t) = K all seq s, t i...j 1 ) K all seq s a, t b) = K all seq s a, t) + δa, b)ks, t) t 2.4.1.4 p p-mer) p v Fp-mer s Φ n str : σ Fn-gram R σ n Φ n str = ϕn us)) u σ n Kn-grams, t) = Φ n str s), Φn str t) Fn-gram = u σ p ϕ n us)ϕ n t s) ϕ n us) = {i s i...i+n 1 = u} : Kn-grams, t) = s n+1 i=1 t n+1 j=1 δs i...i+n 1, t j...j+n 1 ) ROUGE-N Kn-gramC, R) n-gram Kn-grams, t) 1-gram 1-mer BLEU BP 2.4.1.3 v F all seq Ψ p seq : σ Fp-mer R σ p Ψ p seq s) = ψp vs)) v σ ψvs) = { i s[ i] = v} Kp-mers, t) = Ψ p seq s), Ψp seq t) Fp-mer = ψvs) p ψvt) p v σ p ψvs) p = { i s[ i] = v} ROUGE-S K 2-mer C, R) 2-mer ROUGE-SU K 1-mer,2-mer C, R) 1-mer, 2-mer Kp-mers, t) 2.4.1.5 : p-mer λ ESK [6] p v Fp-mer c 2015 Information Processing Society of Japan 6

1 ) ) [0, 1] [0, ] [0, ] [ 1, 1] IMPACT 2.3.2.2 [8] Score γ) K ) 2.4.1.1 all str n-gram) ROUGE-N 2.3.1.3 BLEU 2.3.2.1[7] Score γ) Kn-gram LRscore 2.3.2.4[10] n- 2.4.1.2 Score γ) K ) 2.4.1.3 all seq p-mer) ROUGE-SU) 2.3.1.4 [3], [5] Score γ) Kp-mer ESK 2.3.1.5[6] Score γ) Kgap p-mer 2.4.2.1 Score rank θ p-mer 2.4.1.4 p-mer 2.4.1.5 ) Score footrule d footruleθ=1) RIBES? 2.3.2.3 [9] Score Spearman d Spearmanθ=2) 2) Spearman s ρ Pearson s LRscore 2.3.2.4[10] Score Hamming d Hamming 2.4.2.2 RIBES 2.3.2.3[9] Score Kendall d Kendall Kendall s τ ) LRscore 2.3.2.4[10] d Caylay d Ulam ) ROUGE-L 2.3.1.1 Score LCS d Levenshtein 2.2.3 ) ROUGE-W 2.3.1.2[3] Score γ) WLCS ) Score LCStr Kendall d Kendall 1, 4, 3, 2), 1, 2, 3, 4)) = 3 1 4 3 2 1 4 2 3 ) 1 4 2 3 1 2 4 3 ) 1 2 4 3 1 2 3 4 ) d Kendall 2, 3, 1, 4), 1, 2, 3, 4)) = 2 2 3 1 4 2 1 3 4 ) 2 1 3 4 1 2 3 4 ) Caylay d Caylay 1, 4, 3, 2), 1, 2, 3, 4)) = 1 1 4 3 2 1 2 3 4 ) Ulam d Ulam 1, 4, 3, 2), 1, 2, 3, 4)) = 2 1 4 3 2 1 2 4 3 ) 1 2 4 3 1 2 3 4 ) d Caylay 2, 3, 1, 4), 1, 2, 3, 4)) = 2 2 3 1 4 1 3 2 4 ) 1 3 2 4 1 2 3 4 ) d Ulam 2, 3, 1, 4), 1, 2, 3, 4)) = 1 2 3 1 4 1 2 3 4 ) 1 c 2015 Information Processing Society of Japan 7

Kgap p-mers, t) = Ψ gap seq p p s), Ψgap seq t) Fp-mer = v σ p ψ gap p v s) ψ gap p t) ψv gap p s) = i:v=s[ i] λl i) li) = s i1,...,i v i = i 1,..., i v ) 2.4.2 [12] m µ, ν S m 2 2.4.2.1 m θ- : m d Rank θ µ, ν) = µi) νi) θ ) 1/θ i=1 θ = 1 Spearman footrule m d Footrule µ, ν) = µi) νi) ) i=1 θ = 2 Euclid Euclid 2 Spearman m d Spearman µ, ν) = µi) νi) 2 ) i=1 Spearman Euclid 2 [-1, 1] Spearman ρ Spearman s ρ = 1 6 d Spearman µ, ν) m 3 m µ, ν Pearson *1 Hamming d Hamming µ, ν) = v m δµi), νi)) i=1 Hamming 1) *1 2.4.2.2 µ ν Levenshtein 1 Kendall : Kendall d Kendall Swap) Kendall mm 1) 2 d Kendall = minarg max δπ q q=1 π 2k q, k q +1)) µ, ν)) q d Kendall = m m i=1 j=i+1 χi, j) χ i, j 0 1 : χ = { 1 if µi) µj))νi) νj)) < 0, 0 if µi) µj))νi) νj)) 0 [0,1] : Score Kendall = 1 2 d Kendall µ, ν) m 2 m [-1,1] Kendall τ Kendall s τ = 1 4 d Kendall µ, ν) m 2 m Cayley : Cayley d Caylay Swap) d Caylay = minarg max δπ q q=1 π 2k q, l q )) µ, ν)) q Ulam : Ulam d Ulam i, i + 1,..., j 1, j c 2015 Information Processing Society of Japan 8

µ ν Ulam d Ulam µ, ν) = m LCSµ, ν) [0,1] : Score Ulam µ, ν) = 1 d Ulam µ, ν) m LCSµ, ν) = m = Score LCS µ, ν) [13] swap Kendall Ulam 2.4.2.3 Spearman s ρ Kendall s τ Daniels : 1 3m + 2) 2m + 1) τ m 2 m 2 ρ 1 m 1 3τ 2ρ 1 d Caylay d Kendall Footrule Kendall Cayley Diaconis-Graham inequality): d Kendall + d Caylay d Footrule 2 d Kendall Spearman Kendall Durbin-Stuart inequality): 4 3 d Kendall 1 + d Kendall m ) d Spearman 2.5 1 [14] 1 Score {Score } ) ω Score = log Score = ω ΠScore ω 1 ω w log Score ) substring : n-gram ) subsequence : p-mer ) Ulam 2.4.2.3 2.6 2 BCCWJ-SUMM) GROSS) RETELLING) 3 ) ) 2 2.6.1 BCCWJ-SUMM C BCCWJ-SUMM C BCCWJ Yahoo! 15 ) BCCWJ 1 19 BCCWJ PN A) 40 50-100 PC 100 200 2014 9 c 2015 Information Processing Society of Japan 9

2 BCCWJ-SUMM C 100-200 19 BCCWJ-SUMM L 3 47 8 GROSS C 71,111,113 GROSS L 4 7,6,3 RETELLING I 10 5 RETELLING K 3 3,3,3 3 RETELLING M 4 10 4 BCCWJ-SUMM L 19 3 3 BCCWJ-SUMM C FileID A 01 106 198 A 02 112 195 B 02 98 149 B 03 74 100 C 01 63 100 C 02 63 99 C 03 53 100 D 01 55 100 D 02 55 100 D 03 48 99 D 05 55 99 E 01 58 99 E 02 46 98 E 03 54 100 E 04 60 99 E 05 48 100 E 06 56 98 F 01 57 100 F 02 58 100 2.6.2 BCCWJ-SUMM L BCCWJ-SUMM L BCCWJ BCCWJ- SUMM C 50-100 3 ) 8 47 4 FileID A 01 16 6 A 02 15 5 B 02 15 5 B 03 18 6 C 01 15 5 C 02 15 5 C 03 15 5 Q 30 10 4.3 ) 2014 8 10 1 : BCCWJ-SUMM LP)) BCCWJ-SUMM LT)) 2.6.3 GROSS C GROSS C Yahoo! 15 ) 6.6) 6.4) 6.0) 3 *2 150 250 3 300 :71:111 :113295/300) 2.6.4 GROSS L GROSS L 8 20-50 ) GROSS C 10 6.6) 6.4) 6.0) 3 2 5 *2 [15] c 2015 Information Processing Society of Japan 10

4 7 4 6 4 3 4 145 max 227, min 85 ) 1 : GROSS LP)) GROSS LT)) 2.6.5 RETELLING I Retelling [16], [17] [18] 5 10 3 ) 5 10 50 ) ) 1 3 5 7 9 ) 2 4 6 8 10 ) RETELLING IT)) 2.6.6 RETELLING K [19] 3 *3 1 3 3 9 4 3 2 2 ) ) ) RETELLING K *3 1 20 2 30 3 20 1 : RETELLING KP)) RETELLING KT)) 2.6.7 RETELLING M [20] 10 20-50 ), 10 ) 4 284 min:150 max:451 ) 107 min:74 max:152 ) 10 4 40 ) 1 : RETELLING MP)) RETELLING MT)) 2.7 30 n-gram 1,2,3,4) char/mrph) n-gram 2, 3, 4) char/mrph) p-mer 2,3,4) char/mrph) p-mer 2, 3, 4) char/mrph) 1-gram +Footrule char/mrph) =Spearman) 1-gram +Kendall char/mrph) A 1,A 2 Mean) SD) c m MeCab-0.98+IPADIC-2.7.0 ) p 0.05 2.7.1 3 n-gram1),n-gram2),pmer2),kendall unigramn-gram1)) GROSS LT) BCCWJ-SUMM LT) Bigramn-gram2)), skip-bigramp-mer2)) c 2015 Information Processing Society of Japan 11

Bigramn-gram2)) skip-bigramp-mer2)) bigram Kendall bi-gram BCCWJ-SUMM C BCCWJ-SUMM LP), GROSS C GROSS LP)) ) 2.7.2 6 F ) 0.05 2 ) * 4 BCCWJ-SUMM LP) GROSS LP) RETELLING KP) RETELLING MP) BCCWJ-SUMM LP) GROSS LP) Kendall char n-gram2,3,4) char, n- gram2,3,4, 2, 3, 4) mrph, Kendall mrph Footrule mrph, BCCWJ-SUMM LP) RETELLING KP) n-gram3,4) mrph BCCWJ-SUMM LP) RETELLING KM) GROSS LP) RETELLING {K,M}P) RETELLING KP) RETELLING MP) n-gram 3, 4) mrph,p-mer3,4, 3, 4) n-gram1) *4 = 0.05 2 ) BCCWJ-SUMM LT) GROSS LT) RETELLING IT) RETELLING KT) RETELLING MT) BCCWJ-SUMM LT) GROSS LT) Kendall char n-gram2,3,4) char, n- gram2,3,4, 2, 3, 4) mrph, Kendall mrph Footrule mrph, BCCWJ-SUMM LT) RETELLING {I,K,M}T) GROSS LT) RETELLING {I,K,M}T) RETELLING IT) RETELLING KT) n-gram1,4, 2) char, p- mer2, 2) char RETELLING IT) RETELLING MT) Kendall char RETELLING IT) RETELLING MT) n-gram2, 2, 3, 4) char, p-mer2,3,4, 2, 3, 4) char n-gram1,2, 2, 3, 4) mrph, p- mer2,3,4, 2, 3, 4) mrph RETELLING {I,K}) RETELLING M) RETELLING I) RETELLING K) BCCWJ-SUMM C GROSS C c 2015 Information Processing Society of Japan 12

$"!!##!",!##!"+!##!"*!##!")!##!"!##!"'!##!"&!##!"%!##!"$!##!"!!## -../0123445.# -../01234456789# $"!!##!",!##!"+!##!"*!##!")!##!"!##!"'!##!"&!##!"%!##!"$!##!"!!## -../0123445.# -../01234456789#!"#$%&')*&$+,! -../012344567:9# ;<=225.# ;<=2256789# ;<=22567:9# <>:>66?@;5?7:9# <>:>66?@;5A789#!"#$%&')#%!*! -../012344567:9# ;<=225.# ;<=2256789# ;<=22567:9# 3 <>:>66?@;5A7:9# <>:>66?@;54789# <>:>66?@;547:9# <>:>66?@;5?7:9# <>:>66?@;5A789# <>:>66?@;5A7:9# <>:>66?@;54789# <>:>66?@;547:9# ) BCCWJ-SUMM C BCCWJ-SUMM LP) n-gram2) char, n-gram3) char, n-gram4) char BCCWJ-SUMM C) BCCWJ-SUMM LP)) ) GROSS C GROSS LP) n-gram2,3,4) char, $"!!##!",!##!"+!##!"*!##!")!##!"!##!"'!## 4BCD#!"&!##!"%!##!"$!## 2E#!"!!## $"!!##!",!##!"+!##!"*!##!")!##!"!##!"'!## 4BCD#!"&!##!"%!##!"$!## 2E#!"!!## -../0123445.# -../01234456789#!"#$%&')*&$+,! -../012344567:9# ;<=225.# ;<=2256789# ;<=22567:9# <>:>66?@;5?7:9# <>:>66?@;5A789#!"#$%&&')*+! <>:>66?@;5A7:9# <>:>66?@;54789# <>:>66?@;547:9# -../0123445.# -../01234456789# -../012344567:9# ;<=225.# ;<=2256789# ;<=22567:9# <>:>66?@;5?7:9# <>:>66?@;5A789# <>:>66?@;5A7:9# <>:>66?@;54789# <>:>66?@;547:9# n-gram1),n-gram2),p-mer2),kendall: ) n-gram2,3,4) mrph, Footrule mrph, Kendall mrph wikipedia BCCWJ-SUMM LP) BCCWJ-SUMM LT), GROSS LP) GROSS LT), RETELLING KP) RETELLING KT), RETELLING MP) RETELLING MT) 2.7.3 4BCD# 2E# 4BCD# n-gram n-gram p-mer, Footrule, Kendall n-gram, p-mer n, p n-gram, p-mer n or p) n or p) n or p) n-gram, p-mer n-gram1) * Kendall * n- gram1) * Kendall * 2E# c 2015 Information Processing Society of Japan 13

2.8 n-gram p-mer 7 ) 3 4 3. Information Structure)[21] : Information Status) Topic) Focus) ) Götze[2] BCCWJ-SUMM 3.1 Götze[2] 5 Götze ) discourse referents) 5 Götze ) [2] Layers Tags Description giv Given ) Information Status) acc Accessible) new cat nil New ) cataphor ) ab Aboutness topic non-referential ) Topic) fs Frame setting topic nf New Information Focus Focus) cf Contrastive Focus ) information status) retrievability) given) accessible) new Prince [22] Prince { discourse-old) discoursenew)} { hearer-old) hearer-new)} 6) Prince [22] giv =, ) acc =, ) new =, ) Prince, ) topic) Jacobs[23] aboutness topic) frame setting topic) 2 what the sentence is about ) the frame within which the sentence holds ) : Frame-setting [23], p.656) X,Y) X Y Y X focus) contrast focus) Götze[2] Cook [24] c 2015 Information Processing Society of Japan 14

6 Prince Prince giv ) evoked acc) unused - new ) brand-new 588 aboutness topics Fleiss κ)0.19 0.57 3.2 3.2.1 NP NP new new - ) acc-inf new - acc-inf new - 3.2.2 Götze[2] 7 giv-active: giv-inactive *5 : acc-sit: ) acc-aggr: [2] Peter went shopping with Maria. They bought many flowers. acc-inf set-rel) giv acc-inf acc-inf: [2] part-whole: The garden beautiful. Its entrance is just across this river. set-rel: The flowers in the garden blossom. The flowers near the gate blossom violet. set-rel: The children swam in the lake. The family experienced a beautiful day. entity-attribute: The flowers enchanted Peter. Their scent was wonderful. - - - -- referent given) *5 inactive discourse-new + hearer-new semi-active textually accessible c 2015 Information Processing Society of Japan 15

7 Coarse) Fine) Description giv underspecified) given: giv-active giv-inactive active: inactive: acc underspecified) accessible: new nil cat) acc-sit acc-aggr acc-inf acc-gen situationally accessiblly accessible: aggregation: inferable: part-whole, set-relsubset/superset),entityattribute ) general: new: non-referential: cataphor: acc-gen: Type) ) Token) acc-gen acc-gen ) new ) new new: nil cat catphora) *6 3.3 MAMA [25] 1 [2] 3 3 1 2 { ) *6 2 ) } 3 3 2 4 4 discourse-old) ) giv-active ) giv-inactive giv-active ) acc-aggr discourse-new) acc-sit acc-inf activeness) - ) hearer-old) ) acc-gen ) ) hearer-new) new 3 ) ) ) ) 4 8 Fine)) Fine) c 2015 Information Processing Society of Japan 16

9 4 B-C ) B-C a-g a-i g-a g-i new nil acc-gen 3 3 acc-inf 1 3 2 6 giv-active 1 1 giv-inactive 2 4 6 new 2 2 1 2 10 17 nil 4 1 6 20 17 6 11 1 8 16 22 64 Coarse) 2 A, B, C 2 Cohen s)κ 3 Fleiss s)κ 1 B-C 4 0.50 κ 9 4 B-C nil ) acc ) 3.4 ) giv ) acc ) BCCWJ NAIST [26] 2014 - acc-aggr [27] 4. 2 ) 4.1 Kennedy [29] Dundee Eye- Tracking Corpus 10 20 5 40 Dundee Eye-Tracking Corpus Demberg [30] Gibson Dependency Locality TheoryDLT) integration cost[31] Hale suprisal[32] Dundee Corpus Roland [33] Demberg [34] Dundee c 2015 Information Processing Society of Japan 17

8 Cohen s κ) A-B) B-C) C-A) Fleiss s κ 1 Fine) 1 54 0.54 0.67 0.40 0.53 Coarse) 0.53 0.71 0.38 0.54 2 Fine) 3 77 + 69 + 37 = 183 0.43 0.41 0.42 0.42 Coarse) 0.54 0.44 0.47 0.48 3 Fine) 3 37 + 51 + 42 = 130 0.38 0.40 0.36 0.37 Coarse) 0.41 0.49 0.43 0.44 4 Fine) 3 16 + 33 + 15 0.50 0.49 0.49 0.49 Coarse) 0.49 0.52 0.51 0.50 Corpus 4.2 4.2.1 [35] Linger *7 5 1 Yes/No Question 4.2.2 SRResearch EyeLinkCL 5 1 Yes/No Question 5 1/2 interest area grid ) interest area interest area BCCWJ 4.4 *7 http://tedlab.mit.edu/~dr/linger/ 5 4.3 : 5 : {,,,, } 5 : 15 [15] *8 [37] [38] *8 http://www.kecl.ntt.co.jp/icl/lirg/resources/ goitokusei/goi-test.html c 2015 Information Processing Society of Japan 18

4.4 1 1 1 First pass time Total time Regression path time ) ) ) A 3 A 4,A 5, ) 4.5 2014 12 10 98 43 5 5. 2 1 3 accessible 4 [39] 5.1 5.1.1 6 1b) 1a) [40] 1) a. The lady i [ that GAP i visited the banker ]... b. The lady i [ that the banker visited GAP i ]... )[41][42] ) [31] 6 ) 6 ) 2) a. GAP i i... b. GAP i i... [43] Roland [44] c 2015 Information Processing Society of Japan 19

1a) 57% 43% 1b) 2% 98% 4a) 4b) Roland [44] abstract ) 3) a. ) The banker was very friendly. b. ) There was a dinner party on Saturday night. 4) a. The lady i [ that GAP i visited the banker ] enjoyed the dinner very much. b. The lady i [ that the banker visited the GAP i ] enjoyed the dinner [45] 4 ) BCCWJ 400 70% 30% 20% 80% 5) [45]p.63) ) 5)1-a. 1-b. 2-a. 2-b. 1 ) 1 ) 2 ) ) 2 ) ) 3. 3 ) 5.1.2 [46][47][48] Gordon [46] 6) Gordon [46] *9 ) 6) a. It was the barber/john that saw the lawyer/bill in the parking lot. b. It was the barber/john that the lawyer/bill saw in the parking lot. [49][50][51] [50] 5 ) 7a) 7b) [50] 5 ) ) 7) a. b. [49] 300 2085 1756 84%) 329 16%) 656 6a) 31% 6b) 8% 6c) 45% [49] ) 8) a. GAP i i b. GAP i i *9 Gordon description: the barber, the lawyer ) name: John, Bill ) c 2015 Information Processing Society of Japan 20

c. ) GAP i 9 15 i d. - / + / 10 10 [49] ) - + 0.57 205/357) - + 0.75 86/114) [52] [52] ) 9) 0. 2 1. 2 2-a. 2-b. focus) ) ) ) ) ) 5.2 2 3 accessible accessible-new 4 5.1 kappa PC BCCWJ BCCWJ [53] 6. 2 3 4 2 5.1 c 2015 Information Processing Society of Japan 21

5.2 BCCWJ-Trans 11 BCCWJ-Trans A.1 2 2 : σ : s = s 1,..., s m, t = t 1,..., t m character), character-based): σ s i σ σ morpheme), morpheme-based): σ s i σ σ string): character) character-based string) morpheme) morpheme-based) substring): n n-gram s i n-gram s i,...,i n+1 subsequence): p p-mer s p-mer i = i 1,..., i p 1 i 1 < i 2 < < i p s ) s[ i] / reference): / R / candidate): / / C distance): X 2 d : X X R dx, y) 0), x = y dx, y) 0), dx, y) = dy, x)), dx, y) + dy, z) dx, z)) absolute value): 0 x θ- norm): x =, x 1,..., x n θ- x θ = n i=1 x i p ) 1/p θ x ) 2- : ) similarity): correlation): [-1,1] 1-1 0 kernel function): ) Ks, t) cosine Ks,t) Ks,s) Kt,t) ) score): [0,1] score prefix): suffix): subset): k k-element rank vector): i i m S m µ = µ1),..., µm) µi) i order vector): i i m T m µi) µ 1 = µ 1 1),..., µ 1 m) µ 1 i) i ) concordant): i j c 2015 Information Processing Society of Japan 22

11 BCCWJ-Trans 6 319 OY 1, OC 1, PN 1, PB 1, PM 1, OW 1 ) 6 319 OY 1, OC 1, PN 1, PB 1, PM 1, OW 1 16 436 OY 6, OC 6, PN 1, PB 1, PM 1, OW 1 10 337 OY 3, OC 3, PN 1, PB 1, PM 1, OW 1 : OY, OC, PN, PB, PM, OW µi) µj))νi) νj)) 0 discordant): : insertion) deletion) substitution) : symmetric group) : :permutation) µ 1 = µ 1 1),..., µ 1 m) µ 1 k 1 ), µ 1 k 2 ),..., µ 1 k r ) µ 1 k 1 ) µ 1 k 2 ), µ 1 k 2 ) µ 1 k 3 ),... µ 1 k 1 ) µ 1 k 2 )... µ 1 k r ) µ 1 k 2 ) µ 1 k 3 )... µ 1 k 1 ) π r = k 1, k 2,..., k r ) 2 ) transposition) π 2 = i, j) adjacent transposition) π 2 = i, i + 1) δ : δi, j) = A.2 { ) 1 i = j) 0 i j) A 1 A 2 BCCWJ-SUMM LT) RETELLING KT),RETELLING MT)) A.3 A 3 0-origin) 0-10754 ) A 4,A 5 ) ) ) 95% [-2.0, 2.0] 2.0 2.0 * 2.5 **) NTT CS JSPS B) 25284083, B) 26770156 B)26770167 [1] Maekawa, K., Yamazaki, M., Ogiso, T., Maruyama, T., Ogura, H., Kashino, W., Koiso, H., Yamaguchi, M., Tanaka, M. and Den, Y.: Balanced corpus of contemporary written Japanese, Language Resources and Evaluation, Vol. 48, pp. 345 371 2014). [2] Götze, M., Weskott, T., Endriss, C., Fiedler, I., Hinterwimmer, S., Petrova, S., Schwarz, A., Skopeteas, S. and Stoel, R.: Information structure, Interdisciplinary Studies on Information Structure Dipper, S., Göetze, M., Skopeteas, S. and Stoel, R., eds.), Working Papers of the SFB 632, Vol. 7, Universiätsverlag, Potsdam, 2nd edition updated version 2014) edition, chapter Information Structure, pp. 147 187 2007). [3] Lin, C.-Y.: ROUGE: A Package for Automatic Evaluation of Summaries, Proc. of Workshop on Summarization Branches Out, Post Conference Workshop of ACL 2004, pp. 74 81 2004). [4] Lin, C.-Y. and Hovy, E.: Automatic Evaluation of Summaries Using N-gram Co-occurrence Statistics, Proc. of the 4th Meeting of the North American Chapter of the Association for Computational Linguistics and Human Language Technology, pp. 150 157 2003). [5] Lin, C.-Y. and Och, F. J.: Automatic Evluation of Machine Translation Quality Using Longest Common Subsequence and Skip-Bigram Statistics, Proceedings of c 2015 Information Processing Society of Japan 23

A 1 score BCCWJ-SUMM C BCCWJ-SUMM LP) BCCWJ-SUMM LT) GROSS C GROSS LP) GROSS LT) Mean SD Mean SD Mean SD Mean SD Mean SD Mean SD n-gram1) c 0.63 0.12 0.64 0.11 0.68 0.27 0.68 0.09 0.65 0.06 0.83 0.07 n-gram2) c 0.36 0.17 0.33 0.15 0.48 0.32 0.24 0.10 0.22 0.09 0.56 0.16 n-gram3) c 0.26 0.17 0.22 0.15 0.40 0.34 0.11 0.07 0.09 0.06 0.41 0.18 n-gram4) c 0.20 0.16 0.16 0.14 0.35 0.34 0.05 0.06 0.04 0.04 0.31 0.18 n-gram 2) c 0.52 0.13 0.53 0.12 0.61 0.28 0.55 0.09 0.51 0.07 0.75 0.10 n-gram 3) c 0.45 0.14 0.45 0.12 0.55 0.29 0.47 0.08 0.42 0.06 0.68 0.11 n-gram 4) c 0.40 0.14 0.39 0.12 0.51 0.30 0.41 0.08 0.36 0.06 0.62 0.12 p-mer2) c 0.34 0.14 0.35 0.13 0.49 0.33 0.39 0.10 0.34 0.07 0.63 0.14 p-mer3) c 0.18 0.13 0.18 0.12 0.37 0.35 0.20 0.08 0.15 0.05 0.45 0.17 p-mer4) c 0.09 0.10 0.09 0.10 0.30 0.36 0.09 0.06 0.06 0.03 0.31 0.17 p-mer 2) c 0.34 0.14 0.35 0.13 0.49 0.33 0.39 0.10 0.34 0.07 0.63 0.14 p-mer 3) c 0.18 0.13 0.18 0.12 0.37 0.35 0.20 0.08 0.16 0.05 0.45 0.17 p-mer 4) c 0.10 0.11 0.09 0.10 0.30 0.36 0.10 0.06 0.06 0.03 0.31 0.17 Footrule c 0.50 0.15 0.50 0.14 0.59 0.29 0.48 0.10 0.45 0.08 0.69 0.14 Kendall c 0.48 0.14 0.47 0.13 0.57 0.29 0.44 0.08 0.41 0.06 0.64 0.13 n-gram1) m 0.60 0.12 0.62 0.11 0.67 0.27 0.62 0.10 0.58 0.07 0.79 0.09 n-gram2) m 0.25 0.16 0.24 0.15 0.41 0.34 0.13 0.08 0.11 0.07 0.42 0.18 n-gram3) m 0.15 0.15 0.14 0.14 0.34 0.35 0.04 0.06 0.03 0.04 0.27 0.18 n-gram4) m 0.10 0.13 0.09 0.12 0.29 0.35 0.02 0.05 0.01 0.02 0.19 0.16 n-gram 2) m 0.46 0.13 0.48 0.12 0.57 0.29 0.48 0.09 0.43 0.07 0.67 0.11 n-gram 3) m 0.38 0.13 0.39 0.13 0.51 0.30 0.39 0.08 0.34 0.06 0.58 0.12 n-gram 4) m 0.32 0.13 0.33 0.12 0.47 0.31 0.33 0.07 0.28 0.05 0.52 0.13 p-mer2) m 0.30 0.14 0.32 0.14 0.47 0.33 0.32 0.10 0.27 0.07 0.55 0.15 p-mer3) m 0.15 0.12 0.15 0.12 0.35 0.35 0.14 0.07 0.11 0.04 0.35 0.17 p-mer4) m 0.07 0.09 0.07 0.10 0.28 0.36 0.06 0.05 0.03 0.02 0.22 0.16 p-mer 2) m 0.31 0.14 0.33 0.14 0.47 0.33 0.32 0.10 0.28 0.07 0.55 0.15 p-mer 3) m 0.16 0.12 0.16 0.13 0.36 0.35 0.14 0.07 0.11 0.04 0.36 0.17 p-mer 4) m 0.08 0.10 0.08 0.10 0.29 0.36 0.06 0.05 0.04 0.02 0.23 0.17 Footrule m 0.48 0.15 0.48 0.14 0.59 0.30 0.44 0.10 0.42 0.10 0.66 0.14 Kendall m 0.46 0.15 0.46 0.14 0.56 0.29 0.41 0.09 0.39 0.08 0.61 0.13 the 42nd Annual Meeting on Association for Computational Linguistics, pp. 311 318 2004). [6] Vol. 47, No. 6, pp. 1753 1765 2006). [7] Papineni, K., Roukos, S., Ward, T. and Zhu, W.-J.: Bleu: a Method for Automatic Evaluation of Machine Translation, Technical report, IBM Research Report RC22176 W0109-022) 2001). [8] Echizen-ya, H. and Araki, K.: Automatic Evaluation of Machine Translation based on Recursive Acquisition of an Intuitive Common Parts Continuum, Proceedings of the MT Summit XII Workshop on Patent Translation, pp. 151 158 2007). [9] Kevin, D. Vol. 21, No. 3, pp. 411 444 2014). [10] Birch, A. and Osborne, M.: LRscore for Evaluation Lexical and Reordering Quality in MT, Proceedings of the Joint 5th Workshop on Statistical Machine Translation and MetricsMATR, pp. 327 332 2010). [11] Shawe-Taylor, J. Cristianini, N. ) Kernel Methods for Pattern Analysis) chapter 11 : 2010). [12] SIG-DMSM-A902-07 2009). [13] Nivre, J.: Non-Projective Dependency Parsing in Expected Linear Time, Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, pp. 351 359 2009). [14] Vol. 22, No. 2, pp. 115 126 2007). [15] 1 CD-ROM 1999). [16] 31 pp. 190 193 2013). [17] 32 2013). [18] 2014 2014). [19] 29 2012). [20] 33 2014). [21] Lambrecht, K.: Information Structure and Sentence Form, Cambredge University Press 1994). [22] Prince, E. F.: Discourse description: Diverse linguistic c 2015 Information Processing Society of Japan 24

A 2 score RETELLING IT) RETELLING KP) RETELLING KT) RETELLING MP) RETELLING MT) Mean SD Mean SD Mean SD Mean SD Mean SD n-gram1) c 0.96 0.09 0.86 0.05 0.91 0.19 0.87 0.08 0.92 0.13 n-gram2) c 0.85 0.12 0.59 0.10 0.78 0.21 0.58 0.20 0.75 0.16 n-gram3) c 0.73 0.17 0.39 0.13 0.65 0.22 0.44 0.21 0.64 0.18 n-gram4) c 0.61 0.19 0.23 0.13 0.50 0.21 0.33 0.18 0.52 0.18 n-gram 2) c 0.94 0.09 0.80 0.06 0.89 0.19 0.80 0.11 0.88 0.13 n-gram 3) c 0.91 0.09 0.76 0.07 0.86 0.19 0.74 0.12 0.84 0.13 n-gram 4) c 0.89 0.10 0.73 0.07 0.84 0.18 0.70 0.13 0.81 0.14 p-mer2) c 0.91 0.13 0.72 0.09 0.85 0.19 0.72 0.13 0.83 0.14 p-mer3) c 0.86 0.16 0.60 0.11 0.79 0.21 0.58 0.16 0.74 0.16 p-mer4) c 0.81 0.18 0.49 0.13 0.73 0.22 0.45 0.16 0.64 0.19 p-mer 2) c 0.91 0.13 0.72 0.09 0.85 0.19 0.72 0.13 0.83 0.14 p-mer 3) c 0.86 0.16 0.60 0.11 0.79 0.21 0.58 0.16 0.74 0.16 p-mer 4) c 0.81 0.18 0.49 0.13 0.73 0.22 0.45 0.16 0.64 0.19 Footrule c 0.88 0.11 0.73 0.07 0.83 0.18 0.75 0.13 0.85 0.14 Kendall c 0.81 0.11 0.66 0.06 0.77 0.17 0.68 0.12 0.78 0.13 n-gram1) m 0.94 0.15 0.85 0.06 0.92 0.18 0.83 0.07 0.90 0.13 n-gram2) m 0.75 0.18 0.48 0.10 0.72 0.21 0.45 0.18 0.64 0.17 n-gram3) m 0.58 0.21 0.21 0.13 0.48 0.19 0.28 0.15 0.47 0.16 n-gram4) m 0.39 0.21 0.09 0.07 0.30 0.15 0.10 0.08 0.27 0.15 n-gram 2) m 0.92 0.15 0.78 0.08 0.88 0.18 0.73 0.10 0.84 0.13 n-gram 3) m 0.89 0.15 0.73 0.08 0.85 0.18 0.66 0.11 0.78 0.13 n-gram 4) m 0.86 0.15 0.69 0.09 0.82 0.18 0.60 0.11 0.73 0.13 p-mer2) m 0.90 0.18 0.70 0.11 0.86 0.19 0.64 0.11 0.78 0.14 p-mer3) m 0.85 0.19 0.56 0.13 0.79 0.19 0.46 0.13 0.66 0.17 p-mer4) m 0.79 0.20 0.44 0.14 0.73 0.20 0.32 0.12 0.54 0.18 p-mer 2) m 0.90 0.18 0.70 0.11 0.86 0.19 0.64 0.11 0.78 0.14 p-mer 3) m 0.85 0.19 0.56 0.13 0.79 0.19 0.46 0.13 0.66 0.17 p-mer 4) m 0.79 0.20 0.44 0.14 0.73 0.20 0.33 0.12 0.54 0.18 Footrule m 0.86 0.16 0.71 0.07 0.83 0.18 0.72 0.10 0.83 0.14 Kendall m 0.80 0.16 0.63 0.06 0.76 0.17 0.66 0.10 0.77 0.13 A 3 ) ID 14 1 0 1 7 209 203 14 1 7 2 231 434 204 14 1 15 3 459 623 165 14 1 18 4 743 772 30 14 1 5 5 825 1026 202 14 1 5 6 1047 1072 26 14 1 4 7 1164 1223 60 14 1 5 8 1245 1298 54 14 1 8 9 1317 1600 284 14 1 4 10 1614 1689 76 14 1 18 11 1722 1836 115 14 1 20 12 2127 2164 38 14 2 5 13 2236 2388 153 14 2 3 14 2399 2860 462 14 2 11 15 3170 3239 70 14 2 12 16 3258 3490 233 14 2 16 17 3507 3738 232 14 2 26 18 3766 3993 228 14 2 24 19 4004 4089 86 14 1 31 20 4371 4396 26 c 2015 Information Processing Society of Japan 25

A 4 :) ID ID First Pass Total Reg. Path 14 1 0 203 0.69 203 0.27 203 0.09 14 1 1 204-0.17 906 0.92 204-0.38 14 1 2 195-0.21 348-0.36 1203 0.52 14 2 0 615 1.76 672 0.58 615 0.09 14 2 1 303 0.95 303 0.31 303 0.09 14 2 2 232 0.30 232-0.24 232-0.17 14 2 3 314 0.65 1048 1.63 314-0.09 14 2 4 144-0.42 907 0.92 144-0.43 14 2 5 612 0.19 612-1.30 612-0.76 14 2 6 322 0.86 322 0.16 422 0.10 14 2 7 0-0.69 0-0.77 0-0.38 14 3 0 0-0.35 0-0.39 0-0.19 14 3 1 553 0.11 1299 0.47 553-0.72 14 3 2 260 0.07 481-0.06 260-0.33 14 3 3 353 0.82 603 0.61 1348 0.84 14 3 4 0-0.69 0-0.77 0-0.38 analyses of a fund-raising text, chapter The ZPG letter: Subjects, definiteness, and information status, pp. 295 325, Benjamins 1992). [23] Jacobs, J.: The dimensions of topic-comment, Linguistics, Vol. 39, pp. 641 681 2001). [24] Cook, P. and Bildhauer, F.: Identifying aboutness topics : two annotation experiments, Dialogue and Discourse, Vol. 4, No. 2, pp. 118 141 2013). [25] Pustejovsky, J. and Stubbs, A.: Natural Language Annotatino for Machine Learning A Guide to Corpus- Buildling for Applications, O Reilly 2012). [26] NAIST Vol. 17, No. 2, pp. 25 50 2010). [27] 2009). [28] 1 2012). [29] Kennedy, A. and Pynte, J.: Parafoveal-on-foveal effects in normal reading, Vision Research, Vol. 45, pp. 153 168 2005). [30] Demberg, V. and Keller, F.: Data from eye-tracking corpora as evidence for theories of syntactic processing complexity, Cognition, Vol. 109, No. 2, pp. 193 210 2008). [31] Gibson, E.: Linguistic complexity: Locality of syntactic dependencies, Cognition, Vol. 68, pp. 1 76 1998). [32] Hale, J.: A probabilistic earley parser as a psycholinguistic model, Proc. of the second conference of the North American chapter of the association for computational linguistics, Vol. 2, pp. 159 166 2001). [33] Roland, D., Mauner, G., O Maera, C. and Yun, H.: Discourse expectations and relative clause processing, Journal of Memory and Language, Vol. 66, No. 3, pp. 479 508 2012). [34] Demberg, V. and Keller, F.: Eye-tracking evidence for integration cost effects in corpus data, Proc. of the 29th meeting of the cognitive science society CogSci-07) 2007). [35] Just, M. A., Carpenter, P. A. and Woolley, J. D.: Paradigms and Processes in Reading Comprehension, Journal of Experimental Psychology: General, Vol. 3, pp. 228 238 1982). [36] ChaKi.NET 2) 5 pp. 39 48 2014). [37] Amano, S. and Kondo, T.: Estimation of mental lexicon size with word familiarity database, Proceedings of International Conference on Spoken Language Processing, Vol. 5, pp. 2119 2122 1998). [38] 2002). [39] 3 Part 2 2012 2012). [40] King, J. and Just, M. A.: Individual Differences in Syntactic Processing: The role of Working Memory, Journal of Memory and Language, Vol. 30, pp. 580 602 1991). [41] Hawkins, J. A.: Processing Complexity and Filler-Gap Dependencies Across Grammars, Language, Vol. 75, pp. 244 285 1999). [42] O Grady, W.: Syntactic Development, The University of Chicago Press 1997). [43] Miyamoto, E. T. and Nakamura, M.: Subject/object asymmetries in the processing of relative clauses in Japanese, Proceedings of the 22nd West Coast Conference on Formal Linguistics Garding, G. and Tsujimura, M., eds.), pp. 342 355 2003). [44] Roland, D., O Meara, C., Yun, H. and Mauner, G.: Processing object relative clauses: Discourse or frequency, Poster presented at the CUNY sentence processing conference 2007). [45] 2011). [46] Gordon, P. C., Hendrick, R. and Johnson, M.: Memory Interference During Language Processing, Jounal of Experimental Psychology: Learning, Memory, and Cognition, Vol. 27, No. 6, pp. 1411 1423 2001). [47] Gordon, P. C., Hendrick, R. and Levine, W. H.: Memory-load -interference in Syntacitic Processing, Psychological Science, Vol. 13, No. 5, pp. 425 430 2002). c 2015 Information Processing Society of Japan 26

A 5 :) ID ID First Pass Total Reg. Path 14 1 0 203 1.10 203 0.61 203 0.15 14 1 1 342 1.71 418 1.27 342 0.18 14 1 2 204 0.66 488 1.63 204-0.06 14 1 3 0-0.45 0-0.43 0-0.21 14 1 4 0-0.45 0-0.43 0-0.21 14 1 5 165 0.81 165 0.41 165 0.08 3 14 1 6 30-0.67 183 0.07 1038 1.41 14 1 7 0-0.90 0-0.86 0-0.42 14 1 8 0-0.45 0-0.43 0-0.21 14 2 0 0-0.45 0-0.43 0-0.21 14 2 1 615 3.80** 672 2.57** 615 0.66 14 2 2 0-0.45 0-0.43 0-0.21 14 2 3 0-0.45 0-0.43 0-0.21 14 2 4 303 1.86 303 1.12 303 0.32 9 14 2 5 0-0.45 0-0.43 0-0.21 14 2 6 232 0.87 232 0.32 232-0.01 14 2 7 0-0.45 0-0.43 0-0.21 14 2 8 0-0.45 0-0.43 0-0.21 14 2 9 86 0.21 86 0.01 86-0.06 4 14 2 10 228 1.29 534 2.30* 375 0.45 14 2 11 147 0.67 147 0.32 734 1.08 14 2 12 281 1.70 281 1.00 281 0.29 14 2 13 144 0.20 398 1.17 144-0.17 14 2 14 339 1.69 339 0.87 339 0.18 14 2 15 170 0.85 170 0.44 170 0.09 14 2 16 0-0.45 0-0.43 0-0.21 14 2 17 116-0.01 116-0.27 116-0.22 14 2 18 0-0.90 0-0.86 0-0.42 14 2 19 0-0.90 0-0.86 0-0.42 14 2 20 0-0.45 0-0.43 0-0.21 14 2 21 231 1.31 231 0.75 231 0.20 14 2 22 265 1.57 265 0.92 934 1.44 14 2 23 0-0.45 0-0.43 0-0.21 14 2 24 0-0.45 0-0.43 0-0.21 14 2 25 0-0.90 0-0.86 0-0.42 14 2 26 0-0.45 0-0.43 0-0.21 14 2 27 0-0.90 0-0.86 0-0.42 14 2 28 322 2.01* 322 1.21 422 0.53 14 2 29 0-0.90 0-0.86 0-0.42 14 2 30 0-0.45 0-0.43 0-0.21 14 2 31 0-0.45 0-0.43 0-0.21 [48] Waters, G., Caplan, D. and Yampolsky, S.: On-line syntactic processing under concurrent memory load, Psychonomic Bulletin and Review, Vol. 10, pp. 88 95 2003). [49] Kahraman, B., Sato, A., Ono, H. and Sakai, H.: Why object clefts are easier than subject clefts in Japanese: Frequency or expectation?, IEICE Technical Report, Vol. 111, No. 170, pp. 67 72 2011). [50] Kahraman, B.: Processing gap-filler dependencies in Japanese and Turkish: Regarding the incrementality of sentence processsing, PhD Thesis, 2011). [51] Kahraman, B., Sato, A., Ono, H. and Sakai, H.: Incremental processing of gap-filler dependencies: Evidence from the processing of subject and object clefts in Japanese, The Proceedings of the 12th Tokyo Conference on Psycholinguistics, pp. 133 147 2011). [52] gap-filler ERP 146 pp. 252 257 2013). [53] Vol. 3, No. 2, pp. 63 83 2012). [54] Vol. 23, No. 1, pp. 10 16 2008). c 2015 Information Processing Society of Japan 27