/ 60 : 1. GLM? 2. A: (pwer functin) x y?
|
|
- えの みつだ
- 5 years ago
- Views:
Transcription
1 / 60 ( ) GLM 1. GLM :, link,, deviance (20 ) 2. GLM : (60 ) 3. GLM ( ): ffset (40 ) kub/ce/ecsj2009.html
2 / 60 : 1. GLM? 2. A: (pwer functin) x y?
3 R OS free sftware S R (2007) The R-Tips (2005) Statistics: An Intrductin Using R M. Crawley (2005) / 60
4 1. GLM?
5 : : ( ) / 60
6 / 60 : : (linear predictr): ( ) = ( ) + ( 1) ( 1) + ( 2) ( 2) + ( 3) ( 3) + ( )
7 / 60 (GLMM) GLM + randm effects ( ) (GLM) +
8 GLM / 60
9 / 60 :?! ( )? :? :?
10 / 60! GLM
11 / 60 (Pissn distributin)? lambda = 1.4 y i {0, 1, 2,, } (paramter: λ) λ y exp( λ) 0.0 lambda = y! prbability λ λ lambda = : y
12 (generalized linear mdel; GLM) link : link f () = f ( ) : β 0 + β 1 x 1 + β 2 + x i β i x i (cefficient) ({x i } {y i }) {β i } GLM / 60
13 R : glm() ( ) rbinm() glm(family = binmial) rbinm() glm(family = binmial) rpis() glm(family = pissn) rnbinm() glm.nb() in library(mass) ( ) rgamma() glm(family = gamma) rnrm() glm(family = gaussian) glm() GLM GLM / 60
14 / 60 R glm() :? ( z):? link : z (y)? family:?
15 glm() (1) family: pissn, (0, 1, 2, ) link : "lg" family = pissn link ( z): y ~ x family = pissn(link = "lg")? / 60
16 glm() (2) family: pissn, link : "lg" ( z): y ~ x z = a + bx a, b λ lg(λ) = z λ = exp(z) = exp(a + bx) λ : y Pis(λ) / 60
17 / A: (pwer functin)
18 A:? plant weight (g) number f flwers x y? 100 : i = 1, 2,, 100 x i y i / 60
19 : x y? x (0, 0) :?? / 60
20 1. y i λ i : y i Pis(λ i ) 2. λ i x i : λ i = Ax b i λ i = Ax b i λ i = exp(lg(a) + b lg(x i )) a = lg(a) lg(λ i ) = a + b lg(x i ) / 60
21 GLM! family: pissn, link : "lg" : y ~ lg.x x lg.x z = a + b lg.x a, b λ lg(λ) = z λ = exp(z) = exp(a + b lg.x) λ : y Pis(λ) / 60
22 / 60 R data.frame d = > lad("d.rdata") > head(d) # 6 x y lg.x > d$lg.x <- lg(d$x) > head(d) x y lg.x
23 glm() / 60
24 / 60 R glm() > fit <- glm(y ~ lg.x, data = d, family = pissn) > print(summary(fit)) Call: glm(frmula = y ~ lg.x, family = pissn, data = d) (......) Cefficients: Estimate Std. Errr z value Pr(> z ) (Intercept) lg.x (......) Cefficients
25 GLM plant weight (g) number f flwers plant weight (g) number f flwers / 60
26 : 1. : 2. GLM :, link, 3. R glm() deviance / 60
27 / GLM :, link,, deviance (20 ) 2. GLM : (60 ) 3. GLM ( ): ffset (40 )
28 / 60 : 1. : 2. ffset :
29 / 60 1.
30 / 60 : ( )
31 : ( ) / 60
32 / x x ( )
33 ? ( ) / 60
34 / 60 : : (linear predictr): ( ) = ( ) + ( 1) ( 1) + ( 2) ( 2) + ( 3) ( 3) + ( )
35 / 60
36 / 60! GLM
37 (generalized linear mdel; GLM) link : link f () = f ( ) : β 0 + β 1 x 1 + β 2 + x i β i x i (cefficient) ({x i } {y i }) {β i } GLM / 60
38 / 60 i N i k i i p i = k i /N i j p j = k j /N j i j p
39 ? / : ? ( ) / 60
40 / 60 : specific leaf area (SLA) : ffset : N k :
41 / ffset
42 B:? x {0.1, 0.2,, 1.0} 10 glm(..., family = pissn) / 60
43 ?!! x A = /! glm() ffset / 60
44 / 60 R data.frame: Area, x, y > lad("d2.rdata") > head(d, 8) # 8 Area x y
45 / 60 vs plt(d$x, d$y / d$area) d$y/d$area d$x?
46 / 60 A vs y plt(d$area, d$y) d$y d$area A y
47 / 60 x ( ) plt(d$area, d$y, cex = d$x * 2) d$y d$area?
48 / 60 x y x!
49 = 1. i y i λ i : y i Pis(λ i ) 2. λ i A i x i λ i = A i exp(a + bx i ) λ i = exp(a + bx i + lg(a i )) lg(λ i ) = a + bx i + lg(a i ) lg(a i ) ffset / 60
50 GLM! ffset family: pissn, link : "lg" : y ~ x ffset : lg(area) z = a + b x + lg(area) a, b λ lg(λ) = z λ = exp(z) = exp(a + b x + lg(area)) λ : y Pis(λ) / 60
51 glm() / 60
52 / 60 R glm() > fit <- glm(y ~ x, family = pissn(link = "lg"), data = d, ffset = lg(area)) > print(summary(fit)) Call: glm(frmula = y ~ x, family = pissn(link = "lg"), data = d, ffset = lg(area)) (......) Cefficients: Estimate Std. Errr z value Pr(> z ) (Intercept) x e-06 Cefficients
53 / 60 d$y d$area x = 0.9, x = 0.1 glm()
54 / 60 : glm() ffset ffset = exp( ) d$y d$area
55 / 60 : ffset
56 GLM summary(glm(...)) Wald Pr deviance AIC (verdispersin) GLMM kub/ce/ / 60
57 / 60 GLM (1)? y = 0, 1, 2, 3, (y ) (family = pissn) y = {0, 1}, y = {0, 1, 2,, N} (family = binmial) (family = Gamma) (family = gaussian)
58 R : glm() ( ) rbinm() glm(family = binmial) rbinm() glm(family = binmial) rpis() glm(family = pissn) rnbinm() glm.nb() in library(mass) ( ) rgamma() glm(family = gamma) rnrm() glm(family = gaussian) glm() glm.nb() MASS library GLM / 60
59 GLM (2)!! GLMM! GLMM/ randm effects GLM / 60
60 kub/ce/ / 60
/ 55 2 : : (GLM) 1. 1/23 ( )? GLM? (GLM ) 2.! 1/25 ( ) ffset (GLM )
2012 01 25 1/ 55 ( II) : (2012 1 ) 2 2 (GLM) 2012 01 25! kub@ees.hkudai.ac.jp http://g.gl/76c4i 2012 01 25 2/ 55 2 : : (GLM) 1. 1/23 ( )? GLM? (GLM ) 2.! 1/25 ( ) ffset (GLM ) 2012 01 25 3/ 55 1. : 2.
More information一般化線形 (混合) モデル (2) - ロジスティック回帰と GLMM
.. ( ) (2) GLMM kubo@ees.hokudai.ac.jp I http://goo.gl/rrhzey 2013 08 27 : 2013 08 27 08:29 kubostat2013ou2 (http://goo.gl/rrhzey) ( ) (2) 2013 08 27 1 / 74 I.1 N k.2 binomial distribution logit link function.3.4!
More information講義のーと : データ解析のための統計モデリング. 第3回
Title 講義のーと : データ解析のための統計モデリング Author(s) 久保, 拓弥 Issue Date 2008 Doc URL http://hdl.handle.net/2115/49477 Type learningobject Note この講義資料は, 著者のホームページ http://hosho.ees.hokudai.ac.jp/~kub ードできます Note(URL)http://hosho.ees.hokudai.ac.jp/~kubo/ce/EesLecture20
More information60 (W30)? 1. ( ) 2. ( ) web site URL ( :41 ) 1/ 77
60 (W30)? 1. ( ) kubo@ees.hokudai.ac.jp 2. ( ) web site URL http://goo.gl/e1cja!! 2013 03 07 (2013 03 07 17 :41 ) 1/ 77 ! : :? 2013 03 07 (2013 03 07 17 :41 ) 2/ 77 2013 03 07 (2013 03 07 17 :41 ) 3/ 77!!
More informationkubostat2017e p.1 I 2017 (e) GLM logistic regression : : :02 1 N y count data or
kubostat207e p. I 207 (e) GLM kubo@ees.hokudai.ac.jp https://goo.gl/z9ycjy 207 4 207 6:02 N y 2 binomial distribution logit link function 3 4! offset kubostat207e (https://goo.gl/z9ycjy) 207 (e) 207 4
More informationkubostat2015e p.2 how to specify Poisson regression model, a GLM GLM how to specify model, a GLM GLM logistic probability distribution Poisson distrib
kubostat2015e p.1 I 2015 (e) GLM kubo@ees.hokudai.ac.jp http://goo.gl/76c4i 2015 07 22 2015 07 21 16:26 kubostat2015e (http://goo.gl/76c4i) 2015 (e) 2015 07 22 1 / 42 1 N k 2 binomial distribution logit
More information(2/24) : 1. R R R
R? http://hosho.ees.hokudai.ac.jp/ kubo/ce/2004/ : kubo@ees.hokudai.ac.jp (2/24) : 1. R 2. 3. R R (3/24)? 1. ( ) 2. ( I ) : (p ) : cf. (power) p? (4/24) p ( ) I p ( ) I? ( ) (5/24)? 0 2 4 6 8 A B A B (control)
More informationkubostat2017c p (c) Poisson regression, a generalized linear model (GLM) : :
kubostat2017c p.1 2017 (c), a generalized linear model (GLM) : kubo@ees.hokudai.ac.jp http://goo.gl/76c4i 2017 11 14 : 2017 11 07 15:43 kubostat2017c (http://goo.gl/76c4i) 2017 (c) 2017 11 14 1 / 47 agenda
More informationk2 ( :35 ) ( k2) (GLM) web web 1 :
2012 11 01 k2 (2012-10-26 16:35 ) 1 6 2 (2012 11 01 k2) (GLM) kubo@ees.hokudai.ac.jp web http://goo.gl/wijx2 web http://goo.gl/ufq2 1 : 2 2 4 3 7 4 9 5 : 11 5.1................... 13 6 14 6.1......................
More information講義のーと : データ解析のための統計モデリング. 第5回
Title 講義のーと : データ解析のための統計モデリング Author(s) 久保, 拓弥 Issue Date 2008 Doc URL http://hdl.handle.net/2115/49477 Type learningobject Note この講義資料は, 著者のホームページ http://hosho.ees.hokudai.ac.jp/~kub ードできます Note(URL)http://hosho.ees.hokudai.ac.jp/~kubo/ce/EesLecture20
More information1 15 R Part : website:
1 15 R Part 4 2017 7 24 4 : website: email: http://www3.u-toyama.ac.jp/kkarato/ kkarato@eco.u-toyama.ac.jp 1 2 2 3 2.1............................... 3 2.2 2................................. 4 2.3................................
More information12/1 ( ) GLM, R MCMC, WinBUGS 12/2 ( ) WinBUGS WinBUGS 12/2 ( ) : 12/3 ( ) :? ( :51 ) 2/ 71
2010-12-02 (2010 12 02 10 :51 ) 1/ 71 GCOE 2010-12-02 WinBUGS kubo@ees.hokudai.ac.jp http://goo.gl/bukrb 12/1 ( ) GLM, R MCMC, WinBUGS 12/2 ( ) WinBUGS WinBUGS 12/2 ( ) : 12/3 ( ) :? 2010-12-02 (2010 12
More informationkubostat7f p GLM! logistic regression as usual? N? GLM GLM doesn t work! GLM!! probabilit distribution binomial distribution : : β + β x i link functi
kubostat7f p statistaical models appeared in the class 7 (f) kubo@eeshokudaiacjp https://googl/z9cjy 7 : 7 : The development of linear models Hierarchical Baesian Model Be more flexible Generalized Linear
More informationUse R
Use R! 2008/05/23( ) Index Introduction (GLM) ( ) R. Introduction R,, PLS,,, etc. 2. Correlation coefficient (Pearson s product moment correlation) r = Sxy Sxx Syy :, Sxy, Sxx= X, Syy Y 1.96 95% R cor(x,
More informationkubostat2017j p.2 CSV CSV (!) d2.csv d2.csv,, 286,0,A 85,0,B 378,1,A 148,1,B ( :27 ) 10/ 51 kubostat2017j (http://goo.gl/76c4i
kubostat2017j p.1 2017 (j) Categorical Data Analsis kubo@ees.hokudai.ac.jp http://goo.gl/76c4i 2017 11 15 : 2017 11 08 17:11 kubostat2017j (http://goo.gl/76c4i) 2017 (j) 2017 11 15 1 / 63 A B C D E F G
More informationk3 ( :07 ) 2 (A) k = 1 (B) k = 7 y x x 1 (k2)?? x y (A) GLM (k
2012 11 01 k3 (2012-10-24 14:07 ) 1 6 3 (2012 11 01 k3) kubo@ees.hokudai.ac.jp web http://goo.gl/wijx2 web http://goo.gl/ufq2 1 3 2 : 4 3 AIC 6 4 7 5 8 6 : 9 7 11 8 12 8.1 (1)........ 13 8.2 (2) χ 2....................
More information,, Poisson 3 3. t t y,, y n Nµ, σ 2 y i µ + ɛ i ɛ i N0, σ 2 E[y i ] µ * i y i x i y i α + βx i + ɛ i ɛ i N0, σ 2, α, β *3 y i E[y i ] α + βx i
Armitage.? SAS.2 µ, µ 2, µ 3 a, a 2, a 3 a µ + a 2 µ 2 + a 3 µ 3 µ, µ 2, µ 3 µ, µ 2, µ 3 log a, a 2, a 3 a µ + a 2 µ 2 + a 3 µ 3 µ, µ 2, µ 3 * 2 2. y t y y y Poisson y * ,, Poisson 3 3. t t y,, y n Nµ,
More information> > <., vs. > x 2 x y = ax 2 + bx + c y = 0 2 ax 2 + bx + c = 0 y = 0 x ( x ) y = ax 2 + bx + c D = b 2 4ac (1) D > 0 x (2) D = 0 x (3
13 2 13.0 2 ( ) ( ) 2 13.1 ( ) ax 2 + bx + c > 0 ( a, b, c ) ( ) 275 > > 2 2 13.3 x 2 x y = ax 2 + bx + c y = 0 2 ax 2 + bx + c = 0 y = 0 x ( x ) y = ax 2 + bx + c D = b 2 4ac (1) D >
More informationkubostat2018d p.2 :? bod size x and fertilization f change seed number? : a statistical model for this example? i response variable seed number : { i
kubostat2018d p.1 I 2018 (d) model selection and kubo@ees.hokudai.ac.jp http://goo.gl/76c4i 2018 06 25 : 2018 06 21 17:45 1 2 3 4 :? AIC : deviance model selection misunderstanding kubostat2018d (http://goo.gl/76c4i)
More informationkubostat2017b p.1 agenda I 2017 (b) probability distribution and maximum likelihood estimation :
kubostat2017b p.1 agenda I 2017 (b) probabilit distribution and maimum likelihood estimation kubo@ees.hokudai.ac.jp http://goo.gl/76c4i 2017 11 14 : 2017 11 07 15:43 1 : 2 3? 4 kubostat2017b (http://goo.gl/76c4i)
More information2 / 39
W707 s-taiji@is.titech.ac.jp 1 / 39 2 / 39 1 2 3 3 / 39 q f (x; α) = α j B j (x). j=1 min α R n+2 n ( d (Y i f (X i ; α)) 2 2 ) 2 f (x; α) + λ dx 2 dx. i=1 f B j 4 / 39 : q f (x) = α j B j (x). j=1 : x
More information講義のーと : データ解析のための統計モデリング. 第2回
Title 講義のーと : データ解析のための統計モデリング Author(s) 久保, 拓弥 Issue Date 2008 Doc URL http://hdl.handle.net/2115/49477 Type learningobject Note この講義資料は, 著者のホームページ http://hosho.ees.hokudai.ac.jp/~kub ードできます Note(URL)http://hosho.ees.hokudai.ac.jp/~kubo/ce/EesLecture20
More informationkubo2017sep16a p.1 ( 1 ) : : :55 kubo ( ( 1 ) / 10
kubo2017sep16a p.1 ( 1 ) kubo@ees.hokudai.ac.jp 2017 09 16 : http://goo.gl/8je5wh : 2017 09 13 16:55 kubo (http://goo.gl/ufq2) ( 1 ) 2017 09 16 1 / 106 kubo (http://goo.gl/ufq2) ( 1 ) 2017 09 16 2 / 106
More informationH22 BioS t (i) treat1 treat2 data d1; input patno treat1 treat2; cards; ; run; 1 (i) treat = 1 treat =
H BioS t (i) treat treat data d; input patno treat treat; cards; 3 8 7 4 8 8 5 5 6 3 ; run; (i) treat treat data d; input group patno period treat y; label group patno period ; cards; 3 8 3 7 4 8 4 8 5
More informationMicrosoft PowerPoint - GLMMexample_ver pptx
Linear Mixed Model ( 以下 混合モデル ) の短い解説 この解説のPDFは http://www.lowtem.hokudai.ac.jp/plantecol/akihiro/sumida-index.html の お勉強 のページにあります. ver 20121121 と との間に次のような関係が見つかったとしよう 全体的な傾向に対する回帰直線を点線で示した ところが これらのデータは実は異なる
More information6 2 2 x y x y t P P = P t P = I P P P ( ) ( ) ,, ( ) ( ) cos θ sin θ cos θ sin θ, sin θ cos θ sin θ cos θ y x θ x θ P
6 x x 6.1 t P P = P t P = I P P P 1 0 1 0,, 0 1 0 1 cos θ sin θ cos θ sin θ, sin θ cos θ sin θ cos θ x θ x θ P x P x, P ) = t P x)p ) = t x t P P ) = t x = x, ) 6.1) x = Figure 6.1 Px = x, P=, θ = θ P
More information1 環境統計学ぷらす 第 5 回 一般 ( 化 ) 線形混合モデル 高木俊 2013/11/21
1 環境統計学ぷらす 第 5 回 一般 ( 化 ) 線形混合モデル 高木俊 shun.takagi@sci.toho-u.ac.jp 2013/11/21 2 予定 第 1 回 : Rの基礎と仮説検定 第 2 回 : 分散分析と回帰 第 3 回 : 一般線形モデル 交互作用 第 4.1 回 : 一般化線形モデル 第 4.2 回 : モデル選択 (11/29?) 第 5 回 : 一般化線形混合モデル
More information今回 次回の要点 あぶない 時系列データ解析は やめましょう! 統計モデル のあてはめ Danger!! (危 1) 時系列データの GLM あてはめ (危 2) 時系列Yt 時系列 Xt 各時刻の個体数 気温 とか これは次回)
生態学の時系列データ解析でよく見る あぶない モデリング 久保拓弥 mailto:kubo@ees.hokudai.ac.jp statistical model for time-series data 2017-07-03 kubostat2017 (h) 1/59 今回 次回の要点 あぶない 時系列データ解析は やめましょう! 統計モデル のあてはめ Danger!! (危 1) 時系列データの
More information1 Stata SEM LightStone 4 SEM 4.. Alan C. Acock, Discovering Structural Equation Modeling Using Stata, Revised Edition, Stata Press 3.
1 Stata SEM LightStone 4 SEM 4.. Alan C. Acock, 2013. Discovering Structural Equation Modeling Using Stata, Revised Edition, Stata Press 3. 2 4, 2. 1 2 2 Depress Conservative. 3., 3,. SES66 Alien67 Alien71,
More informationJA2008
A1 1 10 vs 3 2 1 3 2 0 3 2 10 2 0 0 2 1 0 3 A2 3 11 vs 0 4 4 0 0 0 0 0 3 6 0 1 4 x 11 A3 5 4 vs 5 6 5 1 0 0 3 0 4 6 0 0 1 0 4 5 A4 7 11 vs 2 8 8 2 0 0 0 0 2 7 2 7 0 2 x 11 A5 9 5 vs 3 10 9 4 0 1 0 0 5
More information(lm) lm AIC 2 / 1
W707 s-taiji@is.titech.ac.jp 1 / 1 (lm) lm AIC 2 / 1 : y = β 1 x 1 + β 2 x 2 + + β d x d + β d+1 + ϵ (ϵ N(0, σ 2 )) y R: x R d : β i (i = 1,..., d):, β d+1 : ( ) (d = 1) y = β 1 x 1 + β 2 + ϵ (d > 1) y
More information17 17 17 17 11 21 28 1 24 12 36 2,000 2 22 11 3.67 3.38 22 2.97 21 10 1.7 1.12 22 10 13 2.75 11 10 15 24 10 12 14 3 17 17 2006 4 17 10 24 12 17 5 15 17 17 11 40 6 17 40 17 11 7 24 17 24 17 8 40 17 17 9
More informationPowerPoint Presentation
2 9/ 3 3 9/ 9 4 5 , PR () 6 ,,, (11) 7 PR 8 9 10 11 TEL. 106 8/131512/291/3 TEL. 107 12/291/3 12 http://www.f-turn.jp/ 13 21 4 21 14 200910 U 200911 U 200911 20102 15 20102 PR 20103 20103 16 20103 20104
More informationuntitled
,337 37 35 0,349,09 35 55 988 3 0 0 3,387 7 90 0,369,46 5 57 5 0 90 38 8,369 3 4 5 6 7 8 9 0 3 4 5 6 7 8 9 0 3 4 5 6 8 9 30 3 3 5,400 7,00 9,000 0,800,600 4,400 6,00 8,000 9,800,600 3,400 5,00 7,000 8,800
More information,877 61,524 33, ,292, ,653 57,601 95,188 2,416 1,767,
02 02 02 180,771 07 02 01 1,377 07 02 02 1,051,703 07 02 05 220,099 07 03 01 926,597 08 02 04 1,877,566 08 04 02 2,973,603 08 05 03 672,950 10 06 03 778,433 10 06 04 735,789 10 06 06 225,392 10 06 07 365,442
More informationkubostat2018a p.1 統計モデリング入門 2018 (a) The main language of this class is 生物多様性学特論 Japanese Sorry An overview: Statistical Modeling 観測されたパターンを説明する統計モデル
p.1 統計モデリング入門 2018 (a) The main language of this class is 生物多様性学特論 Japanese Sorry An overview: Statistical Modeling 観測されたパターンを説明する統計モデル 久保拓弥 (北海道大 環境科学) Why in Japanese? because even in Japanese, statistics
More information統計モデリング入門 2018 (a) 生物多様性学特論 An overview: Statistical Modeling 観測されたパターンを説明する統計モデル 久保拓弥 (北海道大 環境科学) 統計モデリング入門 2018a 1
統計モデリング入門 2018 (a) 生物多様性学特論 An overview: Statistical Modeling 観測されたパターンを説明する統計モデル 久保拓弥 (北海道大 環境科学) kubo@ees.hokudai.ac.jp 1/56 The main language of this class is Japanese Sorry Why in Japanese? because
More information2 H23 BioS (i) data d1; input group patno t sex censor; cards;
H BioS (i) data d1; input group patno t sex censor; cards; 0 1 0 0 0 0 1 0 1 1 0 4 4 0 1 0 5 5 1 1 0 6 5 1 1 0 7 10 1 0 0 8 15 0 1 0 9 15 0 1 0 10 4 1 0 0 11 4 1 0 1 1 5 1 0 1 1 7 0 1 1 14 8 1 0 1 15 8
More informationp.1/22
p.1/22 & & & & Excel / p.2/22 & & & & Excel / p.2/22 ( ) ( ) p.3/22 ( ) ( ) Baldi Web p.3/22 ( ) ( ) Baldi Web ( ) ( ) ( p.3/22 ) Text Mining for Clementine True Teller Text Mining Studio Text Miner Trustia
More information漸化式のすべてのパターンを解説しましたー高校数学の達人・河見賢司のサイト
https://www.hmg-gen.com/tuusin.html https://www.hmg-gen.com/tuusin1.html 1 2 OK 3 4 {a n } (1) a 1 = 1, a n+1 a n = 2 (2) a 1 = 3, a n+1 a n = 2n a n a n+1 a n = ( ) a n+1 a n = ( ) a n+1 a n {a n } 1,
More informationGLM PROC GLM y = Xβ + ε y X β ε ε σ 2 E[ε] = 0 var[ε] = σ 2 I σ 2 0 σ 2 =... 0 σ 2 σ 2 I ε σ 2 y E[y] =Xβ var[y] =σ 2 I PROC GLM
PROC MIXED ( ) An Introdunction to PROC MIXED Junji Kishimoto SAS Institute Japan / Keio Univ. SFC / Univ. of Tokyo e-mail address: jpnjak@jpn.sas.com PROC MIXED PROC GLM PROC MIXED,,,, 1 1.1 PROC MIXED
More informationMicrosoft Word - 触ってみよう、Maximaに2.doc
i i e! ( x +1) 2 3 ( 2x + 3)! ( x + 1) 3 ( a + b) 5 2 2 2 2! 3! 5! 7 2 x! 3x! 1 = 0 ",! " >!!! # 2x + 4y = 30 "! x + y = 12 sin x lim x!0 x x n! # $ & 1 lim 1 + ('% " n 1 1 lim lim x!+0 x x"!0 x log x
More information( 28 ) ( ) ( ) 0 This note is c 2016, 2017 by Setsuo Taniguchi. It may be used for personal or classroom purposes, but not for commercial purp
( 28) ( ) ( 28 9 22 ) 0 This ote is c 2016, 2017 by Setsuo Taiguchi. It may be used for persoal or classroom purposes, but ot for commercial purposes. i (http://www.stat.go.jp/teacher/c2epi1.htm ) = statistics
More informationDAA09
> summary(dat.lm1) Call: lm(formula = sales ~ price, data = dat) Residuals: Min 1Q Median 3Q Max -55.719-19.270 4.212 16.143 73.454 Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) 237.1326
More informationnakata/nakata.html p.1/20
http://www.me.titech.ac.jp/ nakata/nakata.html p.1/20 1-(a). Faybusovich(1997) Linear systems in Jordan algebras and primal-dual interior-point algorithms,, Euclid Jordan p.2/20 Euclid Jordan V Euclid
More information: (GLMM) (pseudo replication) ( ) ( ) & Markov Chain Monte Carlo (MCMC)? /30
PlotNet 6 ( ) 2006-01-19 TOEF(1998 2004), AM, growth6 DBH growth (mm) 1998 1999 2000 2001 2002 2003 2004 10 20 30 40 50 70 DBH (cm) 1. 2. - - : kubo@ees.hokudai.ac.jp http://hosho.ees.hokudai.ac.jp/ kubo/show/2006/plotnet/
More informationH22 BioS (i) I treat1 II treat2 data d1; input group patno treat1 treat2; cards; ; run; I
H BioS (i) I treat II treat data d; input group patno treat treat; cards; 8 7 4 8 8 5 5 6 ; run; I II sum data d; set d; sum treat + treat; run; sum proc gplot data d; plot sum * group ; symbol c black
More information²¾ÁÛ¾õ¶·É¾²ÁË¡¤Î¤¿¤á¤Î¥Ñ¥Ã¥±¡¼¥¸DCchoice ¡Ê»ÃÄêÈÇ¡Ë
DCchoice ( ) R 2013 2013 11 30 DCchoice package R 2013/11/30 1 / 19 1 (CV) CV 2 DCchoice WTP 3 DCchoice package R 2013/11/30 2 / 19 (Contingent Valuation; CV) WTP CV WTP WTP 1 1989 2 DCchoice package R
More information°ÌÁê¿ô³ØII
July 14, 2007 Brouwer f f(x) = x x f(z) = 0 2 f : S 2 R 2 f(x) = f( x) x S 2 3 3 2 - - - 1. X x X U(x) U(x) x U = {U(x) x X} X 1. U(x) A U(x) x 2. A U(x), A B B U(x) 3. A, B U(x) A B U(x) 4. A U(x),
More information1 Tokyo Daily Rainfall (mm) Days (mm)
( ) r-taka@maritime.kobe-u.ac.jp 1 Tokyo Daily Rainfall (mm) 0 100 200 300 0 10000 20000 30000 40000 50000 Days (mm) 1876 1 1 2013 12 31 Tokyo, 1876 Daily Rainfall (mm) 0 50 100 150 0 100 200 300 Tokyo,
More information/22 R MCMC R R MCMC? 3. Gibbs sampler : kubo/
2006-12-09 1/22 R MCMC R 1. 2. R MCMC? 3. Gibbs sampler : kubo@ees.hokudai.ac.jp http://hosho.ees.hokudai.ac.jp/ kubo/ 2006-12-09 2/22 : ( ) : : ( ) : (?) community ( ) 2006-12-09 3/22 :? 1. ( ) 2. ( )
More information18 ( ) I II III A B C(100 ) 1, 2, 3, 5 I II A B (100 ) 1, 2, 3 I II A B (80 ) 6 8 I II III A B C(80 ) 1 n (1 + x) n (1) n C 1 + n C
8 ( ) 8 5 4 I II III A B C( ),,, 5 I II A B ( ),, I II A B (8 ) 6 8 I II III A B C(8 ) n ( + x) n () n C + n C + + n C n = 7 n () 7 9 C : y = x x A(, 6) () A C () C P AP Q () () () 4 A(,, ) B(,, ) C(,,
More informationii 3.,. 4. F. (), ,,. 8.,. 1. (75%) (25%) =7 20, =7 21 (. ). 1.,, (). 3.,. 1. ().,.,.,.,.,. () (12 )., (), 0. 2., 1., 0,.
24(2012) (1 C106) 4 11 (2 C206) 4 12 http://www.math.is.tohoku.ac.jp/~obata,.,,,.. 1. 2. 3. 4. 5. 6. 7.,,. 1., 2007 (). 2. P. G. Hoel, 1995. 3... 1... 2.,,. ii 3.,. 4. F. (),.. 5... 6.. 7.,,. 8.,. 1. (75%)
More information第13回:交差項を含む回帰・弾力性の推定
13 2018 7 27 1 / 31 1. 2. 2 / 31 y i = β 0 + β X x i + β Z z i + β XZ x i z i + u i, E(u i x i, z i ) = 0, E(u i u j x i, z i ) = 0 (i j), V(u i x i, z i ) = σ 2, i = 1, 2,, n x i z i 1 3 / 31 y i = β
More information2 p T, Q
270 C, 6000 C, 2 p T, Q p: : p = N/ m 2 N/ m 2 Pa : pdv p S F Q 1 g 1 1 g 1 14.5 C 15.5 1 1 cal = 4.1855 J du = Q pdv U ( ) Q pdv 2 : z = f(x, y). z = f(x, y) (x 0, y 0 ) y y = y 0 z = f(x, y 0 ) x x =
More information10:30 12:00 P.G. vs vs vs 2
1 10:30 12:00 P.G. vs vs vs 2 LOGIT PROBIT TOBIT mean median mode CV 3 4 5 0.5 1000 6 45 7 P(A B) = P(A) + P(B) - P(A B) P(B A)=P(A B)/P(A) P(A B)=P(B A) P(A) P(A B) P(A) P(B A) P(B) P(A B) P(A) P(B) P(B
More informationn ( (
1 2 27 6 1 1 m-mat@mathscihiroshima-uacjp 2 http://wwwmathscihiroshima-uacjp/~m-mat/teach/teachhtml 2 1 3 11 3 111 3 112 4 113 n 4 114 5 115 5 12 7 121 7 122 9 123 11 124 11 125 12 126 2 2 13 127 15 128
More informationECCS. ECCS,. ( 2. Mac Do-file Editor. Mac Do-file Editor Windows Do-file Editor Top Do-file e
1 1 2015 4 6 1. ECCS. ECCS,. (https://ras.ecc.u-tokyo.ac.jp/guacamole/) 2. Mac Do-file Editor. Mac Do-file Editor Windows Do-file Editor Top Do-file editor, Do View Do-file Editor Execute(do). 3. Mac System
More information0201
2018 10 17 2019 9 19 SI J cal 1mL 1ºC 1999 cal nutrition facts label calories cal kcal 1 cal = 4.184 J heat capacity 1 K 1 J K 1 mol molar heat capacity J K mol (specific heat specific heat capacity) 1
More informationuntitled
2011/6/22 M2 1*1+2*2 79 2F Y YY 0.0 0.2 0.4 0.6 0.8 0.000 0.002 0.004 0.006 0.008 0.010 0.012 1.0 1.5 2.0 2.5 3.0 3.5 4.0 Y 0 50 100 150 200 250 YY A (Y = X + e A ) B (YY = X + e B ) X 0.00 0.05 0.10
More informationx x x 2, A 4 2 Ax.4 A A A A λ λ 4 λ 2 A λe λ λ2 5λ + 6 0,...λ 2, λ 2 3 E 0 E 0 p p Ap λp λ 2 p 4 2 p p 2 p { 4p 2 2p p + 2 p, p 2 λ {
K E N Z OU 2008 8. 4x 2x 2 2 2 x + x 2. x 2 2x 2, 2 2 d 2 x 2 2.2 2 3x 2... d 2 x 2 5 + 6x 0 2 2 d 2 x 2 + P t + P 2tx Qx x x, x 2 2 2 x 2 P 2 tx P tx 2 + Qx x, x 2. d x 4 2 x 2 x x 2.3 x x x 2, A 4 2
More information第85 回日本感染症学会総会学術集会後抄録(III)
β β α α α µ µ µ µ α α α α γ αβ α γ α α γ α γ µ µ β β β β β β β β β µ β α µ µ µ β β µ µ µ µ µ µ γ γ γ γ γ γ µ α β γ β β µ µ µ µ µ β β µ β β µ α β β µ µµ β µ µ µ µ µ µ λ µ µ β µ µ µ µ µ µ µ µ
More informationx () g(x) = f(t) dt f(x), F (x) 3x () g(x) g (x) f(x), F (x) (3) h(x) = x 3x tf(t) dt.9 = {(x, y) ; x, y, x + y } f(x, y) = xy( x y). h (x) f(x), F (x
[ ] IC. f(x) = e x () f(x) f (x) () lim f(x) lim f(x) x + x (3) lim f(x) lim f(x) x + x (4) y = f(x) ( ) ( s46). < a < () a () lim a log xdx a log xdx ( ) n (3) lim log k log n n n k=.3 z = log(x + y ),
More information( ) FAS87 FAS FAS87 v = 1 i 1 + i
( ) ( 7 6 ) ( ) 1 6 1 18 FAS87 FAS87 7 1 FAS87 v = 1 i 1 + i 10 14 6 6-1 - 7 73 2 N (m) N L m a N (m) L m a N m a (m) N 73 9 99 18 4-2 - 4 143 2 145 3 37 4 37 4 40 6 40 6 41 10 41 10 13 10 14 4 24 3 145
More information1 4 1 ( ) ( ) ( ) ( ) () 1 4 2
7 1995, 2017 7 21 1 2 2 3 3 4 4 6 (1).................................... 6 (2)..................................... 6 (3) t................. 9 5 11 (1)......................................... 11 (2)
More information