3: 2: 2. 2 Semi-supervised learning Semi-supervised learning [5,6] Semi-supervised learning Self-training [13] [14] Self-training Self-training Semi-s

Similar documents
Duplicate Near Duplicate Intact Partial Copy Original Image Near Partial Copy Near Partial Copy with a background (a) (b) 2 1 [6] SIFT SIFT SIF

IPSJ SIG Technical Report Vol.2012-CG-149 No.13 Vol.2012-CVIM-184 No /12/4 3 1,a) ( ) DB 3D DB 2D,,,, PnP(Perspective n-point), Ransa

2014/3 Vol. J97 D No. 3 Recognition-based segmentation [7] 1 DP 1 Conditional random field; CRF [8] [10] CRF / OCR OCR [11], [1

THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS TECHNICAL REPORT OF IEICE. TRECVID2012 Instance Search {sak

3 2 2 (1) (2) (3) (4) 4 4 AdaBoost 2. [11] Onishi&Yoda [8] Iwashita&Stoica [5] 4 [3] 3. 3 (1) (2) (3)

IPSJ SIG Technical Report Vol.2013-CVIM-187 No /5/30 1,a) 1,b), 1,,,,,,, (DNN),,,, 2 (CNN),, 1.,,,,,,,,,,,,,,,,,, [1], [6], [7], [12], [13]., [

4. C i k = 2 k-means C 1 i, C 2 i 5. C i x i p [ f(θ i ; x) = (2π) p 2 Vi 1 2 exp (x µ ] i) t V 1 i (x µ i ) 2 BIC BIC = 2 log L( ˆθ i ; x i C i ) + q

一般社団法人電子情報通信学会 THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGIN

(MIRU2008) HOG Histograms of Oriented Gradients (HOG)

(MIRU2010) Geometric Context Randomized Trees Geometric Context Rand

(a) (b) (c) Canny (d) 1 ( x α, y α ) 3 (x α, y α ) (a) A 2 + B 2 + C 2 + D 2 + E 2 + F 2 = 1 (3) u ξ α u (A, B, C, D, E, F ) (4) ξ α (x 2 α, 2x α y α,

IPSJ SIG Technical Report Vol.2009-CVIM-167 No /6/10 Real AdaBoost HOG 1 1 1, 2 1 Real AdaBoost HOG HOG Real AdaBoost HOG A Method for Reducing

28 Horizontal angle correction using straight line detection in an equirectangular image

IPSJ SIG Technical Report GPS LAN GPS LAN GPS LAN Location Identification by sphere image and hybrid sensing Takayuki Katahira, 1 Yoshio Iwai 1

本文6(599) (Page 601)

IPSJ SIG Technical Report Vol.2010-CVIM-170 No /1/ Visual Recognition of Wire Harnesses for Automated Wiring Masaki Yoneda, 1 Ta

スライド 1

Microsoft PowerPoint - SSII_harada pptx

2007/8 Vol. J90 D No. 8 Stauffer [7] 2 2 I 1 I 2 2 (I 1(x),I 2(x)) 2 [13] I 2 = CI 1 (C >0) (I 1,I 2) (I 1,I 2) Field Monitoring Server

Google Goggles [1] Google Goggles Android iphone web Google Goggles Lee [2] Lee iphone () [3] [4] [5] [6] [7] [8] [9] [10] :

,,.,.,,.,.,.,.,,.,..,,,, i

SICE東北支部研究集会資料(2017年)

2 Fig D human model. 1 Fig. 1 The flow of proposed method )9)10) 2.2 3)4)7) 5)11)12)13)14) TOF 1 3 TOF 3 2 c 2011 Information

(a) 1 (b) 3. Gilbert Pernicka[2] Treibitz Schechner[3] Narasimhan [4] Kim [5] Nayar [6] [7][8][9] 2. X X X [10] [11] L L t L s L = L t + L s

1., 1 COOKPAD 2, Web.,,,,,,.,, [1]., 5.,, [2].,,.,.,, 5, [3].,,,.,, [4], 33,.,,.,,.. 2.,, 3.., 4., 5., ,. 1.,,., 2.,. 1,,


IPSJ SIG Technical Report iphone iphone,,., OpenGl ES 2.0 GLSL(OpenGL Shading Language), iphone GPGPU(General-Purpose Computing on Graphics Proc

1 Fig. 1 Extraction of motion,.,,, 4,,, 3., 1, 2. 2.,. CHLAC,. 2.1,. (256 ).,., CHLAC. CHLAC, HLAC. 2.3 (HLAC ) r,.,. HLAC. N. 2 HLAC Fig. 2

1 Kinect for Windows M = [X Y Z] T M = [X Y Z ] T f (u,v) w 3.2 [11] [7] u = f X +u Z 0 δ u (X,Y,Z ) (5) v = f Y Z +v 0 δ v (X,Y,Z ) (6) w = Z +

[1] SBS [2] SBS Random Forests[3] Random Forests ii

untitled

% 2 3 [1] Semantic Texton Forests STFs [1] ( ) STFs STFs ColorSelf-Simlarity CSS [2] ii

Silhouette on Image Object Silhouette on Images Object 1 Fig. 1 Visual cone Fig. 2 2 Volume intersection method Fig. 3 3 Background subtraction Fig. 4

Sobel Canny i

[2] OCR [3], [4] [5] [6] [4], [7] [8], [9] 1 [10] Fig. 1 Current arrangement and size of ruby. 2 Fig. 2 Typography combined with printing

gengo.dvi

IPSJ SIG Technical Report Vol.2010-MPS-77 No /3/5 VR SIFT Virtual View Generation in Hallway of Cybercity Buildings from Video Sequen

情報処理学会研究報告 IPSJ SIG Technical Report Vol.2013-CVIM-186 No /3/15 EMD 1,a) SIFT. SIFT Bag-of-keypoints. SIFT SIFT.. Earth Mover s Distance

LED a) A New LED Array Acquisition Method Focusing on Time-Gradient and Space- Gradient Values for Road to Vehicle Visible Light Communication Syunsuk

3 1 Table 1 1 Feature classification of frames included in a comic magazine Type A Type B Type C Others 81.5% 10.3% 5.0% 3.2% Fig. 1 A co

(a) (b) 2 2 (Bosch, IR Illuminator 850 nm, UFLED30-8BD) ( 7[m] 6[m]) 3 (PointGrey Research Inc.Grasshopper2 M/C) Hz (a) (b

11) 13) 11),12) 13) Y c Z c Image plane Y m iy O m Z m Marker coordinate system T, d X m f O c X c Camera coordinate system 1 Coordinates and problem

44 4 I (1) ( ) (10 15 ) ( 17 ) ( 3 1 ) (2)

生活設計レジメ

I II III 28 29


,,,,,,,,,,,,,,,,,,, 976%, i

2.2 6).,.,.,. Yang, 7).,,.,,. 2.3 SIFT SIFT (Scale-Invariant Feature Transform) 8).,. SIFT,,. SIFT, Mean-Shift 9)., SIFT,., SIFT,. 3.,.,,,,,.,,,., 1,

Convolutional Neural Network A Graduation Thesis of College of Engineering, Chubu University Investigation of feature extraction by Convolution

bag-of-words bag-of-keypoints Web bagof-keypoints Nearest Neighbor SVM Nearest Neighbor SIFT Nearest Neighbor bag-of-keypoints Nearest Neighbor SVM 84

28 TCG SURF Card recognition using SURF in TCG play video

IPSJ SIG Technical Report Vol.2015-MUS-107 No /5/23 HARK-Binaural Raspberry Pi 2 1,a) ( ) HARK 2 HARK-Binaural A/D Raspberry Pi 2 1.

yoo_graduation_thesis.dvi

(3.6 ) (4.6 ) 2. [3], [6], [12] [7] [2], [5], [11] [14] [9] [8] [10] (1) Voodoo 3 : 3 Voodoo[1] 3 ( 3D ) (2) : Voodoo 3D (3) : 3D (Welc

1 (PCA) 3 2 P.Viola 2) Viola AdaBoost 1 Viola OpenCV 3) Web OpenCV T.L.Berg PCA kpca LDA k-means 4) Berg 95% Berg Web k-means k-means

Abstract

P361

THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS TECHNICAL REPORT OF IEICE.

Fig. 1. Example of characters superimposed on delivery slip.

図 2: 高周波成分を用いた超解像 解像度度画像とそれらを低解像度化して得られる 低解像度画像との差により低解像度の高周波成分 を得る 高解像度と低解像度の高周波成分から位 置関係を保ったままパッチ領域をそれぞれ切り出 し 高解像度パッチ画像と低解像度パッチ画像の ペアとしてデータベースに登録する

情報処理学会研究報告 IPSJ SIG Technical Report Vol.2014-GN-90 No.6 Vol.2014-CDS-9 No.6 Vol.2014-DCC-6 No /1/23 Bullet Time 1,a) 1 Bullet Time Bullet Time

BDH Cao BDH BDH Cao Cao Cao BDH ()*$ +,-+.)*$!%&'$!"#$ 2. 1 Weng [4] Metric Learning Weng DB DB Yang [5] John [6] Sparse Coding sparse coding DB [7] K

Microsoft Word - toyoshima-deim2011.doc

IS1-09 第 回画像センシングシンポジウム, 横浜,14 年 6 月 2 Hough Forest Hough Forest[6] Random Forest( [5]) Random Forest Hough Forest Hough Forest 2.1 Hough Forest 1 2.2

列生成を困難にする要因となっている 既存研究では 与え られた画像からグレイスケールに変換し 画像 1 枚から抽出 を行っているため 外乱 ( 影や光 ) の影響を受けると文字列を 正しく抽出できない (Yin et al., 2014) さらに 情景内の単一 の文字は既存研究では考慮されていない

IPSJ SIG Technical Report Vol.2013-CVIM-188 No /9/2 1,a) D. Marr D. Marr 1. (feature-based) (area-based) (Dense Stereo Vision) van der Ma

1 Table 1: Identification by color of voxel Voxel Mode of expression Nothing Other 1 Orange 2 Blue 3 Yellow 4 SSL Humanoid SSL-Vision 3 3 [, 21] 8 325

kut-paper-template.dvi

IPSJ SIG Technical Report Vol.2010-CVIM-171 No /3/19 1. Web 1 1 Web Web Web Multiple Kernel Learning(MKL) Web ( ) % MKL 68.8% Extractin

IPSJ SIG Technical Report 1,a) 1,b) 1,c) 1,d) 2,e) 2,f) 2,g) 1. [1] [2] 2 [3] Osaka Prefecture University 1 1, Gakuencho, Naka, Sakai,

Optical Flow t t + δt 1 Motion Field 3 3 1) 2) 3) Lucas-Kanade 4) 1 t (x, y) I(x, y, t)

xx/xx Vol. Jxx A No. xx 1 Fig. 1 PAL(Panoramic Annular Lens) PAL(Panoramic Annular Lens) PAL (2) PAL PAL 2 PAL 3 2 PAL 1 PAL 3 PAL PAL 2. 1 PAL

VRSJ-SIG-MR_okada_79dce8c8.pdf

Web Basic Web SAS-2 Web SAS-2 i

研究成果報告書

2017 (413812)

2). 3) 4) 1.2 NICTNICT DCRA Dihedral Corner Reflector micro-arraysdcra DCRA DCRA DCRA 3D DCRA PC USB PC PC ON / OFF Velleman K8055 K8055 K8055

1. HNS [1] HNS HNS HNS [2] HNS [3] [4] [5] HNS 16ch SNR [6] 1 16ch 1 3 SNR [4] [5] 2. 2 HNS API HNS CS27-HNS [1] (SOA) [7] API Web 2

4 4 2 RAW (PCA) RAW RAW [5] 4 RAW 4 Park [12] Park 2 RAW RAW 2 RAW y = Mx + n. (1) y RAW x RGB M CFA n.. R G B σr 2, σ2 G, σ2 B D n ( )

IPSJ SIG Technical Report Vol.2014-MBL-70 No.49 Vol.2014-UBI-41 No /3/15 2,a) 2,b) 2,c) 2,d),e) WiFi WiFi WiFi 1. SNS GPS Twitter Facebook Twit

(4) ω t(x) = 1 ω min Ω ( (I C (y))) min 0 < ω < C A C = 1 (5) ω (5) t transmission map tmap 1 4(a) t 4(a) t tmap RGB 2 (a) RGB (A), (B), (C)

Fig. 2 Signal plane divided into cell of DWT Fig. 1 Schematic diagram for the monitoring system

B HNS 7)8) HNS ( ( ) 7)8) (SOA) HNS HNS 4) HNS ( ) ( ) 1 TV power, channel, volume power true( ON) false( OFF) boolean channel volume int

10_08.dvi

THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS TECHNICAL REPORT OF IEICE {s-kasihr, wakamiya,

光学

258 5) GPS 1 GPS 6) GPS DP 7) 8) 10) GPS GPS ) GPS Global Positioning System

Accuracy Improvement by Compound Discriminant Functions for Resembling Character Recognition Takashi NAKAJIMA, Tetsushi WAKABAYASHI, Fumitaka KIMURA,

(fnirs: Functional Near-Infrared Spectroscopy) [3] fnirs (oxyhb) Bulling [4] Kunze [5] [6] 2. 2 [7] [8] fnirs 3. 1 fnirs fnirs fnirs 1

LBP 2 LBP 2. 2 Local Binary Pattern Local Binary pattern(lbp) [6] R

2. 研 究 実 施 内 容 ( 文 中 に 番 号 がある 場 合 は(3-1)に 対 応 する) (1) 黄 瀬 グループ 局 所 特 徴 量 と 最 近 傍 探 索 を 用 いる 文 字 認 識 手 法 の 開 発 を 進 めた 特 に 今 年 度 は 高 速 化 に 注 力 した 具 体 的

2003/3 Vol. J86 D II No Fig. 1 An exterior view of eye scanner. CCD [7] CCD PC USB PC PC USB RS-232C PC

SICE東北支部研究集会資料(2013年)

THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS TECHNICAL REPORT OF IEICE. UWB UWB

27 AR

i


Wide Scanner TWAIN Source ユーザーズガイド

(a) (b) (c) Fig. 2 2 (a) ; (b) ; (c) (a)configuration of the proposed system; (b)processing flow of the system; (c)the system in use 1 GPGPU (

( )

Transcription:

THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS TECHNICAL REPORT OF IEICE. 599-8531 1-1 E-mail: tsukada@m.cs.osakafu-u.ac.jp, {masa,kise}@cs.osakafu-u.ac.jp Semi-supervised learning 2 2 Segmentation by RecognitionLabeling by RecognitionSemi-supervised Learning 1. [1, 2] 1 (i) (ii) (iii) (iv) NEOCR [3] Street View House Numbers Dataset [4] (ii) (iv) 1: (ii) (iii) (iv) Semi-supervised learning [5, 6] 2 2. Semi-supervised learning 2. 1 [7] [8 11] [12] [2] [2] [2] [2] 1 1 1

3: 2: 2. 2 Semi-supervised learning Semi-supervised learning [5,6] Semi-supervised learning Self-training [13] [14] Self-training Self-training Semi-supervised learning Self-training Nearest Neighbor Nearest Neighbor 3. 3. 1 2 (I) PCA-SIFT [15] Dense Sampling (II) [16] d nn d 2nn d nn d 2nn < t d (1) t d 3 (III)Reference Point(RP) [17] RP RP RP (IV) (i)rp d c RP (ii) RP n c (iii) RP x y (V) 4 2 x d x x d x d x d x x S x S x S x = S d x dx y 2

S dx dy S d y S dy S d x dx d x 4: d y 5: 3. 2 5 (I) (II) (III) (IV) (III) (V) (II)(IV) Reliability Check 3. 2. 1 Dense Sampling ID 3. 1 (1) 1/N i N i i ID 3. 2. 2 Reliability Check Reliability Check s 1 2 s 2 t l < s 2 s 1 < t u (2) t l t s 1 (1) 4. 1 3. 1 2 3. 2 3 4. 1 Street View House Numbers Dataset Street View House Numbers Dataset [4] Full Numbers Full Numbers 4. 2 1 train 3

70 60 # labeled data : 100 # labeled data : 500 # labeled data : 1000 Recall [%] 50 40 30 20 10 10 20 30 40 50 60 70 80 Precision [%] (a) 100 (b) 1000 6: 1Recall-Precision 7: 1 96 96[pixels] 200 100500 1,000 extra 1,000 t d d c n c 0.93035 RecallPrecision d c n c [18] 6 6 7 100 1,000 7 8 8 1 4. 3 2 2 3 1 3. 2 2 Reliability Check 3 10 10,000 train extra 8: 1 test t d t l t u 0.900.2 4. 3. 1 1 (2) a. b. (1) c. (1) d. b d a c b d (1) b d a c ab cd 4

1: 2-1 a. b. c. d. [%] 46.3 44.8 46.3 44.8 [%] 42.7 50.8 50.3 64.7 41,360 46,834 30,624 33,531 [%] 63.7 68.2 81.4 89.0 2: 3 [%] 77.5 [%] 78.8 4,037 [%] 73.0 4. 3. 2 Reliability Check 9(a) 1,000 10,000 1,000 (Ground Truth) Correct CheckCorrect Check (1) Reliability Check Ground Truth Correct Check Correct Check Reliability Check 9(b) 9(a) Reliability Check Recognition Rate [%] # of Retrained Data 100 90 80 70 60 50 Ground Truth Correct Check Proposed Method 40 0 2000 4000 6000 8000 10000 100000 90000 80000 70000 60000 50000 40000 30000 20000 10000 # of Unlabeled Data per Class (a) Ground Truth Correct Check Proposed Method 0 0 2000 4000 6000 8000 10000 # of Unlabeled Data per Class (b) 9: 2-2 4. 3. 3 10 5

Recognition Rate [%] 80 70 60 50 40 30 20 # of Labeled Data: 10 # of Labeled Data: 50 10 # of Labeled Data: 500 # of Labeled Data: 100 # of Labeled Data: 1000 0 0 100 500 1000 # of Labeled Data + Unlabeled Data per Class 10: 2-3 4. 4 3 3 2 Reliability Check Reliability Check (2) 2 2 train 1,000 extra 5,000 13,215 test t d d c n c t l t u 0.930550.30.7 2 1.3% 4,037 73% 2,947 2 Reliability Check Reliability Check 5. Reliabiilty Check JST CREST [1] M. Iwamura, T. Tsuji, and K. Kise, Memory-based recognition of camera-captured characters, Proc. DAS, 2010. [2] M. Iwamura, T. Kobayashi, and K. Kise, Recognition of multiple characters in a scene image using arrangement of local features, Proc. ICDAR, 2011. [3] A.D. Robert Nagy and K. Meyer-Wegener, NEOCR: A configurable dataset for natural image text recognition, Proc. CBDAR, 2011. [4] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A.Y. Ng, Reading digits in natural images with unsupervised feature learning, NIPS Workshop on Deep Learning and Unsupervised Feature Learning, 2011. [5] O. Chapelle, B. Schölkopf, and A. Zien eds., Semisupervised learning, Cambridge, MIT Press, Sept. 2006. [6] X. Zhu and A.B. Goldberg, Introduction to semi-supervised learning, Morgan and Claypool Publishers, Sept. 2009. [7] J.-J. Lee, P.-H. Lee, S.-W. Lee, A.L. Yuille, and C. Koch, Adaboost for text detection in natural scene, Proc. IC- DAR, 2011. [8] B. Epshtein, E. Ofek, and Y. Wexler, Detecting text in natural scenes with stroke width transform, Proc. CVPR, 2010. [9] P. Sanketi, H. Shen, and J.M. Coughlan, Localizing blurry and low-resolution text in natural images, Proc. IEEE Workshop on Applications of Computer Vision, 2011. [10] C. Yao, Z. Tu, and Y. Ma, Detecting texts of arbitrary orientations in natural images, Proc. CVPR, 2012. [11] L. Neumann and J. Matas, Real-time scene text localization and recognition, Proc. CVPR, 2012. [12] Y.-F. Pan, X. Hou, and C.-L. Liu, A hybrid approach to detect and localize texts in natural scene images, IEEE Trans. on Image Processing, vol.20, no.3, pp.800 813, March 2011. [13] M. Tsukada, M. Iwamura, and K. Kise, Expanding recognizable distorted characters using self-corrective recognition, Proc. DAS, 2012. [14] V. Frinken, M. Baumgartner, A. Fischer, and H. Bunke, Semi-supervised learning for cursive handwriting recognition using keyword spotting, Proc. ICFHR, 2012. [15] Y. Ke and R. Sukthankar, PCA-SIFT: a more distinctive representation for local image descriptors, Proc. CVPR, pp.506 513, 2004. [16] pp.73 78PRMU2012-142Feb. 2013 [17] M. Klinkigt and K. Kise, Using a reference point for local configuration of sift-like features for object recognition with serious background clutter, IPSJ Trans. on Computer Vision and Applications, vol.3, pp.110 121, Dec. 2011. [18] C. Wolf and J.-M. Jolion, Object count/area graphs for the evaluation of object detection and segmentation algorithms, IJDAR, vol.8, no.4, 2006. 6