Top Banner
Speaker : Meng-Shun Su Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Adviser : Chih-Hung Lin Ten-Chuan Hsi Ten-Chuan Hsia Date : 2010/01/26 Date : 2010/01/26 ©2010 STUT. ©2010 STUT. CSIE. CSIE. Multimedia and Information Multimedia and Information Security Lab. J205-1 Security Lab. J205-1
26

Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

Dec 30, 2015

Download

Documents

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

Speaker : Meng-Shun SuSpeaker : Meng-Shun SuAdviser : Chih-Hung LinAdviser : Chih-Hung Lin Ten-Chuan HsiaoTen-Chuan HsiaoDate : 2010/01/26Date : 2010/01/26

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Page 2: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

OutlineOutline

11 IntroductionIntroductionIntroductionIntroduction

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

22 Related WorkRelated WorkRelated WorkRelated Work

44 Experimental ResultsExperimental ResultsExperimental ResultsExperimental Results

55 ConclusionsConclusionsConclusionsConclusions

2©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Page 3: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

3

1. Elaborate tracking systems,2. Input devices and computer vision techniques   improve the registration of images3. User interaction in AR setups.

In recent years, augmented reality (AR) research has focused on various subareas of interest.

11 IntroductionIntroductionIntroductionIntroduction

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Page 4: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

4

‧Real-time process complex visual computations.

In mobile PC of necessity condition

Example :     A cell phone with a camera and a wide screen can serve as an inexpensive AR device for the general public. Emerging wearable computers are also excellent platforms for enabling AR anywhere users may go.

11 IntroductionIntroductionIntroductionIntroduction

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Page 5: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

5

Paper present and evaluate new techniques for markerless computer-vision-based tracking in arbitrary physical tabletop environments.

Three-dimensional coordinate systems   can be established on a tabletop surface at will, using a  user’s outstretched hand as a temporary initialization  pattern. Building upon an existing framework for fingertip and hand pose tracking

11 IntroductionIntroductionIntroductionIntroduction

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Page 6: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

We want to lower the barrier to initiating AR systems, so that users can easily experience AR anywhere, Considering a mobile user entering a new workplace, such as an office desk environment. Several marker-based AR systems have shown uccessful registration of virtual objects on top of cardboard arkers. In 1999 year H. Kato et.al succeed shown virtual object.

6

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Fig 1. (a) MIS LAB. LOGO image. (b) STUT-CSIE String. (c) 7-ELEVEN Mascot image.

(a) (b) (c)

Page 7: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

Interactive “tabletop” systems. Hand touch a virtual object, In 1999 -2005 years of research, has made great contribution. For fast and accurate tracking, optical-flow-based algorithms are widely used. Algorithms in intel OpenCV library.

7

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Fig 2. Used in the automotive mechanical specifications.

Page 8: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

‧Handy AR Using the “Handy AR” approach, we estimate a six-degree-of-freedom camera pose. The hand is segmented by a skin-color-based classifier with an adaptively learned skin color histogram.

8

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Fig 3. Generated to detect the hand coordinate system.

Page 9: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

While the hand is tracked over time, fingertips are located on the contour of the hand by finding points that have high curvature values.

9

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Fig 4. Snapshots of Handy AR: (a) the hand’s coordinate system, (b) selecting and inspecting world-stabilized augmented objects, and (c), (d) inspecting a virtual object from various angles.

Page 10: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

‧Invariant Feature Detection In the image formation process, several unknown variables play a role in varying the image properties, viewing angle, illumination, lens distortion, and so forth. In these wide-baseline cases, the feature matching problem requires the features to maintain invariant properties for large differences in viewing angles and camera translation. We apply the scale invariant feature transform (SIFT) to extract keypoints from a captured video frame.

10

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Page 11: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

11

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Difference-of-Gaussian (DoG) in scale space are detected as keypoints,characterized accurately by four components:

, the pixel location, , the scale, , the major orientation,

),( yx

Page 12: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

12

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

4 x 4 array of eightbin histograms results in 4 x 4 x 8 = 128 dimensions. The keypoints selected by the SIFT algorithm are shown to be invariant to scale and orientation changes, and the descriptors remain robust to illumination changes. From here on, we will refer to these keypoints together with their descriptors as SIFT features. SIFT features are shown to be useful both for drift-free tracking and for initially recognizing a previously stored scene.

Page 13: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

13

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Optical-Flow-Based Tracking (OFBT)Given two images at time and at time with a fixed size window, we compute the spatial derivative.

Image(t) Image’(t+1)

I I’

t dtt

)(xI t )(' xI dtt

Fig 5. OFBT Consciousness Chart.

Page 14: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

14

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

IIIIIIxyx

yxxxG 2

2

)(

IIII

txb tx

ty

),(

),()(),( 1 txbxGtxu

: 2x2 symmetric matrix is invertible)(xG

This gives the criteria for Shi and Tomasi’s Good Features to Track algorithm.

Page 15: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

15

22 Related WorkRelated WorkRelated WorkRelated Work

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Fig 6. Procedural flowchart and conceptual time sequence for each tracked entity, establishing a coordinate system for a tabletop AR environment, using Handy AR and our hybrid feature tracking method.

Page 16: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

16

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Initializing a Coordinate System Using the Handy AR system, we enable a user to initialize a coordinate system for AR using the simple gesture of putting the hand on the working space surface.

1. User places a hand on a desktop surface. ‧ Handy AR estimates the coordinate system. ‧ SIFT features are detected in the camera frame. ‧ The coordinate system is propagated to the surface.

Page 17: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

17

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Fig 7. Establishing a coordinate system using Handy AR (a), propagating it to the scene as we detect scene features (b). After moving the hand out of the scene (c), new features are detected from the area that the hand occluded (d).

2. User removes the hand.‧SIFT features are detected in the camera frame, matching previously stored features and providing new features in the space previously occupied by the hand.

Page 18: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

18

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Computing 3D Feature Locations

Planar Structures:

Fig 8. Planar Structures

Page 19: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

19

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Estimated camera pose from Handy AR, the 3D locations of the features are computed in a new world coordinate system according to the projection model:

11

43 Z

Y

X

y

x

P 1yx

1ZYX

This homogeneous representations of the SIFT features in the image plane and the coordinate system

43P This projection matrix of the camera

Page 20: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

20

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Nonplanar Structures 3D locations of the feature points can be computed using two or more views of the camera. or linear method . For example, suppose that a feature point is tracked between two images with the projection model:

XPx 11 XPx 22

and3

21, IR This image measurements of the feature point in homogeneous

representation.

21,PP This 3 x 4 projection matrices of the two views.

X This construct a linear system for the unknown variable.

Page 21: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

21

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Hybrid Feature Tracking

Fig 9. Illustration of tracking events in frame sequences for our hybrid feature tracking mechanism.

Page 22: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

22

33 Method DescriptionMethod DescriptionMethod DescriptionMethod Description

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Hybrid Feature Tracking

Fig 10. Diagram of dependency among threads and data. Since the SIFT detection thread and the optical flow tracking thread critically share the feature data, synchronization is required.

Page 23: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

23

44 Experimental ResultsExperimental ResultsExperimental ResultsExperimental Results

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Regard to the speed of the system and robustness of the hybrid feature tracking algorithm. evaluations is described in Fig 11.

Fig 11. Flowchart of the AR application task sequence, as used in performance experiments.

Page 24: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

24

44 Experimental ResultsExperimental ResultsExperimental ResultsExperimental Results

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

1. Initial stage of the program, the user establishes the coordinate system either by hand gesture using Handy AR or by the system recognizing a stored scene.

2. The tracking region is expanded while the system performs hybrid feature tracking.

3. When the region is expanded enough for user interactions, the user can select an object.

4. The user can place an object.

Page 25: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

25

55 ConclusionsConclusionsConclusionsConclusions

©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11

Our markerless tracking is initialized by a simple hand gesture using the Handy AR system, which estimates a camera pose from a user’s outstretched hand.The established coordinate system is propagated to the scene, and then the tabletop workspace for AR is continuously expanded to cover an area far beyond a single reference frame.

Page 26: Speaker : Meng-Shun Su Adviser : Chih-Hung Lin Ten-Chuan Hsiao Ten-Chuan Hsiao Date : 2010/01/26 ©2010 STUT. CSIE. Multimedia and Information Security.

26©2010 STUT.©2010 STUT. CSIE.CSIE. Multimedia and Information Security Lab. J205-Multimedia and Information Security Lab. J205-11