簡易檢索 / 詳目顯示

研究生: 黃彥翔
Yen-Shiung Huang
論文名稱: 應用在擴增實境之植基於邊緣檢測的即時視覺追蹤技術
Real-Time Visual Tracking Techniques Based on Edge Detection Applied for Augmented Reality
指導教授: 范欽雄
Chin-Shyurng Fahn
口試委員: 邱舉明
Ge-Ming Chiu
莊仁輝
Jen-Hui Chuang
王榮華
Jung-Hua Wang
學位類別: 碩士
Master
系所名稱: 電資學院 - 資訊工程系
Department of Computer Science and Information Engineering
論文出版年: 2009
畢業學年度: 97
語文別: 英文
論文頁數: 68
中文關鍵詞: 擴增實境視覺追蹤邊緣檢測標籤辨識即時處理
外文關鍵詞: augmented reality, visual tracking, edge detection, marker recognition, real-time processing
相關次數: 點閱:225下載:4
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  •   在本篇論文中,我們提出了應用在擴增實境的即時視覺追蹤技術,其視覺感測元件為一般的USB攝影機,用以捕捉來源影像;而被追蹤的物體為含有特定圖案且其尺寸已知的自製標籤。我們所提出的方法使電腦能感知各個標籤是否存在於攝影機的視野之中、該標籤具有何種身分,以及得知該標籤的姿態與位置。有了這些資訊,電腦即可將各個數位三維物件合成於指定的標籤上,如此,使用者就可以在與標籤互動的同時,直接與虛擬物件互動。
      在本論文中所描述的追蹤方法包含四邊形偵測、標籤辨識,以及三維的姿態與位置估算,其中四邊形偵測的方法係植基於邊緣檢測,亦即在來源影像中,我們利用邊緣及其骨架與方向來搜尋四邊形的頂點,與植基於二值化的方法相比,我們的方法所適用的環境明暗變化範圍較大,接著,我們利用所搜尋出的頂點,將該四邊形正規化,以進行標籤辨識,並利用這些頂點來估算在真實世界中,標籤與攝影機之間的空間關係;為了將辨識錯誤降到最低,我們標籤辨識方法中的正規化處理程序,除了將四邊形的形狀正規化之外,也對其顏色進行正規化,於此,我們所設計的每張標籤都含有一個16位元的編碼,這提供了一個可觀數量的標籤供使用者應用。另外,在已知攝影機的垂直視角與焦距長度的情況下,應用我們的姿態與位置估算方法之前,不需要進行攝影機校正。
      根據實驗的結果,在使用一個時脈頻率為2.4 GHz的單核心處理器,且同時追蹤5張標籤時,一張320 × 240影像的平均處理時間為9.4毫秒,相當於每秒可處理104.6張影像;而在追蹤相同數量的標籤時,一張640 × 480影像的平均處理時間為32.7毫秒,相當於每秒可處理30.6張影像,這顯示我們提出的方法能夠同時對於多個物件進行即時追蹤。


    In this thesis, we present real-time visual tracking techniques applied for augmented reality. The visual sensor employed is a common USB camera for capturing source images, and the objects tracked are the artificial square markers with particular patterns and given sizes. Our proposed method enables computers to sense each marker whether exists in the sight of the camera, which identity the marker has, and what pose and position it is. With this information, computers can superimpose virtual objects on the specific markers respectively, so that users can interact with the virtual objects simultaneously by interacting with the markers.
    The tracking method described in this thesis includes the quadrangle detection, marker recognition, and three-dimensional pose and position estimation. Our quadrangle detection method is based on the edge detection, and we use the edges with their skeletons and directions to search the vertices of quadrangles in the source image. In comparison with a threshold based method, our method is able to work in a wider range of the environmental lightness. The vertices are used to normalize the corresponding quadrangle for marker recognition and to estimate the spatial relations between the marker and the camera in the real world. The normalization process in our marker recognition method is not only to normalize the shape of the quadrangle but also to normalize the color of it for minimizing the errors of the marker identification. Each of our designed markers contains a 16-bit code which gives a good variety of identities to them. Our pose and position estimation method is camera-calibration free as long as the vertical angle of view and the focal length of the camera are given.
    According to the experimental results, the average processing time of a 320 × 240 image for tracking 5 markers is 9.4 milliseconds corresponding to 104.6 FPS, and it takes 32.7 milliseconds corresponding to 30.6 FPS on an average to process a 640 × 480 image for tracking the same number of markers with a 2.4 GHz, single-core processor. It means that our tracking techniques can track multiple objects in real time.

    致謝 i 摘要 ii Abstract iii Contents v List of Figures vii List of Tables ix Chapter 1 Introduction 1 1.1 Overview 1 1.2 Background and motivation 2 1.3 Our proposed method 4 1.4 Thesis organization 6 Chapter 2 Related Work 8 2.1 Reviews of edge detection 8 2.2 Reviews of thinning 10 2.3 Reviews of geometric transformation 12 Chapter 3 Quadrangle Detection 15 3.1 Edge detection 15 3.2 Thinning 18 3.3 Quadrangle searching 21 Chapter 4 Marker Recognition 27 4.1 Quadrangle normalization 27 4.2 Grayscale normalization 29 4.3 Error reduction 32 4.4 Marker identification 34 Chapter 5 Pose and Position Estimation 37 5.1 Estimation of the rotation matrix 37 5.2 Translation estimation 40 Chapter 6 Experimental Results and Discussion 43 6.1 Tracking results in different angles and distances 43 6.2 Tracking results in different lightness of environment 47 6.3 Timing results 50 Chapter 7 Conclusion and Future Works 52 7.1 Conclusion 52 7.2 Future works 52 References 54

    [1] F. Zhou, H. B. L. Duh, and M. Billinghurst, “Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR,” in proceedings of the 7th IEEE/ACM International Symposium on Mixed and Augmented Reality, Cambridge, UK, pp. 193 - 202, 2008.
    [2] J. P. Rolland, L. Davis, and Y. Baillot, A survey of tracking technology for virtual environments, In Fundamentals of Wearable Computers and Augmented Reality, 1st ed., W. Barfield and T. Caudell, Eds. Mahwah, NJ: CRC, pp. 67-112, 2001.
    [3] J. Newman, D. Ingram, and A. Hopper, “Augmented reality in a wide area sentient environment,” in Proceedings of IEEE/ACM International Symposium on Augmented Reality, New York, USA, pp. 77 - 86, 2001.
    [4] G. Klinker, R. Reicher, and B. Brugge, “Distributed user tracking concepts for augmented reality applications,” in Proceedings of IEEE/ACM International Symposium on Augmented Reality, Munich, Germany, pp. 37 - 44, 2000.
    [5] J. Newman, M. Wagner, M. Bauer, A. MacWilliams, T. Pintaric, D. Beyer, D. Pustka, F. Strasser, D. Schmalstieg, and G. Klinker, “Ubiquitous tracking for augmented reality,” in Proceedings of the 3rd IEEE/ACM International Symposium on Mixed and Augmented Reality, Arlington, USA, pp. 192 - 201, 2004.
    [6] H. Kato and M. Billinghurst, “Marker Tracking and HMD Calibration for a Video-based Augmented Reality Conferencing System,” in Proceedings of the 2nd IEEE/ACM International Workshop on Augmented Reality, San Francisco, USA, pp. 85 - 94, 1999.
    [7] D. Reiners, D. Stricker, G. Klinker, and S. Muller, “Augmented reality for construction tasks: Doorlock assembly,” in Proceedings of the 1st International Workshop on Augmented Reality, San Francisco, USA, pp. 31 - 46, 1998.
    [8] I. Poupyrev, D. Tan, M. Billinghurst, H. Kato, H. Regenbrecht, and N. Tetsutani, “Developing a generic augmented reality interface,” in Computer, vol. 35, no. 3, pp. 44 - 50, March 2002.
    [9] X. Zhang, N. Navab, and S. Liou, “E-commerce direct marketing using augmented reality,” in Proceedings of IEEE International Conference on Multimedia and Expo, New York, USA, pp. 88 - 91, 2000.
    [10] J. Park, S. You, and U. Neumann, “Natural feature tracking for extendible robust augmented realities,” in Proceedings of the 1st International Workshop on Augmented Reality, San Francisco, USA, 1998.
    [11] A. Comport, E. Marchand, and F. Chaumette, “A real-time tracker for markerless augmented reality,” in Proceedings of the 2nd IEEE/ACM International Symposium on Mixed and Augmented Reality, Tokyo, Japan, pp. 36 - 45, 2003.
    [12] G. Reitmayr and T. Drummond, “Going out: robust model-based tracking for outdoor augmented reality,” in Proceedings of IEEE/ACM International Symposium on Mixed and Augmented Reality, Santa Barbara, USA, pp. 109 - 118, 2006.
    [13] L. Vacchetti, V. Lepetit, and P. Fua, “Combining edge and texture information for real-time accurate 3D camera tracking,” in Proceedings of 3rd IEEE/ACM International Symposium on Mixed and Augmented Reality, Arlington, USA, pp. 48 - 56, 2004.
    [14] ARToolKit. http://www.hitl.washington.edu/artoolkit/ referenced at July 7, 2009.
    [15] I. Sobel and G. Feldman, “A 3 × 3 Isotropic Gradient Operator for Image Processing,” presented at a talk at the Stanford Artificial Project, 1968.
    [16] R. Sukthankar, R. G. Stockton, M. D. Mullin, “Automatic Keystone Correction for camera-assisted Presentation interface,” in Proceedings of International Conference on Multimodal Interfaces, Beijing, China, pp. 601 - 714, 2000.
    [17] L. G. Roberts, “Machine Perception of Three-Dimensional Solids,” in Optical and Electro-Optical Information Processing, Tippet, J. T. (ed.), MIT Press, Cambridge, Mass, 1965.
    [18] J. M. S. Prewitt, “Object Enhancement and Extraction” in Picture Processing and Psychopictorics, Lipkin, B. S., and Rosenfeld, A. (eds.), Academic Press, New York, 1970.
    [19] J. Canny, “A Computational Approach to Edge Detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 8, no. 6, pp. 679 - 698, 1986.
    [20] R. C. Gonzalez and R. E. Woods, Digital Image Processing Second Edition. New Jersey: Prentice Hall, 2002.

    QR CODE