簡易檢索 / 詳目顯示

研究生: 廖崇儒
Chung-ju Liao
論文名稱: 可供輪椅使用者使用之基於電腦視覺的手勢辨識系統
Vision-based hand gesture recognition system for users on wheelchairs
指導教授: 蘇順豐
Shun-Feng Su
口試委員: 郭重顯
Chung-Hsien Kuo
陶金旺
Chin-Wang Tao
鄭錦聰
Jin-Tsong Jeng
學位類別: 碩士
Master
系所名稱: 電資學院 - 電機工程系
Department of Electrical Engineering
論文出版年: 2010
畢業學年度: 98
語文別: 英文
論文頁數: 86
中文關鍵詞: 手勢辨識離群值二階段膚色偵測
外文關鍵詞: hand gesture recognition, outliers, two stages of skin color detection
相關次數: 點閱:412下載:4
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 本文提出了一個可供輪椅使用者使用的手勢辨識系統,大多數的手勢辨識相關文獻只探討如何提升手勢辨識的準確度,但是在這篇論文中除了手勢辨識之外,還有許多的問題需要解決,例如偵測手什麼時候進入影片當中或者手掌的區域是不是有超出影片的周圍。本文提出的系統之流程可以分為四個步驟,首先要先判斷手何時出現在影片中,接下來是如何把手從背景中擷取出來,然後判斷手掌是否超出影片的四周圍,最後則是手勢辨識。在擷取手部區域方面,我們從統計學的離群值(outlier)觀念做延伸應用到膚色偵測上,不同於大多數的手勢辨識系統只對膚色做了單一階段的偵測,我們做了兩階段的膚色偵測,藉由此方法去除跟手相連的膚色雜訊。至於判斷手掌是否超出影片的四周圍的階段,我們利用人體工學的概念來判斷手掌是否超出影片範圍外。從實驗結果可發現本文提出的系統除了背景的膚色雜訊過大或者手跟攝影機的距離太過接近無法克服之外,整體而言是相當成功的,如果把本文所提出的系統搭配硬體或者晶片加以實現,將有助於提升輪椅使用者單獨在家時候的便利性。


    In this study, a hand gesture recognition system is proposed for users on wheelchairs. Unlike other studies that focus on the stage of hand gesture recognition, many problems are considered, such as detection of when a hand reaches in the field of the camera view or detection of a full palm. There are 4 stages in the proposed system, detection of the appearance of hands, segmentation of hand regions, detection of full palm and hand gesture recognition. Detection of the appearance of hands is to find out when a hand appears in the front of the camera. The hand region is then obtained in the next stage. In this stage, some morphological techniques, along with 2 stage skin color detection, are used to alleviate the effect of noise. The 2 stage skin color detection approach is adopted from the idea of handling outliers to extract the palm from complicated backgrounds. Following that, detection of full palm is conducted to know whether the hand reaches beyond the field of the camera view. The concept of ergonomics is employed to determine whether the hand is beyond the field of the camera view. Finally, hand gesture recognition is performed. Experimental results show that the proposed system is quite promising except that a very small number of frames are misjudged because the system cannot deal with some problems such as the area of the noise being too large or the hand being too close to the camera.

    摘要 I Abstract II Table of Contents III List of Figures IV List of Tables VI Chapter 1 Introduction 1 1.1 Research Motivation and Objective 1 1.2 Environment 3 1.3 Thesis Organization 4 Chapter 2 Literature Review 5 2.1 Hand Gesture Recognition Techniques 5 2.1.1 Glove-Based Techniques 6 2.1.2 Vision-Based Techniques 7 2.2 Skin Color Detection Techniques 10 2.2.1 RGB Color Space 12 2.2.2 Normalized RGB Color Space 14 2.2.3 HSV Color Space 16 2.2.4 YCbCr Color Space 17 Chapter 3 System Description 20 3.1 Detection of the Appearance of Hands 20 3.2 Segmentation of Hand Regions 25 3.2.1 Image Binarization 27 3.2.2 Morphological Techniques 28 3.2.3 Component Labeling and Scan Line 31 3.2.4 2nd Stage of Skin Color Detection 35 3.2.5 Region Filling 37 3.2.6 Opening and Component Labeling 39 3.3 Detection of Full Palm 40 3.4 Hand Gesture Recognition 46 Chapter 4 Experimental Results 55 4.1 Experimental results in simple backgrounds 58 4.2 Experimental results in complicated backgrounds 69 Chapter 5 Conclusions and Future Work 82 References 84

    [1] 內政部統計處, http://www.moi.gov.tw/stat/
    [2] M. Saito, “Expanding Welfare Concept and Assistive Technology”,
    Proceedings of the 2nd Int. Cognitive Technology Conference, pp. 271-278,
    1997.
    [3] V.I. Pavlovic, R. Sharma, and T.S. Huang, “Visual Interpretation of Hand
    Gestures for Human-Computer Interaction: A Review,” IEEE Transactions on
    Pattern Analysis and Machine Intelligence, pp. 677-695, 1997.
    [4] W. Ying and T.S. Huang, “Capturing Articulated Human Hand Motion: A
    Divide-and-Conquer Approach,” The Proceedings of the Seventh IEEE
    International Conference on Computer Vision, pp. 606-611, 1999.
    [5] J. Lee and T. L. Kunii, “Model-Based Analysis of Hand Posture,” IEEE
    Computer Graphics and Applications, pp. 77-86, 1995.
    [6] K. Kwon, H. Zhang., and F. Dornaika, “Hand Pose Recovery with a Single
    Video Camera,” IEEE International Conference on Robotics and Automation,
    pp. 1194-1200, 2001.
    [7] C. Kervrann and F. Heitz, “Learning Structure and Deformation Modes of
    Non rigid Objects in Long Image Sequences,” International Workshop on
    Automatic Face- and Gesture-Recognition, 1995.
    [8] K. Cho and S.M. Dunn, “Learning Shape Classes,” IEEE Transactions on
    Pattern Analysis and Machine Intelligence, pp. 882-888, 1994.
    [9] R. Cipolla and N.J. Hollinghurst, “Human-Robot Interface by Pointing With
    Uncalibrated Stereo Vision,” Image and Vision Computing 14, pp. 171-178,
    1996.
    [10] T.E. Starner and A. Pentland, “Visual Recognition of American Sign
    Language Using Hidden Markov Models,” Proc. Int’l Workshop on
    Automatic Face and Gesture Recognition, pp. 189-194, 1995 .
    [11] B. Bauer and H. Hienz, “Relevant Features for Video-Based Continuous
    Sign Language Recognition,“ Proc. of Fourth IEEE International
    Conference on Automatic Face and Gesture Recognition, pp. 440-445, 2000.

    [12] K. Grobel and M. Assam, “Isolated Sign Language Recognition Using Hidden
    Markov Models,” Proceedings of the IEEE International Conference on
    Computational Cybernetics and Simulation, pp. 162-167, 1997
    [13] Y. Hamada, N. Shimada, and Y. Shirai, ”Hand Shape Estimation under
    Complex Backgrounds for Sign Language Recognition,” Proceedings of 6th
    Int. Conf. on Automatic and Gesture Recognition, pp. 589-594, 2004.
    [14] J. Issacs and J.S. Foo, “Hand Pose Estimation for American Sign Language
    Recognition,” Proc. of the Thirty-Sixth Southeastern Symposium on System
    Theory, pp. 132-136, 2004.
    [15] D.S. Zhang and G. Lu, “Review of Shape Representation and Description
    Techniques,” Pattern Recognition, pp. 1-19, 2004.
    [16] E.J. Holden and R. Owens, “Recognizing Moving Hand Shapes,” Proceedings
    of the 12th International Conference on Image Analysis and Processing,
    pp. 14-19, 2003
    [17] C.G. Rafael and E.W. Richard, “Digital Image Processing 2/e,” Prentice
    Hall, 2002.
    [18] K.Kristensen, P. Nilsson, and V. Owall, “Background Segmentation beyond
    RGB,” 7th Asian Conference on Computer Vision. Pp. 602-612, 2006.
    [19] E. Osuna, F. Cirosi, and R. Freund, “Training Support Vector Machines:
    an Application to Face Detection,” Proceedings of CVPR’97, pp. 241-246,
    1997.
    [20] L. Bretzner, I. Laptev, and T. Lindeberg, “Hand Gesture Recognition
    using Multi-Scale Colour Features, Hierarchical Models and Particle
    Filtering,” Proceedings on Automatic Face and Gesture Recognition, pp.
    423-428, 2002.
    [21] J. Yang, W. Lu, and W. Alex, “Skin-Color Modeling and Adaptation,”
    Computer Vision – ACCV’98, pp.687-694, 1997
    [22] M.H. Yang and N. Ahuja, “Gaussian Mixture Model for Human Skin Color and
    Its Application in Image and Video Databases,” Conference on Storage and
    Retrieval for Image and Video Databases (SPIE 99), pp. 458-466, 1999.
    [23] J. Li, B. Zhao, H. Zhang, and J. Jiao, “Dual-Space Skin-Color Cue Based
    Face Detection for Eye Location,” International Conference on
    Information Engineering and Computer Science, pp. 1-4, 2009.
    [24] M.H. Yang, D.J. Kriegman, and N. Ahuja, “Detecting Faces in Images: a
    Survey,“ IEEE Trans. Pattern Anal. Mach. Intell., pp. 34-58, 2002.
    [25] 王貞元, “火災偵測和區域定位應用於影像監控”, 台灣科技大學電機工程學系碩士
    論文, 2008.
    [26] 曹文潔, “猜拳機”, 中央大學電機工程研究所碩士論文, 2007.
    [27] Barnett, V. Lewis, “Outliers in Statistical Data 3/e,” John Wiley &
    Sons, 1994.
    [28] WolframMathworld, http://mathworld.wolfram.com/
    [29] 黃俊捷, “互動雙足式機器人之設計與實現(I)手勢辨識“, 中央大學電機工程研究
    所碩士論文, 2008.
    [30] 陳培森, “看護系統之監控與辨識功能建立“, 中央大學電機工程研究所碩士論文,
    2006.
    [31] 涂又仁, “利用人臉及手勢辨識之人機介面“, 中正大學電機工程研究所碩士論文,
    2007.
    [32] D. Bowman, C. Wingrave, J. Campbell, and V. Ly, “Using Pinch Gloves for
    Both Natural and Abstract Interaction Techniques in Virtual
    Environments,” HCI international, 2001.
    [33] L. Gan, K.S. Lik, G. Hall, and U. Zeshan, “Dynamic Hand Gesture Tracking
    and Recognition for Real-Time Immersive Virtual Object Manipulation, ”
    CW ’09 International Conference on CyberWorlds, pp. 29-35, 2009.
    [34] L. Dipietro, A.M. Sabatini, and P. Dario, “A Survey of Glove-Based
    Systems and Their Applications,” IEEE Transactions on SMC, pp. 461-482, 2008.
    [35] C.Y. Tsai, and Y.H. Lee, “Multiple-Camera-Based Gesture Recognition by
    MDA Method,” 5th International Conference on Fuzzy Systems and Knowledge
    Discovery, pp. 599-603, 2008.
    [36] R. Gopalan, and B. Dariush, “Toward a Vision-Based Hand Gesture
    Interface for Robotic Grasping,” IEEE International Conference on
    Intelligent Robots and Systems, pp.1452-1459, 2009.
    [37] Y. Sato, M. Saito, and H. Koike, “Real Time Input of 3D Pose and
    Gestures of A User’s Hand and Its Application for HCI,” Virtual Reality
    Annual International Symposium, pp. 79-86, 2001.
    [38] S. Wan, and H.T. Nguyen, “Human Computer Interaction using Hand
    Gesture,” Annual International Conference of the IEEE Engineering in
    Medicine and Biology Society, pp. 2357-2360, 2008.
    [39] D.J. Sturman, and D. Zeltzer, “ A Survey of Glove-Based Input,”
    Computer Graphics and Applications, IEEE, pp. 30-39, 1994.
    [40] X.M. Yin, and M. Xie, “Hand Gesture Segmentation, Recognition and
    Application,” Computational Intelligence in Robotics and Automation
    IEEE, pp. 438-443, 2001.
    [41] C. Qing, N.D. Georganas, and E.M. Petriu, “Real-Time Vision-Based Hand
    Gesture Recognition using Haar-Like Features,” IEEE Instrumentation and
    Measurement Technology Conference, pp. 1-6, 2007.
    [42] A. Malima, E. Ozgur, and M. Cetin, “A Fast Algorithm for Vision-Based
    Hand Gesture Recognition for Robot Control,” IEEE 14th Signal Processing
    and Communication Applications Conference, pp. 1-4, 2006.
    [43] K. Sobottka, and I. Pitas, “A Novel Method for Automatic Face
    Segmentation, Facial Feature Extraction and Tracking,” Signal
    Processing: Image Communication, pp. 263-281, 1998.
    [44] M. Soriano, B. Martinkauppi, S. Huovinen, and M. Laaksonen, “Using the
    Skin Locus to Cope with Changing Illumination Conditions in Color-Based
    Face Tracking,” IEEE Nordic Signal Processing Symposium, pp. 393-386,
    2000.
    [45] X. Yishen, G. Jihua, T. Zhi, and W. Di, “Bare Hand Gesture
    Recognition,” International Congress on Image and Signal Processing,
    2009.
    [46] H.I. Park, and J.W. Lee, “Hand Gesture Recognition for Table-Top
    Interaction System,” International Symposium on Ubiquitous VR, 2007.
    [47] Y.K. Chen, T.Y. Cheng, and S.T. Chiu, “Motion Detection with Entropy in
    Dynamic Background,” International Asia Conference on Informatics in
    Control, Automation, and Robotics, pp. 263-266, 2009.

    QR CODE