研究生: |
廖崇儒 Chung-ju Liao |
---|---|
論文名稱: |
可供輪椅使用者使用之基於電腦視覺的手勢辨識系統 Vision-based hand gesture recognition system for users on wheelchairs |
指導教授: |
蘇順豐
Shun-Feng Su |
口試委員: |
郭重顯
Chung-Hsien Kuo 陶金旺 Chin-Wang Tao 鄭錦聰 Jin-Tsong Jeng |
學位類別: |
碩士 Master |
系所名稱: |
電資學院 - 電機工程系 Department of Electrical Engineering |
論文出版年: | 2010 |
畢業學年度: | 98 |
語文別: | 英文 |
論文頁數: | 86 |
中文關鍵詞: | 手勢辨識 、離群值 、二階段膚色偵測 |
外文關鍵詞: | hand gesture recognition, outliers, two stages of skin color detection |
相關次數: | 點閱:412 下載:4 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
本文提出了一個可供輪椅使用者使用的手勢辨識系統,大多數的手勢辨識相關文獻只探討如何提升手勢辨識的準確度,但是在這篇論文中除了手勢辨識之外,還有許多的問題需要解決,例如偵測手什麼時候進入影片當中或者手掌的區域是不是有超出影片的周圍。本文提出的系統之流程可以分為四個步驟,首先要先判斷手何時出現在影片中,接下來是如何把手從背景中擷取出來,然後判斷手掌是否超出影片的四周圍,最後則是手勢辨識。在擷取手部區域方面,我們從統計學的離群值(outlier)觀念做延伸應用到膚色偵測上,不同於大多數的手勢辨識系統只對膚色做了單一階段的偵測,我們做了兩階段的膚色偵測,藉由此方法去除跟手相連的膚色雜訊。至於判斷手掌是否超出影片的四周圍的階段,我們利用人體工學的概念來判斷手掌是否超出影片範圍外。從實驗結果可發現本文提出的系統除了背景的膚色雜訊過大或者手跟攝影機的距離太過接近無法克服之外,整體而言是相當成功的,如果把本文所提出的系統搭配硬體或者晶片加以實現,將有助於提升輪椅使用者單獨在家時候的便利性。
In this study, a hand gesture recognition system is proposed for users on wheelchairs. Unlike other studies that focus on the stage of hand gesture recognition, many problems are considered, such as detection of when a hand reaches in the field of the camera view or detection of a full palm. There are 4 stages in the proposed system, detection of the appearance of hands, segmentation of hand regions, detection of full palm and hand gesture recognition. Detection of the appearance of hands is to find out when a hand appears in the front of the camera. The hand region is then obtained in the next stage. In this stage, some morphological techniques, along with 2 stage skin color detection, are used to alleviate the effect of noise. The 2 stage skin color detection approach is adopted from the idea of handling outliers to extract the palm from complicated backgrounds. Following that, detection of full palm is conducted to know whether the hand reaches beyond the field of the camera view. The concept of ergonomics is employed to determine whether the hand is beyond the field of the camera view. Finally, hand gesture recognition is performed. Experimental results show that the proposed system is quite promising except that a very small number of frames are misjudged because the system cannot deal with some problems such as the area of the noise being too large or the hand being too close to the camera.
[1] 內政部統計處, http://www.moi.gov.tw/stat/
[2] M. Saito, “Expanding Welfare Concept and Assistive Technology”,
Proceedings of the 2nd Int. Cognitive Technology Conference, pp. 271-278,
1997.
[3] V.I. Pavlovic, R. Sharma, and T.S. Huang, “Visual Interpretation of Hand
Gestures for Human-Computer Interaction: A Review,” IEEE Transactions on
Pattern Analysis and Machine Intelligence, pp. 677-695, 1997.
[4] W. Ying and T.S. Huang, “Capturing Articulated Human Hand Motion: A
Divide-and-Conquer Approach,” The Proceedings of the Seventh IEEE
International Conference on Computer Vision, pp. 606-611, 1999.
[5] J. Lee and T. L. Kunii, “Model-Based Analysis of Hand Posture,” IEEE
Computer Graphics and Applications, pp. 77-86, 1995.
[6] K. Kwon, H. Zhang., and F. Dornaika, “Hand Pose Recovery with a Single
Video Camera,” IEEE International Conference on Robotics and Automation,
pp. 1194-1200, 2001.
[7] C. Kervrann and F. Heitz, “Learning Structure and Deformation Modes of
Non rigid Objects in Long Image Sequences,” International Workshop on
Automatic Face- and Gesture-Recognition, 1995.
[8] K. Cho and S.M. Dunn, “Learning Shape Classes,” IEEE Transactions on
Pattern Analysis and Machine Intelligence, pp. 882-888, 1994.
[9] R. Cipolla and N.J. Hollinghurst, “Human-Robot Interface by Pointing With
Uncalibrated Stereo Vision,” Image and Vision Computing 14, pp. 171-178,
1996.
[10] T.E. Starner and A. Pentland, “Visual Recognition of American Sign
Language Using Hidden Markov Models,” Proc. Int’l Workshop on
Automatic Face and Gesture Recognition, pp. 189-194, 1995 .
[11] B. Bauer and H. Hienz, “Relevant Features for Video-Based Continuous
Sign Language Recognition,“ Proc. of Fourth IEEE International
Conference on Automatic Face and Gesture Recognition, pp. 440-445, 2000.
[12] K. Grobel and M. Assam, “Isolated Sign Language Recognition Using Hidden
Markov Models,” Proceedings of the IEEE International Conference on
Computational Cybernetics and Simulation, pp. 162-167, 1997
[13] Y. Hamada, N. Shimada, and Y. Shirai, ”Hand Shape Estimation under
Complex Backgrounds for Sign Language Recognition,” Proceedings of 6th
Int. Conf. on Automatic and Gesture Recognition, pp. 589-594, 2004.
[14] J. Issacs and J.S. Foo, “Hand Pose Estimation for American Sign Language
Recognition,” Proc. of the Thirty-Sixth Southeastern Symposium on System
Theory, pp. 132-136, 2004.
[15] D.S. Zhang and G. Lu, “Review of Shape Representation and Description
Techniques,” Pattern Recognition, pp. 1-19, 2004.
[16] E.J. Holden and R. Owens, “Recognizing Moving Hand Shapes,” Proceedings
of the 12th International Conference on Image Analysis and Processing,
pp. 14-19, 2003
[17] C.G. Rafael and E.W. Richard, “Digital Image Processing 2/e,” Prentice
Hall, 2002.
[18] K.Kristensen, P. Nilsson, and V. Owall, “Background Segmentation beyond
RGB,” 7th Asian Conference on Computer Vision. Pp. 602-612, 2006.
[19] E. Osuna, F. Cirosi, and R. Freund, “Training Support Vector Machines:
an Application to Face Detection,” Proceedings of CVPR’97, pp. 241-246,
1997.
[20] L. Bretzner, I. Laptev, and T. Lindeberg, “Hand Gesture Recognition
using Multi-Scale Colour Features, Hierarchical Models and Particle
Filtering,” Proceedings on Automatic Face and Gesture Recognition, pp.
423-428, 2002.
[21] J. Yang, W. Lu, and W. Alex, “Skin-Color Modeling and Adaptation,”
Computer Vision – ACCV’98, pp.687-694, 1997
[22] M.H. Yang and N. Ahuja, “Gaussian Mixture Model for Human Skin Color and
Its Application in Image and Video Databases,” Conference on Storage and
Retrieval for Image and Video Databases (SPIE 99), pp. 458-466, 1999.
[23] J. Li, B. Zhao, H. Zhang, and J. Jiao, “Dual-Space Skin-Color Cue Based
Face Detection for Eye Location,” International Conference on
Information Engineering and Computer Science, pp. 1-4, 2009.
[24] M.H. Yang, D.J. Kriegman, and N. Ahuja, “Detecting Faces in Images: a
Survey,“ IEEE Trans. Pattern Anal. Mach. Intell., pp. 34-58, 2002.
[25] 王貞元, “火災偵測和區域定位應用於影像監控”, 台灣科技大學電機工程學系碩士
論文, 2008.
[26] 曹文潔, “猜拳機”, 中央大學電機工程研究所碩士論文, 2007.
[27] Barnett, V. Lewis, “Outliers in Statistical Data 3/e,” John Wiley &
Sons, 1994.
[28] WolframMathworld, http://mathworld.wolfram.com/
[29] 黃俊捷, “互動雙足式機器人之設計與實現(I)手勢辨識“, 中央大學電機工程研究
所碩士論文, 2008.
[30] 陳培森, “看護系統之監控與辨識功能建立“, 中央大學電機工程研究所碩士論文,
2006.
[31] 涂又仁, “利用人臉及手勢辨識之人機介面“, 中正大學電機工程研究所碩士論文,
2007.
[32] D. Bowman, C. Wingrave, J. Campbell, and V. Ly, “Using Pinch Gloves for
Both Natural and Abstract Interaction Techniques in Virtual
Environments,” HCI international, 2001.
[33] L. Gan, K.S. Lik, G. Hall, and U. Zeshan, “Dynamic Hand Gesture Tracking
and Recognition for Real-Time Immersive Virtual Object Manipulation, ”
CW ’09 International Conference on CyberWorlds, pp. 29-35, 2009.
[34] L. Dipietro, A.M. Sabatini, and P. Dario, “A Survey of Glove-Based
Systems and Their Applications,” IEEE Transactions on SMC, pp. 461-482, 2008.
[35] C.Y. Tsai, and Y.H. Lee, “Multiple-Camera-Based Gesture Recognition by
MDA Method,” 5th International Conference on Fuzzy Systems and Knowledge
Discovery, pp. 599-603, 2008.
[36] R. Gopalan, and B. Dariush, “Toward a Vision-Based Hand Gesture
Interface for Robotic Grasping,” IEEE International Conference on
Intelligent Robots and Systems, pp.1452-1459, 2009.
[37] Y. Sato, M. Saito, and H. Koike, “Real Time Input of 3D Pose and
Gestures of A User’s Hand and Its Application for HCI,” Virtual Reality
Annual International Symposium, pp. 79-86, 2001.
[38] S. Wan, and H.T. Nguyen, “Human Computer Interaction using Hand
Gesture,” Annual International Conference of the IEEE Engineering in
Medicine and Biology Society, pp. 2357-2360, 2008.
[39] D.J. Sturman, and D. Zeltzer, “ A Survey of Glove-Based Input,”
Computer Graphics and Applications, IEEE, pp. 30-39, 1994.
[40] X.M. Yin, and M. Xie, “Hand Gesture Segmentation, Recognition and
Application,” Computational Intelligence in Robotics and Automation
IEEE, pp. 438-443, 2001.
[41] C. Qing, N.D. Georganas, and E.M. Petriu, “Real-Time Vision-Based Hand
Gesture Recognition using Haar-Like Features,” IEEE Instrumentation and
Measurement Technology Conference, pp. 1-6, 2007.
[42] A. Malima, E. Ozgur, and M. Cetin, “A Fast Algorithm for Vision-Based
Hand Gesture Recognition for Robot Control,” IEEE 14th Signal Processing
and Communication Applications Conference, pp. 1-4, 2006.
[43] K. Sobottka, and I. Pitas, “A Novel Method for Automatic Face
Segmentation, Facial Feature Extraction and Tracking,” Signal
Processing: Image Communication, pp. 263-281, 1998.
[44] M. Soriano, B. Martinkauppi, S. Huovinen, and M. Laaksonen, “Using the
Skin Locus to Cope with Changing Illumination Conditions in Color-Based
Face Tracking,” IEEE Nordic Signal Processing Symposium, pp. 393-386,
2000.
[45] X. Yishen, G. Jihua, T. Zhi, and W. Di, “Bare Hand Gesture
Recognition,” International Congress on Image and Signal Processing,
2009.
[46] H.I. Park, and J.W. Lee, “Hand Gesture Recognition for Table-Top
Interaction System,” International Symposium on Ubiquitous VR, 2007.
[47] Y.K. Chen, T.Y. Cheng, and S.T. Chiu, “Motion Detection with Entropy in
Dynamic Background,” International Asia Conference on Informatics in
Control, Automation, and Robotics, pp. 263-266, 2009.