研究生: |
莊崴丞 Chuang Wei-Cheng |
---|---|
論文名稱: |
在複雜動態環境下利用損失函數方法的手勢操控系統 Hand Gesture Recognition System with the Use of Basic Loss Functions in a Complicated and Dynamic Environment |
指導教授: |
蘇順豐
Shun-Feng Su |
口試委員: |
郭重顯
Chung-Hsien Kuo 王偉彥 Wei-Yen Wang 莊鎮嘉 Chen-Chia Chuang |
學位類別: |
碩士 Master |
系所名稱: |
電資學院 - 電機工程系 Department of Electrical Engineering |
論文出版年: | 2012 |
畢業學年度: | 100 |
語文別: | 英文 |
論文頁數: | 108 |
中文關鍵詞: | 複雜背景 、動態環境 、膚色偵測 、損失函數 |
外文關鍵詞: | complicated background, dynamic environment, skin-color detection, loss function. |
相關次數: | 點閱:191 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
本研究目的在於讓行動不便之者可以不需要藉由手來推動輪椅,而是透過手勢直接導引、操控輪椅。一般的手勢辨識相關文獻只探討如何提高手勢辨識的精準度,在本篇論文,將會討論如何在動態背景下去處理我們的影像,以及在複雜的背景下擷取出完整的手勢前景。本文提出的系統之流程可以分為四個步驟,第一,首先利用背景相減法取得手部前景。第二,由於我們的背景是動態,所以更新我們的背景當 =0.05。並且設立region of interest (ROI)來判斷我們的手是否達到臨界值,然後決定系統是否開始偵測,第三,使用Robust Learning Algorithms中的loss functions概念,去對統計學的離群值(outlier)做延伸應用到膚色偵測上,去濾除一些不必要的膚色雜訊以達到完整擷取出我們的手勢,最後我們利用極座標統計手的像素,找出手指指尖,以方便我們使用指尖來導引輪椅。此方法證明對於在一些複雜場景、膚色背景下的效果都不錯。從實驗結果可發現本文提出的系統除了背景的膚色雜訊過大或者手部光源分布不均,整體而言是相當成功的。
The purpose of this study is to let people with physical disabilities not need to drive the wheel, but guide and control it directly on the wheelchair by gestures. Unlike other studies that focus on the stage of hand gesture recognition, In this paper, we will discuss how to process image in a dynamic and complicated background to obtain the complete hand of the foreground. There are 4 stages in the proposed system. First, it is to use the background subtraction to get the hand image. Since the background is dynamic, the background is updated with =0.05. Secondly, the region of interest (ROI) is considered to decide whether the foreground arrives the threshold value and whether our system starts to detect the hand, Thirdly, in order to extract the completely gesture from the image, unnecessary skin-color noises are removed by a skin-color approach. The proposed skin-color detection approach is adopted from the idea of Robust Learning Algorithms with use of loss functions for outlier. Finally, the method of polar coordinates is employed to count the pixels of the hand, and the fingertips are found to guide conveniently the wheelchair. The method is performed promisingly for complicated sites and skin-color backgrounds. Experimental results show that the proposed system is quite promising except that a very small number of frames are misjudged because the system cannot deal with some problems such as the area of the noise being too large or the distribution of light source being too uneven on the hand region.
References
[1] 內政部統計處, http://www.moi.gov.tw/stat/
[2] 莊鎮嘉, “在極大誤差情況之強健塑模”, 國立台灣科技大學電機工程研究所博士論文, 2000
[3] 廖崇儒, “可供輪椅使用者使用之基於電腦視覺的手勢辨識系統”, 碩士論文, 國立台灣科技大學電機工程研究所, 2010
[4] Kinect for Windows, http://www.microsoft.com/en-us/kinectforwindows/
[5] 張榮財, 陳建宏, “應用主成份分析法與倒傳遞類神經網路於手語手勢辨識”, 2010 9th Conference on Information Technology and Applications in Outlying Islands, pp.2-5.
[6] K.H. Park, H.E. Lee, Y. Kim, and Z.Z. Bien, “A Steward Robot for Human-Friendly Human-Machine Interaction in a Smart House Environment,” IEEE Transactions on Automation Science and Engineering, pp. 21-25, 2008.
[7] S.W. Lee, “Automatic Gesture Recognition for Intelligent Human-Robot Interaction,”1 Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition, pp.1-6, 2006.
[8] Y. Sato, M. Saito, and H. Koike, “Real Time Input of 3D Pose and Gestures of A
User’s Hand and Its Application for HCI,” Virtual Reality Annual International
Symposium, pp. 79-86, 2001.
[9] W. Ying and T.S. Huang, “Capturing Articulated Human Hand Motion: A Divide-and-Conquer Approach,” The Proceedings of the Seventh IEEE
International Conference on Computer Vision, pp. 606-611, 1999.
[10] J. Lee and T. L. Kunii, “Model-Based Analysis of Hand Posture,” IEEE Computer Graphics and Applications, pp. 77-86, 1995.
[11] K. Kwon, H. Zhang., and F. Dornaika, “Hand Pose Recovery with a Single Video Camera,” IEEE International Conference on Robotics and Automation, pp.1194-1200, 2001.
[12] Y. Shirai and Y. Kuno, “Hand Posture Estimation by Combining 2-D Appearance-based and 3-D Model-based Approacbes” IEEE Pattern Recognition, 2000. Proceedings. 15th International Conference on , vol.3, no., pp.705-708 vol.3, 2000.
[13] N. Shimada, K. Kimura, Y. Shirai and Y. Kuno, “Rapid Hand Posture Recognition Using Adaptive Histogram Template of Skin and Hand Edge Contour,” Machine Vision and Image Processing (MVIP), 2010 6th Iranian , vol., no., pp.1-5, 27-28 Oct. 2010
[14] L.W. Howe, F. Wong, A. Chekima, “Comparison of Hand Segmentation Methodologies for Hand Gesture, ” IEEE Information Technology, 2008. ITSim 2008. International Symposium on , vol.2, no., pp.1-7, 26-28
[15] 葉宗樺, “應用於網路攝影機於指尖軌跡圖形辨識”, 碩士論文, 南台科技大學, 2006
[16] J. Issacs and J.S. Foo, “Hand Pose Estimation for American Sign Language Recognition,” Proc. of the Thirty-Sixth Southeastern Symposium on System Theory, pp. 132-136, 2004.
[17] Wikipedia, http://en.wikipedia.org/wiki/Principal_component_analysis
[18] 曹文潔,“猜拳機”, 國立中央大學電機工程所,碩士論文, 2007.
[19] 黃俊捷, “互動雙足式機器人之設計與實現(I)手勢辨識”, 國立中央大學電機工程所,碩士論文,2008
[20] C. Kim and J.N. Hwang, “Fast and Automatic Video Object Segmentation and Tracking for Content-Based Applications,” IEEE Transactions on Circuits and System for Video Technology, 12(7), pp.597-612, 2002.
[21] S. Murali and R. Girisha, “Segmentation of Motion Objects from Surveillance Video Sequences using Temporal Differencing Combined with Multiple Correlation,” 6th IEEE International Conference on Advanced Video and Signal Based Surveillance, AVSS 2009, art. no.5279625, pp.472-477, 2009.
[22] J.L. Barron, D.J. Fleet, and S.S. Beauchemin, “Performance of optical flow techniques,” International Journal of Computer Vision, 12(1), pp.43-77, 1994.
[23] C. Kim and J.N. Hwang, “Fast and Automatic Video Object Segmentation and Tracking for Content-Based Applications,” IEEE Transactions on Circuits and System for Video Technology, 12(7), pp.597-612, 2002.
[24] Z. Zivkovic, “Improved Adaptive Gaussian Mixture Model for Background Subtraction,” Proceedings-International Conference on Pattern Recognition, 2, pp.28-31, 2004.
[25] C.G. Rafael and E.W. Richard, “Digital Image Processing 2/e,” Prentice Hall,
2002.
[26] L. K. Simone and D. G. Kamper, “Design considerations for a wearable monitor to measure finger posture.” Journal of NeuroEngineering and Rehabilitation, 2005.
[27] S. Singh, D.S. Chauhan, M. Vatsa and R. Singh, “A Robust Skin Color Based Face Detection Algorithm ,” Tamkang Journal of Science and Engineering, pp. 227-234, 2003.
[28] Wikipedia, http://zh.wikipedia.org/wiki/File:HSV_cone.jpg
[29] M.H. Yang, D.J. Kriegman, and N. Ahuja, “Detecting Faces in Images: a
Survey,“ IEEE Trans. Pattern Anal. Mach. Intell., pp. 34-58, 2002.
[30] R.L. Hsu, A. Mottaleb and K. Jain, “Face detection in color images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 696 - 706, 2002
[31] V. Tartter and K. Knowlton. “Perception of sign language from an array of 27 moving spots,” Nature 289, pp. 676-678, 1981.
[32] R.L. Hsu, A. Mottaleb and K. Jain, “Face detection in color images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 696 - 706, 2002.
[33] VPL Research Inc., DataGlove Model 2 User’s Manual, Redwood City, CA, 1987.
[34] Virtex Co,“ Company brochure,” Stanford, CA, October, 1992.
[35] Kaiser, Polhemus, “ 3 Space user’s manual, ” A Kaiser Aerospace & Electronics Company, 1987.
[36] V. Frati, D. Prattichizzo, “Using Kinect for hand tracking and rendering in wearable haptics,” IEEE World Haptics Conference, pp. 317-321, 2011
[37] Z. Ren, J. Meng, J. Yuan, “Depth Camera Based Hand Gesture Recognition
and its Applications in Human-Computer-Interaction,” International Conference on Information and Communication Systems,2011
[38] 辛柏陞,“虛擬實境手部功能訓練系統之設計開發與成效探討之研究”, 國立中央大學機械工程研究所, 博士論文, 2005.
[39] 李慶銘, “即時影音教學傳播系統的實現”, 碩士論文,國立台灣科技大學電機工程系,2011.
[40] 王貞元, “火災偵測和區域定位應用於影像監”, 台灣科技大學電機工程學系碩論文, 2008.
[41] 趙于翔, “可攜式台灣手語發聲系統”, 淡江大學電機工程研究所,碩士論文, 2002.
[42] Barnett, V. Lewis, “Outliers in Statistical Data 3/e,” John Wiley & Sons, 1994.
[43] Math Open Reference, http://www.mathopenref.com/tocs/trigfunctionstoc.html
[44] 涂又仁, “利用人臉及手勢辨識之人機介面”, 國立中正大學電機工程研究所, 碩士論文, 2007.
[45] E.J. Holden and R. Owens, “Recognizing Moving Hand Shapes,” Proceedings of
the 12th International Conference on Image Analysis and Processing, pp. 14-19,
2003
[46] MAD, http://en.wikipedia.org/wiki/Median_absolute_deviation
[47] Regression, http://en.wikipedia.org/wiki/Regression_analysis