簡易檢索 / 詳目顯示

研究生: 何耕任
Ho-Keng Jen
論文名稱: 基於追瞳系統之人機介面開發─以瞳孔軌跡姿態辨識與凝視控制為例
On the Development of a Pupil Tracking-based Human Computer Interface Used for Pupil Posture Recognition and Gaze Control
指導教授: 范欽雄
Chin-Shyurng Fahn
口試委員: 傅楸善
none
王榮華
none
古鴻炎
none
學位類別: 碩士
Master
系所名稱: 電資學院 - 資訊工程系
Department of Computer Science and Information Engineering
論文出版年: 2015
畢業學年度: 103
語文別: 英文
論文頁數: 67
中文關鍵詞: 電腦視覺瞳孔追蹤瞳孔軌跡姿態辨識凝視控制人機介面
外文關鍵詞: Computer vision, pupil tracking, pupil posture recognition, gaze control, human computer interface
相關次數: 點閱:247下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 日前人機互動介面的使用方式越來越多樣化,其中,基於種種優點,以電腦視覺的軌跡追蹤與辨識方式,是目前研究領域的一個熱門題目。又人類能夠整日移動瞳孔方向而不會感到疲勞,因此瞳孔追蹤是一個能夠整日使用的理想人機介面互動方式。此外,在移動裝置上,尤其是筆記型電腦、平板電腦、智慧型手機,大多配備了前鏡頭,因此瞳孔追蹤系統的安裝可望在不改動使用者硬體的狀況下完成。由於以上的好處,越來越多的研究機構和消費電子公司,都積極投入瞳孔追蹤控制的研究。
    然而,人的眼睛和瞳孔,在前鏡頭的占據面積並不大,而且容易受到低光源雜訊的干擾,導致偵測準確率並不高,從而在偵測與追蹤的過程之中,很容易發生誤判,從而產生錯誤定位與雜點。因此,在本篇論文中,我們提出改良過的瞳孔偵測方法,並使用卡爾曼濾波器,以及Cam-Shift追蹤法,來達成較好的追蹤結果,提升瞳孔軌跡辨識的正確率。
    本篇論文題出兩種瞳孔控制方法,其中一個即是瞳孔軌跡辨識,藉由辨識到的瞳孔移動行為,來控制電腦。另一個則是使用眼睛注視的區域控制滑鼠指標位置。此兩種系統能夠投入日常使用,特別是需要非接觸式的使用者介面時,或者是提供殘疾人士使用。
    本篇論文使用類神經網路進行瞳孔軌跡辨識,達到80%以上的正確率。由此實證,揭示了本篇論文的系統具有足夠的強健性提供使用者使用。


    The motion tracking is currently a popular research topic by the use of computer vision. Human beings move their pupils all the time without feeling tired. Therefore, pupil tracking is an ideal input method for non-contact computer interface for daylong use. In mobile devices, such as tablets, and smart phones, the front facing camera can be utilized for pupil tracking directly. Because of the benefits of pupil tracking, more and more institutes and companies are developing this kind of technology.
    Because the eyes and pupil occupies small ratio in the frames of a front facing camera, and the quality of front facing cameras is limited due to some manufacture reasons, the trajectory from the detection is rather noisy especially in low light condition. Therefore, in this thesis, we proposed our pupil detection methods and use tracking methods as well as Kalman filter to eliminate noisy data, and obtain better and acceptable recognition results.
    Two human computer interface are designed. One is a pupil posture recognition system, which recognizes pupil movement trajectory as commands. The other is a pupil-mouse mapping, which detects the gaze position to move the cursor of a computer. The proposed methods and systems can be employed into daily use of humans, especially for the occasions the interface must be non-contact, and those who are mentally challenged.
    The neural network classifiers are employed and tested for our proposed systems. The accuracy of the pupil posture is above 80% according to our experimental results, which means our proposed method is robust enough for daily use.

    中文摘要i Abstractii 致謝iii Contentsiv List of Figuresvii List of Tablesx Chapter 1Introduction1 1.1Overview1 1.2Motivation1 1.3Challenges in Pupil Detection from Mobile Devices3 1.4System Description4 1.5 Thesis organization5 Chapter 2Related Works7 2.1Psychological Analysis7 2.2Eye Control8 2.3Gaze Control9 2.4Comparison of the Pupil Tracking Systems13 Chapter 3Eye Tracking and Pupil Tracking15 3.1Eye Detection15 3.1.1Feature Extraction15 3.1.2AdaBoost19 3.1.3Haar Cascading23 3.2Eye Tracking25 3.2.1Mean-Shift25 3.2.2Cam-shift27 3.3Pupil Detection and Tracking27 Chapter 4Human Computer Interface Based on Pupil Tracking36 4.1Preprocessing of Pupil Trajectories36 4.1.1Trajectory Formulation36 4.1.2Kalman Filter37 4.2Pupil Posture Recognition40 4.2.1Recognition by Neural Networks40 4.2.2Recognition by Multiple Layered Perceptron42 4.3Pupil-mouse Mapping46 4.3.1Homography Matching by Four Points Calibration46 4.3.2Homography Matching by Five Points48 4.3.3Gaze Control with Calibrated Results51 Chapter 5Experimental Results and Discussions53 5.1Experimental Setup53 5.2Results of Eye Detection54 5.3Results of Eye Tracking55 5.4Results of Pupil Detection55 5.5Results of Pupil Tracking57 5.6Results of Pupil Posture Recognition59 5.7The Results of Pupil-mouse Mapping62 Chapter 6Conclusions and Future Works64 6.1Conclusions64 6.2Future Works64 References65

    [1]“SentiGaze Eye Movement Tracking System,” [Online] Available http://www.neurotechnology.com/press_release_sentigaze.html (accessed on June 26, 2015)
    [2]H. F. Ho, "Reading Process of Integrated Circuit Layout Debugging: Evidence from Eye Movements," Advanced Materials Research, vol. 787, pp. 855-860, 2013
    [3]R. Ballagas, J. Borchers, M. Rohs, and J. G. Sheridan, "The smart phone: a ubiquitous input device," Pervasive Computing, vol. 5, no. 1, pp. 70-77, 2006.
    [4]S. Instruments, "iView X HED," ed, 2009. [Online] Available http://www.smivision.com/en/gaze-and-eye-tracking-systems/products/overview.html (accessed on June 26, 2015)
    [5]T. Yoshioka, S. Nakashima, J. Odagiri, H. Tomimori, and T. Fukui, "Pupil detection in the presence of specular reflection," in Proceedings of the Symposium on Eye Tracking Research and Applications, pp. 363-364. Florida, 2014.
    [6]“The Eye Tribe,” [Online] Available https://theeyetribe.com/ (accessed on June 12, 2015)
    [7]"Anadolu CV Eye Tracker," [Online] Available https://www.youtube.com/watch?v=8x0dW9BdkcU (accessed on July 3, 2015)
    [8]R. Lienhart and J. Maydt, "An extended set of haar-like features for rapid object detection," in Proceedings of International Conference on Image Processing, vol. 1., pp. 900-903, New York, 2002.

    [9]P. Viola and M. J. Jones, "Robust real-time face detection," International Journal of Computer Vision, vol. 57, no. 2, pp. 137-154, 2004.
    [10]T. Mita, T. Kaneko, and O. Hori, "Joint haar-like features for face detection," in Proceedings of International Conference on Computer Vision, pp. 1619-1626, Beijing, China, 2005.
    [11]Y. Freund, R. Schapire, and N. Abe, "A short introduction to boosting," Journal-Japanese Society for Artificial Intelligence, vol. 14, no. 5, pp. 1612, 1999.
    [12]A. Kasinski and A. Schmidt, "The architecture and performance of the face and eyes detection system based on the Haar cascade classifiers," Pattern Analysis and Applications, vol. 13, no. 2, pp. 197-211, 2010.
    [13]D. Comaniciu and P. Meer, "Mean shift: A robust approach toward feature space analysis," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 5, pp. 603-619, 2002.
    [14]D. Exner, E. Bruns, D. Kurz, A. Grundhöfer, and O. Bimber, "Fast and robust CAMShift tracking," in Proceedings of Computer Vision and Pattern Recognition Workshops, pp. 9-16, San Francisco, CA, 2010.
    [15]E. R. Dougherty, and R. A. Lotufo, Hands-on morphological image processing vol. 71, SPIE press Bellingham, 2003.
    [16]N. Otsu, "A threshold selection method from gray-level histograms," Automatica, vol. 11, no. 1, pp. 23-27, 1975.
    [17]T. Pavlidis, "Algorithms for shape analysis of contours and waveforms," IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. PAMI-2, no. 4, pp. 301-312, 1980.
    [18]R. E. Kalman, "A new approach to linear filtering and prediction problems," Journal of Fluids Engineering, vol. 82, no. 1, pp. 35-45, 1960.
    [19]D. E. Rumelhart, G. E. Hinton, and R. J. Williams, "Learning representations by back-propagating errors," Cognitive Modeling, vol. 5, no. 3, p. 213-220, 1988.
    [20]A. Goh, "Back-propagation neural networks for modeling complex systems," Artificial Intelligence in Engineering, vol. 9, no. 3 , pp. 143-151, 1995.
    [21]Z. Zhang and A. R. Hanson, "3D reconstruction based on homography mapping," in Proceedings of ARPA Image Understanding Workshop, pp. 1007-1012, Cambridge, MA, 1996.
    [22]J. M. Henderson, "Human gaze control during real-world scene perception," Trends in cognitive sciences, vol. 7, no. 11, pp. 498-504, 2003.

    無法下載圖示 全文公開日期 2020/08/12 (校內網路)
    全文公開日期 2025/08/12 (校外網路)
    全文公開日期 2025/08/12 (國家圖書館:臺灣博碩士論文系統)
    QR CODE