簡易檢索 / 詳目顯示

研究生: 張博勛
Po-hsun Chang
論文名稱: 自主性機器人航行在室內環境的字牌偵測與辨識技術
Text Plate Detection and Recognition Techniques Used for Autonomous Robots Navigating in Indoor Environments
指導教授: 范欽雄
Chin-shyurng Fahn
口試委員: 傅立成
Li-chen Fu
洪西進
Shi-jinn Horng
宋開泰
Kai-tai Song
學位類別: 碩士
Master
系所名稱: 電資學院 - 資訊工程系
Department of Computer Science and Information Engineering
論文出版年: 2009
畢業學年度: 97
語文別: 英文
論文頁數: 58
中文關鍵詞: 自主性機器人線性鑑別分析區域切割字牌辨識字牌偵測室內環境
外文關鍵詞: text plate detection, text plate recognition, region segmentation, linear discriminant analysis, autonomous robot, indoor environment
相關次數: 點閱:652下載:2
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 最近幾年,有許多研究者投入心力於服務性機器人的研究,其主要目的在於協助人群處理一些例行事務。日常生活中,有許多地方可由服務性機器人提供支援,譬如居家安全、娛樂,以及遞送物品等。對於室內的環境來說,如何引導這些機器人抵達預定地,並且賦予它們一些智慧去做日常工作,是一個值得探討的問題。本論文將提出一個在室內環境下,自主性機器人對於字牌偵測及辨識的技術,並藉由字牌所提供的資訊,達到導引的目標,進而完成所交付的任務,以提供諸如辦公室環境的便利生活。

    於我們所建構具有字牌偵測與辨識能力的自主性機器人,其視覺系統為主要擷取週遭資訊的配備,它是利用PTZ 攝影機獲得室內環境中的連續影像,在我們的實驗裡,將使用在影像強度上對比較高的字牌提供區域切割的資訊。在自主性機器人行進的途中,會對所取得的連續影像做字牌偵測,我們先利用顏色資訊做分析,當在取得預測區域之後,再使用邊緣偵測法來觀察每一個區域的邊點數,其中邊點數過少的區域將被過濾掉;接著,我們使用線性鑑別分析的方法做字牌辨識,若是辨識結果符合所預定的目標,則機器人會做出相對應的動作。

    在我們的實驗過程中,受到光影的影響之下,字牌偵測率大於93%,而受到遮蔽的影響之下,則高於90%;另外,字牌辨識率若以字母作統計,則平均約有93.5%。在我們的系統,影像的解析度為640×480,而偵測和辨識程序的處理速度平均一秒約有兩個畫面。


    In recent years, many researchers do the best of their abilities in the development of service robots. The main purpose of these researches is to assist people deal with routine matters. In our daily life, many supports could be offered from service robots; for example, home security, entertainment, and delivery. In an indoor environment, there is an issue worthy to be investigated that how to guide robots to reach a destination and embed them some intelligence to handle day-and-day works. In this thesis, we would present text plate detection and recognition techniques used for an autonomous robot navigating in indoor environments. We would use text plates to provide the surrounding information and achieve the goal of navigation. Furthermore, some given tasks could be accomplished by the autonomous robot to supply convenient services in an indoor environment such as an office room.

    The vision system of our autonomous robot which possesses the text plate detection and recognition abilities is the main equipment to capture surrounding information. By use of a PTZ camera, we could obtain the sequential images of indoor environments. In the experiments, we would adopt the text plates that have high contrast of image intensity to provide the information of region segmentation. As the autonomous robot moving forward, the obtained sequential images are used in text plate detection. We analyze the color information in these images to determine candidate regions, and observe the edge points in each candidate region after edge detection. Some regions would be filtered out because the number of edge points is too few in them. We would recognize the text plate by applying the linear discriminant analysis method. When the recognition result matches with the predefined target, the autonomous robot would execute the corresponding actions.

    The experimental results reveal that the text plate detection rate is larger than 93% under the influence of lights and shadows. The detection rate is higher than 90% when the text plates are occluded with blots. Besides, the average text plate recognition rate is about 93.5%, where every letter is counted in statistics. In our system, each captured frame contains 640×480 pixels, and the procedures of text plate detection and recognition work by two frames per second.

    誌謝 i 中文摘要 ii Abstract iii Contents v List of Figures vii Chapter 1 Introduction 1 1.1 Overview 1 1.2 Background and motivation 2 1.3 System description 4 1.4 Thesis organization 8 Chapter 2 Related Works 9 2.1 Review of text detection 10 2.2 Review of text recognition 14 Chapter 3 Text Plate Detection 18 3.1 Color processing of blocks 19 3.1.1 Main color selection 20 3.1.2 Main color occupation 22 3.2 Connected component processing 24 3.2.1 Connected component labeling 25 3.2.2 Labeled component filtering 26 Chapter 4 Text Plate Recognition 29 4.1 Image preprocessing 30 4.1.1 Lighting compensation 31 4.1.2 Letter normalization 32 4.2 Letter classification 33 4.2.1 Feature extraction 33 4.2.2 Linear discriminant analysis 36 Chapter 5 Experimental Results and Discussions 39 5.1 The results of text plate detection 40 5.2 The results of text plate recognition 47 Chapter 6 Conclusions and Future Works 53 6.1 Conclusions 53 6.2 Future works 54 References 56

    [1] M. Y. Shieh, J. C. Hsieh, and C. P. Cheng, “Design of an intelligent hospital service robot and its application,” in Proceedings of the International Conference on Systems, Man, and Cybernetics, Hague, Netherlands, pp. 4377-4382, 2004.
    [2] R. C. Luo, P. K. Wang, T. Y. Hsu, and T. Y. Lin, “Navigation and mobile security system of intelligent security robot,” in Proceedings of the IEEE International Conference on Industrial Technology, Hong Kong, pp. 260-265, 2005.
    [3] B. Tribelhorn and Z. Dodds, “Evaluating the Roomba: A low-cost, ubiquitous platform for robotics research and education,” in Proceedings of IEEE International Conference on Robotics and Automation, Roma, Italy, pp. 1394-1399, 2007.
    [4] W. Wu, X. Chen, and J. Yang, “Detection of text on road signs from video,” IEEE Transactions on Intelligent Transportation Systems, vol. 6, no. 4, pp. 378-390, 2005.
    [5] T. Saoi, H. Goto, and H. Kobayashi, “Text detection in color scene images based on unsupervised clustering of multi-channel wavelet features,” in Proceedings of the 8th International Conference on Document Analysis and Recognition, Seoul, Korea, vol. 2, pp. 690-694, 2005.
    [6] N. Ezaki, K. Kiyota, B. T. Minh, M. Bulacu, and L. Schomaker, “Improved text-detection methods for a camera-based text reading system for blind persons,” in Proceedings of the 8th International Conference on Document Analysis and Recognition, Seoul, Korea, vol. 1, pp. 257-261, 2005.
    [7] G. Miao, G. Zhu, S. Jiang, Q. Huang, X. Changsheng, and W. Gao, “A real-time score detection and recognition approach for broadcast basketball video,” in Proceedings of the IEEE International Conference on Multimedia and Expo., Beijing, China, pp. 1691-1694, 2007.
    [8] S. Hu and M. Chen, “Adaptive Fréchet kernel based support vector machine for text detection,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Philadelphia, Pennsylvania, vol. 5, pp. 365-368, 2005.
    [9] K. I. Kim, K. Jung, and J. H. Kim, “Texture-based approach for text detection in images using support vector machines and continuously adaptive mean shift algorithm,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, no. 12, pp. 1631-1639, 2003.
    [10] P. Dubey, “Edge based text detection for multi-purpose application,” in Proceedings of the 8th International Conference on Signal Processing, Guilin, China, vol. 4, 2006.
    [11] C. Liu, C. Wang, and R. Dai, “Text detection in images based on unsupervised classification of edge-based features,” in Proceedings of the 8th International Conference on Document Analysis and Recognition, Seoul, Korea, vol. 2, pp. 610-614, 2005.
    [12] N. Ezaki, M. Bulacu, and L. Schomaker, “Text detection from natural scene images: Towards a system for visually impaired persons,” in Proceedings of the 17th International Conference on Pattern Recognition, Cambridge, United Kingdom, vol. 2, pp. 683-686, 2004.
    [13] Y. Liu, S. Goto, and T. Ikenaga, “A robust algorithm for text detection in color images,” in Proceedings of the 8th International Conference on Document Analysis and Recognition, Seoul, Korea, vol. 1, pp. 399-403, 2005.
    [14] V. Ganapathy and C. C. H. Lean, “Optical character recognition program for images of printed text using a neural network,” in Proceedings of the IEEE International Conference on Industrial Technology, Mumbai, India, pp. 1171-1176, 2006.
    [15] T. H. Su, T. W. Zhang, H. J. Huang, and Y. Zhou, “HMM-based recognizer with segmentation-free strategy for unconstrained Chinese handwritten text,” in Proceedings of the 9th International Conference on Document Analysis and Recognition, Curitiba, Paraná, Brazil, vol. 1, pp. 133-137, 2007.
    [16] T. B. Chen, D. Ghosh, and S. Ranganath, “Video-text extraction and recognition,” in Proceedings of the IEEE Region 10 Conference, Chiang Mai, Thailand, vol. 1, pp. 319-322, 2004.
    [17] X. Chen, J. Yang., J. Zhang, and A. Waibel, “Automatic detection and recognition of signs from natural scenes,” IEEE Transactions on Image Processing, vol. 13, no. 1, pp. 87-99, 2004.
    [18] K. Suzuki, I. Horiba, and N. Sugie, “Linear-time connected-component labeling based on sequential local operations,” Computer Vision and Image Understanding, vol. 89 , no. 1, pp. 1-23, 2003.
    [19] R. L. Hsu, M. Abdel-Mottaleb, and A. K. Jain, “Face detection in color images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 5, pp. 696-706, 2002.
    [20] R. C. Gonzalez, and R. E. Woods, Digital Image Processing, 2nd Edition, Addison-Wesley, Reading, Massachusetts, 1992.
    [21] X. Lin, J. Ji, and Y. Gu, “The Euler number study of image and its application,” in Proceedings of the IEEE Conference on Industrial Electronics and Applications, Harbin, China, pp. 910-912, 2007.

    QR CODE