研究生: |
楊忠諺 Zhong-yan Yang |
---|---|
論文名稱: |
以單一FPGA平台實現SIFT法則於自走車的地標物之視覺辨識與搜索 Single Platform of FPGA-Based SIFT for Visual Recognition and Landmark Searching of a Mobile Robot |
指導教授: |
黃志良
Chih-lyang Hwang |
口試委員: |
施慶隆
none 翁慶昌 none 李世安 none |
學位類別: |
碩士 Master |
系所名稱: |
電資學院 - 電機工程系 Department of Electrical Engineering |
論文出版年: | 2014 |
畢業學年度: | 102 |
語文別: | 中文 |
論文頁數: | 59 |
中文關鍵詞: | FPGA平台 、SIFT演算法 、影像辨識 、線性搜索的匹配準則 、自走車 、地標物的搜索策略 |
外文關鍵詞: | FPGA platform, SIFT algorithm, Visual recognition, Matching criterion, Mobile robot, Searching strategy of specific landmark. |
相關次數: | 點閱:183 下載:2 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
眾所周知,特徵的萃取為影像處理的物件辨識之基本重要步驟,為了達到萃取此較強健之特徵點,軟體執行相關演算法所需之計算較難以達到即時的效果,有鑑於此,本論文將以FPGA平台實現SIFT演算法達到快速而且能夠獲得較強健之特徵點的計算及描述。本論文亦選定滅火器及垃圾桶為地標物,並事先將此特徵描述向量儲存於FPGA平台中,並設計地標物的搜索策略,接著以自走車上的FPGA平台進行相關地標物之搜索。並以線性搜索的匹配準則,自走車在可辨識之距離進行地標物的搜索、辯識及匹配比對,最後以相關的實驗驗證所發展的平台執行SIFT演算法和地標物搜索策略之有效性及強健性。
It is known that feature extraction is a fundamental part to recognize objects in an image processed. For the purpose of robustly extracting these features, the demanded computation using software approach is difficult to obtain on-line applications. In this thesis, an FPGA-based scale invariant feature transform (SIFT) is implemented to accelerate the recognition and description of these features. The extinguisher and garbage can are chosen as specific landmarks and they are also pre-trained to accomplish their corresponding feature vectors stored in the FPGA platform. The matching criterion using linear search is then constructed to evaluate the successful recognition rate of the specific landmarks. The strategy for the search of the chosen landmarks through a mobile robot is designed and evaluated by different conditions. Finally, the corresponding experiments are given to validate the effectiveness and robustness of the proposed methodology.
[1] T. H. S. Li, S. J. Chang, “Fuzzy target tracking control of autonomous mobile robots by using infrared sensors,” IEEE Trans. Fuzzy Syst., vol. 12, no. 4, pp. 491-501, Aug. 2004.
[2] D. Bank and T. Kampke, “High resolution ultrasonic environment imaging,” IEEE Trans. Robotics, vol. 23, no. 2, pp. 370-381, Apr. 2007.
[3] S. J. Kim and B. K. Kim, “Dynamic ultrasonic hybrid localization system for indoor mobile robot,” IEEE Trans. Ind. Electron., vol. 60, no. 10, pp. 4562-4573, Oct. 2013.
[4] S. Han. H. S. Lim and J. M. Lee, “An efficient localizations scheme for a differential-driving mobile robot based on RFID system,” IEEE Trans. Ind. Electron., vol. 54, no. 6, pp. 3362-3369, Dec. 2007
[5] X. Yang, M. Moallem and R.V. Patel, “A layer global-oriented fuzzy motion planning strategy for mobile robot navigation,” IEEE Syst. Man & Cybern., Part B, vol. 35, no. 6, pp. 1214-1224, Dec. 2005.
[6] A. M. Ladd, K.E. Bekris, A. P. Rudys, D.S. Wallach and L. E. Kavraki, “ On the feasibility of using wireless Ethernet for indoor localization,” IEEE Trans. Robotics & Autom., vol. 20, no. 3, pp. 555-559, Jun. 2004.
[7] J. G. Davis, R. Sloan, and A. J. Peyton, “The 3-D positioning of wireless sensors in dispersive propagation media,” IEEE Trans. Instrum. Meas., vol. 62, no. 8, pp. 2338-2352, Aug. 2013.
[8] D. Schleicher, L. M. Bergasa, M. Ocana, R. Barea and M. E. Lopez, “Real-time hierarchical outdoor SLAM based on stereovision and GPS fusion,” IEEE Trans. Intell. Trans. Syst., vol. 10, no. 3, pp. 440-452, Sep. 2009.
[9] E. Menegatti, A. Pretto, A. Scarpa, and E. Pagello, “Omni-directional vision scan matching for robot localization in dynamic environments,” IEEE Trans. Robotics, vol. 22, no. 3, pp. 523-535, Jun. 2006.
[10] H. C. Huang and C. C. Tsai, “FPGA implementation of an embedded robust adaptive controller for autonomous omnidirectional mobile platform,” IEEE Trans. Ind. Electron., vol. 56, no. 5, pp.1604-1616, May 2009.
[11] S. K. Su, “A Comparison of Vision-Based Autonomous Navigation for Target Grasping of Humanoid Robot by Enhanced SIFT and Traditional HT Algorithms”pp.16-21, May 2013
[12] D. G. Lowe, “Distinctive features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004.