簡易檢索 / 詳目顯示

研究生: 徐靖崴
Ching-Wei Hsu
論文名稱: 多人物人體動作即時辨識與異常動作追蹤系統
Multi Person Real-Time Action Recognition and Abnormal Action Tracking System
指導教授: 施慶隆
Ching-Long Shih
口試委員: 吳修明
李文猶
王乃堅
學位類別: 碩士
Master
系所名稱: 電資學院 - 電機工程系
Department of Electrical Engineering
論文出版年: 2022
畢業學年度: 110
語文別: 中文
論文頁數: 79
中文關鍵詞: 卷積神經網路長短期記憶人體關鍵點多人物動作辨識異常動作追蹤
外文關鍵詞: Convolutional Neural Network, Long Short-Term Memory, Human Body Key Points (Skeleton), Multi-Person Action Recognition, Abnormal Action Tracking
相關次數: 點閱:174下載:0
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 本論文之目的在於多人物人體動作偵測及異常動作追蹤,使用人物關鍵點座標資訊結合神經網路分類方法達到多人物之動作辨識,並即時追蹤出畫面中動作異常的人物。為了達到以上之目的本論文僅使用不具深度值的C615 RGB相機,並設計了卷積神經網路以及長短期記憶網路兩個不同的網路架構並比較兩者架構之測試結果。本文設計八種於室內活動可能出現的動作,並包含兩種異常動作,同時也建立了小型的動作數據集,作為神經網路分類訓練的資料。兩個網路模型都是透過人物的關鍵點座標資訊作為動作辨識的依據,首先將訓練資料作前處理;然後接著將神經網路模型訓練成功並以此模型作為即時系統的動作預測之目標。最後再以多物件追蹤的方法達到畫面中每一個人物的追蹤,並同時成功的完成多人物的動作辨識。


    The purpose of this paper is to perform multi-person human body motion detection and abnormal action tracking by using the key point coordinate information of the person and combined with the neural network classification method in order to achieve the multi-person action identification issue and track the abnormal action of the person in real time. This paper uses only one RGB camera C615 without the depth value, and designs convolutional neural network and a long short-term memory network to reach the action classification problem and to compare these two different network architectures. This paper has planned eight different actions that may occur in in-door, and including two kinds of abnormal actions, and establishes a small action dataset as the data for neural network classification training. Both network models use the key point coordinate information of the character as the basis for action recognition. First, the training data is preprocessed, and then the neural network model is trained successfully, and then used as a real-time system action prediction tool to achieve the goal. Finally, the multi-object tracking method is used to achieve the tracking of each person in each frame, and at the same time the multi-person action real-time recognition is performed successfully.

    摘要 I Abstract II 目錄 III 圖目錄 VI 表目錄 IX 第1章 緒論 1 1.1 研究動機與目的 1 1.2 文獻回顧 2 1.3 論文大綱 3 第2章 系統架構與任務說明 5 2.1 HAR流程 5 2.2 硬體設備介紹 6 2.3 任務說明 7 2.4 動作定義 8 第3章 人體關鍵點資料收集與前處理 11 3.1 收集所有動作影片 12 3.2 YOLO追蹤人體 13 3.3 MediaPipe Holistic人體關鍵點追蹤 14 3.4 動作資料定義與訓練資料集 17 3.5 即時性辨識系統實現 18 第4章 基於神經網路的動作辨識方法 19 4.1 類神經網路 20 4.2 卷積神經網路 21 4.2.1 卷積神經網路架構 22 4.2.2 卷積層 23 4.2.3 激勵函數 24 4.2.4 池化層 25 4.2.5 平坦層與全連接層 26 4.2.6 損失函數與優化器 27 4.3 長短期記憶 27 4.3.1 循環神經網路架構 28 4.3.2 長短期記憶網路架構 29 4.3.3 Cell 32 4.3.4 Forget Gate 33 4.3.5 Output Gate 34 4.3.6 Input Gate 34 4.4 神經網路方法比較 36 第5章 基於多人物動作辨識之方法 38 5.1 MOT基本步驟 38 5.2 SORT 39 5.3 DeepSORT 41 第6章 實驗結果與討論 43 6.1 訓練過程分析 44 6.1.1 卷積神經網路訓練過程分析 44 6.1.2 長短期記憶網路訓練過程分析 45 6.2 驗證資料評估 46 6.2.1 卷積神經網路驗證資料評估 47 6.2.2 長短期記憶網路驗證資料評估 48 6.3 實際測試與機率預測結果 50 6.3.1 實際測試 50 6.3.2 連續圖片演示系統與機率預測結果 52 6.4 多人物動作辨識結果 62 第7章 結論與建議 66 7.1 結論 66 7.2 建議 67 參考資料 68

    1. Yamato, J., J. Ohya, and K. Ishii. Recognizing Human Action in Time-Sequential Images Using Hidden Markov Model. in Proceedings 1992 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 1992.
    2. Schuldt, C., I. Laptev, and B. Caputo. Recognizing Human Actions: a Local SVM Approach. in Proceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004. 2004.
    3. Ren, Z., et al., Robust Part-Based Hand Gesture Recognition Using Kinect Sensor. IEEE Transactions on Multimedia, 2013. 15(5): p. 1110-1120.
    4. Liu, T., et al. Human Action Recognition Based on Depth Images from Microsoft Kinect. in 2013 Fourth Global Congress on Intelligent Systems. 2013.
    5. Ijjina, E.P. and C.K. Mohan. Human Action Recognition Based on Recognition of Linear Patterns in Action Bank Features Using Convolutional Neural Networks. in 2014 13th International Conference on Machine Learning and Applications. 2014.
    6. Wang, X., et al., Beyond Frame-level CNN: Saliency-Aware 3-D CNN With LSTM for Video Action Recognition. IEEE Signal Processing Letters, 2017. 24(4): p. 510-514.
    7. Cho, T.Z.W., M.T. Win, and A. Win. Human Action Recognition System Based on Skeleton Data. in 2018 IEEE International Conference on Agents (ICA). 2018.
    8. Yang, Y., et al. Human Action Recognition Based on Skeleton and Convolutional Neural Network. in 2019 Photonics & Electromagnetics Research Symposium - Fall (PIERS - Fall). 2019.
    9. Redmon, J., et al. You Only Look Once: Unified, Real-Time Object Detection. in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2016.
    10. Redmon, J. and A. Farhadi. YOLO9000: Better, Faster, Stronger. in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2017.
    11. Redmon, J. and A.J.a.e.-p. Farhadi YOLOv3: An Incremental Improvement. 2018. arXiv:1804.02767.
    12. Bochkovskiy, A., C.-Y. Wang, and H.-Y.M.J.a.e.-p. Liao YOLOv4: Optimal Speed and Accuracy of Object Detection. 2020. arXiv:2004.10934.
    13. Cao, Z., et al., OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021. 43(1): p. 172-186.
    14. Bewley, A., et al. Simple Online and Realtime Tracking. in 2016 IEEE International Conference on Image Processing (ICIP). 2016.
    15. Wojke, N., A. Bewley, and D. Paulus. Simple Online and Realtime Tracking with a Deep Association Metric. in 2017 IEEE International Conference on Image Processing (ICIP). 2017.

    無法下載圖示
    全文公開日期 2032/07/05 (校外網路)

    QR CODE