研究生: |
葉芳嘉 Ye, Fang-Jia |
---|---|
論文名稱: |
具單目視覺距離量測之演示學習仿人機器人系統 A Humanoid Robot Learning from Demonstration System with Monocular Vision-Based Distance Measurement |
指導教授: |
王偉彥
Wang, Wei-Yen |
口試委員: |
蘇順豐
Su, Shun-Feng 呂成凱 Lu, Cheng-Kai 王偉彥 Wang, Wei-Yen |
口試日期: | 2023/07/11 |
學位類別: |
碩士 Master |
系所名稱: |
電機工程學系 Department of Electrical Engineering |
論文出版年: | 2023 |
畢業學年度: | 111 |
語文別: | 中文 |
論文頁數: | 61 |
中文關鍵詞: | 演示學習 、單目視覺 、距離量測系統 、人形機器人動作模仿系統 |
英文關鍵詞: | learning from demonstration (LfD), monocular vision, distance measurement system, humanoid robot motion imitation system |
DOI URL: | http://doi.org/10.6345/NTNU202301257 |
論文種類: | 學術論文 |
相關次數: | 點閱:77 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
本論文主要貢獻在於提出了一種基於單目視覺距離測量的仿人機器人演示學習系統。該系統結合基於數據驅動的人體動作識別和使用鏈結向量和虛擬關節的機器人運動控制,使仿人機器人可以模仿人類的動作,此外,提出了一種具單目測距方法的類視覺里程計,該方法中提出了兩種數學模型,可使用相機平面視圖圖像和不同的相機姿態下的圖像進行距離測量,所提出的方法不需要高精度雙目攝像頭或額外的傳感器來測量距離。這種方法可以應用於各種應用領域,如物料搬運、監視和自動車輛系統,具有低成本和易於實施的額外優勢。最後,ㄧ些實際的實驗證明我們的系統對於不同的相機姿態和環境的條件下具有一定的準確度和穩定性。
This thesis proposes a humanoid robot learning from demonstration system with monocular-vision-based distance measurement. The system combines a data-driven-based human action recognition and a robot motion control based on link vectors and virtual joints. This enables the humanoid robot to imitate human actions. Moreover, two mathematical models are proposed for distance measurement using flat view images and different camera poses. The proposed method does not require high-precision binocular cameras or additional sensors to measure the distance. Furthermore, this method can be utilized in various applications, such as material handling, surveillance, and autonomous vehicles, with the added advantages of low cost and ease of implementation. Finally, some practical experiments show that our system is accurate and robust to different camera poses and environmental conditions.
[1]洪正達, “台南驚傳電擊意外!水電工開電箱被炸飛全身65%燒灼傷,” 三立新聞網, 09 04 2021. Available:https://www.setn.com/News.aspx?NewsID=923171.
[2]吳映璠, “消防員無法滅火 歐最大核電廠爆炸威力恐比車諾比大10倍,” 中時新聞網, 04 03 2022. Available:https://www.chinatimes.com/realtimenews/20220304001473-260408?chdtv.
[3]shark-robotics, “shark-robotics,” shark-robotics, Available:https://www.shark-robotics.com/shark-robots#colossus.
[4]B. D. Argall, S. Chernova, M. Veloso , and B. Browning, "A survey of robot learning from demonstration," Robotics and Autonomous Systems, vol. 57, no. 5, pp. 469-483, 2009.
[5]洪權甫, 藉由物件偵測與多動作識別之機器人演示學習系統, 國立臺灣師範大學電機工程學系, 2022.
[6]T. Koch, L. Liebel, F. Fraundorfer, and M. Korner, “Evaluation of CNN-based single-image depth, ” in ECCV, 2018.
[7]A. Kouris, and C.-S. Bouganis, “Learning to fly by myself: a self-supervised CNN-based approach for autonomous navigation, ” in IROS, 2019.
[8]R.-C. Miclea, C. Dughir, A. Florin, S. Florin, and S. Ioan, “Laser and LIDAR in a system for visibility distance estimation in fog conditions, ” Sensors, vol. 20, no. 21, 5 Nov 2020.
[9]K. Karsch, C. Liu, and S. B. Kang, “Depth Transfer: depth extraction from video using non-parametric sampling, ” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 36, no. 11, pp. 2144 - 2158, 01 Nov 2014.
[10]劉邦禹, 以影像邊緣特徵加速光流法於視覺測距, 臺中技術學院資訊科技與應用研究所, 2010.
[11]羅至中, 單視域之遞迴式深度估測補償, 國立交通大學電信工程研究所, 2012.
[12]C.-C. Hsu, M.-C. Lu, W.-Y. Wang, and Y.-Y. Lu, “Distance measurement based on pixel variation of CCD images, ” ISA Trans, vol. 48, no. 4, p. 389–395, 2009.
[13]Z. Xu, L. Wang, and J. Wang, A method for distance measurement of moving objects in a monocular image, 2018.
[14]M. A. Mahammed, A. I. Melhum, and F. A. Kochery, “Object distance measurement by stereo vision, ” International journal of science and applied information technology, vol. 2, no. 2, pp. 5-8, 2013.
[15]J. Jiang, L. Liu, R. Fu, Y. Yan, and W. Shao,“Non-horizontal binocular vision ranging method based on pixels,”Optical and Quantum Electronics, vol. 52, no. 4, 2020.
[16]“Kinect for Windows SDK 2.0,” Microsoft, 2014.
Available: https://www.microsoft.com/en-us/download/details.aspx?id=44561.
[17]Z. Zhang, “A flexible new technique for camera calibration,”IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 11, pp. 1330 - 1334, 11 2000.
[18]H. B. K. Paul, Robot vision, Cambridge, MA: MIT Press, 2003.
[19]“OpenCV: camera calibration and 3D reconstruction,”
Available: https://docs.opencv.org/4.x/d9/d0c/group__calib3d.html.
[20]D. Brown, "Close-range camera calibration,”Photogramm. Eng, 2002.
[21]“camera_calibration-ROS Wiki,”Available: http://wiki.ros.org/camera_calibration.
[22]“OpenCV,”Available: https://opencv.org/.
[23]D. G. Lowe, “Distinctive image features from scale-invariant keypoints, ” International Journal of Computer Vision, vol. 60, no. 2, pp. 91-110, 2004.
[24]H. Bay, A. Ess, T. Tuytelaars, and L. Van Gool, “Speeded-up robust features (SURF), ”Computer Vision and Image Understanding, vol. 110, no. 3, pp. 346-359, 2008.
[25]E. Rublee, V. Rabaud, K. Konolige, and G. Bradski,“Orb: an efficient alternative to SIFT or SURF, ”2011 International Conference on Computer Vision, 2011.
[26]“Ultralytics YOLOv8,” Available: https://github.com/ultralytics/ultralytics.
[27]劉申禾, 基於GPU平行計算之視覺里程計, 國立臺灣師範大學電機工程學系, 2017.
[28]H. Shen, C.-C. Hsu, W.-Y. Wang, and Y.-T. Wang,“Enhanced visual odometry algorithm based on elite selection method and voting system,” in IEEE 7th International Conference on Consumer Electronics, Berlin (ICCE-Berlin), 2017.
[29]C.-H. Lin, W.-Y. Wang, S.-H. Liu, C.-C. Hsu, and C.-C. Chien, “Heterogeneous implementation of a novel indirect visual odometry system, ” IEEE Access, vol. 3, pp. 34631-34644, March 2019.
[30]ROBOTIS, “ROBOTIS OP3,” ROBOTIS,
Available:https://emanual.robotis.com/docs/en/platform/op3/introduction/#introduction.
[31]D. A. Pisner, and D. M. Schnyer, “Chapter 6 - Support vector machine, ” in Machine Learning, A. Mechelli and S. Vieira, Eds., Academic Press, 2020, pp. 101-121.
[32]林富偉, 智慧型仿人行動機器人設計, 銘傳大學電腦與通訊工程學系碩士班, 2014.
[33]Robotis, “Robotis,” Robotis, Available: https://www.robotis.com/.
[34]P.-J. Hwang, C.-C. Hsu, and W.-Y. Wang, “Development of a mimic robot—learning from demonstration incorporating object detection and multiaction recognition, ” IEEE Consumer Electronics Magazine, vol. 9, no. 3, pp. 79-87, 2020.
[35]H.-W. Ji, Y.-H. Chien, Y.-H. Lin, W.-Y. Wang, and C.-K. Lu, “Real-time human action recognition system using K-means algorithm and FNN architecture, ” in Proc. of the 2022 International Conference on Fuzzy Theory and Its Applications (iFUZZY2022), Nov. 2022.
[36]Z. Zhang, Y. Niu, Z. Yan, and S. Lin, “Real-Time whole-body imitation by humanoid robots and task-oriented teleoperation using an analytical mapping method and quantitative evaluation, ” Applied Sciences, vol. 8, no. 10, p. 2005, 22 October 2018.
[37]F.-J. Ye, Y.-H. Chien, M.-J. Hsu, W.-Y. Wang, C.-K. Lu, and C.-C. Hsu, “Monocular-vision-based distance estimation system for a humanoid robot, ” in Proc. of 2023 National SympNational Symposium on System Science and Engineering, June 2023.