English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 60696/93562 (65%)
造訪人次 : 1054016      線上人數 : 39
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library & TKU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋
    請使用永久網址來引用或連結此文件: https://tkuir.lib.tku.edu.tw/dspace/handle/987654321/120548

    題名: Generating a 3D Hand Model from Position of Fingertip Using Image Processing Technique
    作者: Anuwattananon, Natthapach;Ruengittinun, Somchoke;Yen, Shwu-Huey
    關鍵詞: cameras;gesture recognition;human computer interaction;image motion analysis;solid modelling
    日期: 2019-08-05
    上傳時間: 2021-04-20 12:10:41 (UTC+8)
    摘要: A gesture from hands and fingers have rich meanings in communication even without a word of sound. It would be very useful if a computer can understand a hand gesture. Hence, we can use a hand gesture to communicate with a robot and perform certain activities. This study focuses on tracking the position of each fingertip and palm to make a computer knows the gesture of a hand. The proposed solution was initially implemented using a MS Kinect camera while capturing a depth image of a human hand. Then, we applied some image processing algorithms to track the positions of fingertips. Finally, the result was visualized in a real-time 3D hand model based on the movements/signs given by a human hand. The experiment results indicate that the proposed approach can literally track the positions of a fingertip.
    DOI: 10.1109/Ubi-Media.2019.00020
    顯示於類別:[資訊工程學系暨研究所] 會議論文


    檔案 描述 大小格式瀏覽次數



    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library & TKU Library IR teams. Copyright ©   - 回饋