淡江大學機構典藏:Item 987654321/118148
English  |  正體中文  |  简体中文  |  Items with full text/Total items : 62805/95882 (66%)
Visitors : 3953297      Online Users : 781
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library & TKU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version
    Please use this identifier to cite or link to this item: https://tkuir.lib.tku.edu.tw/dspace/handle/987654321/118148


    Title: Visually Guided Picking Control of an Omnidirectional Mobile Manipulator Based on End-to-End Multi-Task Imitation Learning
    Authors: Chi-Yi Tsai;Yung-Shan Chou;Ching-Chang Wong;Yu-Cheng Lai;Chien-Che Huang
    Keywords: Omnidirectional mobile manipulator;visually guided picking control;deep learning;multi-task imitation learning;end-to-end control
    Date: 2019-12-25
    Issue Date: 2020-03-02 12:10:16 (UTC+8)
    Publisher: IEEE
    Abstract: In this paper, a novel deep convolutional neural network (CNN) based high-level multi-task
    control architecture is proposed to address the visual guide-and-pick control problem of an omnidirectional
    mobile manipulator platform based on deep learning technology. The proposed mobile manipulator control
    system only uses a stereo camera as a sensing device to accomplish the visual guide-and-pick control
    task. After the stereo camera captures the stereo image of the scene, the proposed CNN-based high-level
    multi-task controller can directly predict the best motion guidance and picking action of the omnidirectional mobile manipulator by using the captured stereo image. In order to collect the training dataset, we manually controlled the mobile manipulator to navigate in an indoor environment for approaching and picking up an object-of-interest (OOI). In the meantime, we recorded all of the captured stereo images and the corresponding control commands of the robot during the manual teaching stage. In the training stage, we employed the end-to-end multi-task imitation learning technique to train the proposed CNN model by
    learning the desired motion and picking control strategies from prior expert demonstrations for visually
    guiding the mobile platform and then visually picking up the OOI. Experimental results show that the
    proposed visually guided picking control system achieves a picking success rate of about 78.2% on average.
    Relation: IEEE Access 8, p.1882-1891
    DOI: 10.1109/ACCESS.2019.2962335
    Appears in Collections:[Graduate Institute & Department of Electrical Engineering] Journal Article

    Files in This Item:

    File SizeFormat
    index.html0KbHTML92View/Open

    All items in 機構典藏 are protected by copyright, with all rights reserved.


    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library & TKU Library IR teams. Copyright ©   - Feedback