淡江大學機構典藏:Item 987654321/126315
English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 64191/96979 (66%)
造訪人次 : 8333970      線上人數 : 7600
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library & TKU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋
    請使用永久網址來引用或連結此文件: https://tkuir.lib.tku.edu.tw/dspace/handle/987654321/126315


    題名: SW-YOLOX: A YOLOX-based real-time pedestrian detector with shift window-mixed attention mechanism
    作者: Tsai, Chi-Yi;Wang, Run-Yu;Chiu, Yu-Chen
    日期: 2024-08-13
    上傳時間: 2024-09-23 12:05:38 (UTC+8)
    摘要: Pedestrian detection is a critical research area in computer vision with practical applications. This paper addresses this key topic by providing a novel lightweight model named Shift Window-YOLOX (SW-YOLOX). The purpose of SW-YOLOX is to significantly enhance the robustness and real-time performance of pedestrian detection under practical application requirements. The proposed method incorporates a novel Shift Window-Mixed Attention Mechanism (SW-MAM), which combines spatial and channel attention for effective feature extraction. In addition, we introduce a novel up-sampling layer, PatchExpandingv2, to enhance spatial feature representation while maintaining computational efficiency. Furthermore, we propose a novel Shift Window-Path Aggregation Feature Pyramid Network (SW-PAFPN) to integrate with the YOLOX detector, further enhancing feature extraction and the robustness of pedestrian detection. Experimental results validated on challenging datasets such as CrowdHuman, MOT17Det, and MOT20Det demonstrate the competitive performance of the proposed SW-YOLOX compared to state-of-the-art methods and its pedestrian detection performance in crowded and complex scenes.
    關聯: Neurocomputing, Vol. 606, No. 128357, p. 1-16
    DOI: https://doi.org/10.1016/j.neucom.2024.128357
    顯示於類別:[電機工程學系暨研究所] 期刊論文

    文件中的檔案:

    檔案 描述 大小格式瀏覽次數
    index.html0KbHTML45檢視/開啟

    在機構典藏中所有的資料項目都受到原著作權保護.

    TAIR相關文章

    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library & TKU Library IR teams. Copyright ©   - 回饋