English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 49378/84106 (59%)
造訪人次 : 7377477      線上人數 : 97
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library & TKU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋
    請使用永久網址來引用或連結此文件: http://tkuir.lib.tku.edu.tw:8080/dspace/handle/987654321/80822


    題名: New memory-efficient hardware architecture of 2-D dual-mode lifting-based discrete wavelet transform for JPEG2000
    作者: Hsia, Chih-Hsien;Chiang, Jen-Shiun
    貢獻者: 淡江大學電機工程學系
    關鍵詞: 2-D 5/3 mode LDWT;2-D 9/7 mode LDWT;interlaced read scan algorithm (IRSA);lifting-based discrete wavelet transform (LDWT);low-transpose memory
    日期: 2008
    上傳時間: 2013-03-07 14:01:41 (UTC+8)
    出版者: New York: Institute of Electrical and Electronics Engineers (IEEE)
    摘要: This work presents new algorithms and hardware architectures to improve the critical issues of the 2-D dual-mode (supporting 5/3 lossless and 9/7 lossy coding) lifting-based discrete wavelet transform (LDWT). The proposed 2-D dual-mode LDWT architecture has the advantages of low-transpose memory, low latency, and regular signal flow, which is suitable for VLSI implementation. The transpose memory requirement of the N ?? N 2-D 5/3 mode LDWT is 2N, and that of 2-D 9/7 mode LDWT is 4N. According to the comparison results, the proposed hardware architecture surpasses previous architectures in the aspects of lifting-based low-transpose memory size. It can be applied to real-time visual operations such as JPEG2000, MPEG-4 still texture object decoding, and wavelet-based scalable video coding.
    關聯: Communication Systems, 2008. ICCS 2008. 11th IEEE Singapore International Conference on, pp.766-772
    DOI: 10.1109/ICCS.2008.4737288
    顯示於類別:[電機工程學系暨研究所] 會議論文

    文件中的檔案:

    檔案 描述 大小格式瀏覽次數
    index.html0KbHTML185檢視/開啟
    New Memory-Efficient Hardware Architecture.pdf396KbAdobe PDF258檢視/開啟

    在機構典藏中所有的資料項目都受到原著作權保護.

    TAIR相關文章

    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library & TKU Library IR teams. Copyright ©   - 回饋