English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 49291/83828 (59%)
造訪人次 : 7159222      線上人數 : 74
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library & TKU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋
    請使用永久網址來引用或連結此文件: http://tkuir.lib.tku.edu.tw:8080/dspace/handle/987654321/80119


    題名: Reinforced Two-Step-Ahead Weight Adjustment Technique for Online Training of Recurrent Neural Networks
    作者: Chang, Li-Chiu;Chen, Pin-An;Chang, Fi-John
    貢獻者: 淡江大學水資源及環境工程學系
    關鍵詞: Real-time recurrent learning (RTRL) algorithm, recurrent neural network (RNN);streamflow forecast;time series forecast
    日期: 2012-08
    上傳時間: 2013-01-18 00:04:08 (UTC+8)
    出版者: Piscataway: Institute of Electrical and Electronics Engineers
    摘要: A reliable forecast of future events possesses great value. The main purpose of this paper is to propose an innovative learning technique for reinforcing the accuracy of two-step-ahead (2SA) forecasts. The real-time recurrent learning (RTRL) algorithm for recurrent neural networks (RNNs) can effectively model the dynamics of complex processes and has been used successfully in one-step-ahead forecasts for various time series. A reinforced RTRL algorithm for 2SA forecasts using RNNs is proposed in this paper, and its performance is investigated by two famous benchmark time series and a streamflow during flood events in Taiwan. Results demonstrate that the proposed reinforced 2SA RTRL algorithm for RNNs can adequately forecast the benchmark (theoretical) time series, significantly improve the accuracy of flood forecasts, and effectively reduce time-lag effects.
    關聯: IEEE Transactions on Neural Networks and Learning Systems 23(8), pp.1269-1278
    DOI: 10.1109/TNNLS.2012.2200695
    顯示於類別:[水資源及環境工程學系暨研究所] 期刊論文

    文件中的檔案:

    檔案 描述 大小格式瀏覽次數
    index.html0KbHTML168檢視/開啟

    在機構典藏中所有的資料項目都受到原著作權保護.

    TAIR相關文章

    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library & TKU Library IR teams. Copyright ©   - 回饋