کد مقاله کد نشریه سال انتشار مقاله انگلیسی نسخه تمام متن
406610 678101 2014 11 صفحه PDF دانلود رایگان
عنوان انگلیسی مقاله ISI
Learning long-term dependencies in segmented-memory recurrent neural networks with backpropagation of error
ترجمه فارسی عنوان
یادگیری وابستگی های طولانی مدت در شبکه های عصبی مجدد حافظه جداگانه با عقب راندن خطا
کلمات کلیدی
شبکه عصبی مکرر، شبکه عصبی مکرر در حافظه جداگانه، از بین رفتن مشکل شیب، وابستگی های بلند مدت، پیش آموزش بدون مراقبت
موضوعات مرتبط
مهندسی و علوم پایه مهندسی کامپیوتر هوش مصنوعی
چکیده انگلیسی

In general, recurrent neural networks have difficulties in learning long-term dependencies. The segmented-memory recurrent neural network (SMRNN) architecture together with the extended real-time recurrent learning (eRTRL) algorithm was proposed to circumvent this problem. Due to its computational complexity eRTRL becomes impractical with increasing network size. Therefore, we introduce the less complex extended backpropagation through time (eBPTT) for SMRNN together with a layer-local unsupervised pre-training procedure. A comparison on the information latching problem showed that eRTRL is better able to handle the latching of information over longer periods of time, even though eBPTT guaranteed a better generalisation when training was successful. Further, pre-training significantly improved the ability to learn long-term dependencies with eBPTT. Therefore, the proposed eBPTT algorithm is suited for tasks that require big networks where eRTRL is impractical. The pre-training procedure itself is independent of the supervised learning algorithm and can improve learning in SMRNN in general.

ناشر
Database: Elsevier - ScienceDirect (ساینس دایرکت)
Journal: Neurocomputing - Volume 141, 2 October 2014, Pages 54–64
نویسندگان
, , , ,