TAILIEUCHUNG - Mạng thần kinh thường xuyên cho dự đoán P9

A Class of Normalised Algorithms for Online Training of Recurrent Neural Networks A normalised version of the real-time recurrent learning (RTRL) algorithm is introduced. This has been achieved via local linearisation of the RTRL around the current point in the state space of the network. Such an algorithm provides an adaptive learning rate normalised by the L2 norm of the gradient vector at the output neuron. The analysis is general and also covers simpler cases of feedforward networks and linear FIR filters. | Recurrent Neural Networks for Prediction Authored by Danilo P. Mandic Jonathon A. Chambers Copyright 2001 John Wiley Sons Ltd ISBNs 0-471-49517-4 Hardback 0-470-84535-X Electronic 9 A Class of Normalised Algorithms for Online Training of Recurrent Neural Networks Perspective A normalised version of the real-time recurrent learning RTRL algorithm is introduced. This has been achieved via local linearisation of the RTRL around the current point in the state space of the network. Such an algorithm provides an adaptive learning rate normalised by the L2 norm of the gradient vector at the output neuron. The analysis is general and also covers simpler cases of feedforward networks and linear FIR filters. Introduction Gradient-descent-based algorithms for training neural networks such as the backpropagation backpropagation through time recurrent backpropagation RBP and real-time recurrent learning RTRL algorithm typically suffer from slow convergence when dealing with statistically nonstationary inputs. In the area of linear adaptive filters similar problems with the LMS algorithm have been addressed by utilising normalised algorithms such as NLMS. We therefore introduce a normalised RTRL-based learning algorithm with the idea to impose similar stabilisation and convergence effects on training of RNNs as normalisation imposes on the LMS algorithm. In the area of linear FIR adaptive filters it is shown Soria-Olivas et al. 1998 that a normalised gradient-descent-based learning algorithm can be derived starting from the Taylor series expansion of the instantaneous output error of an adaptive FIR filter given by e k 1 e k de k w d e k A w k AW k e k 1 e k dwi k Aw k 2 2 dwi k dwj k k j k 150 OVERVIEW From the mathematical description of LMS1 from Chapter 2 we have and de k dwi k x k i 1 i 1 2 . N Awi k p k e k x k i 1 i 1 2 . N. Due to the linearity of the FIR filter the second- and higher-order partial derivatives in vanish. Combining - yields

TỪ KHÓA LIÊN QUAN
TAILIEUCHUNG - Chia sẻ tài liệu không giới hạn
Địa chỉ : 444 Hoang Hoa Tham, Hanoi, Viet Nam
Website : tailieuchung.com
Email : tailieuchung20@gmail.com
Tailieuchung.com là thư viện tài liệu trực tuyến, nơi chia sẽ trao đổi hàng triệu tài liệu như luận văn đồ án, sách, giáo trình, đề thi.
Chúng tôi không chịu trách nhiệm liên quan đến các vấn đề bản quyền nội dung tài liệu được thành viên tự nguyện đăng tải lên, nếu phát hiện thấy tài liệu xấu hoặc tài liệu có bản quyền xin hãy email cho chúng tôi.
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.