top of page

ARTICLE: Recurrent Neural Networks - how do they work?

  • Forfatters billede: Marco Singh
    Marco Singh
  • 16. nov. 2016
  • 1 min læsning

Recurrent Neural Networks (RNN) in different shapes have been very succesful in recent times. The model captures long-term dependencies, which is very desirable in many applications. Microsoft researchers recently used a combination of Convolutional Neural Networks and LSTM's (a special version of a RNN) to reach human parity in conversational speech, and many more applications are seeing the light. This article explores the standard RNN and I dig into every mathematical detail of the network structure and in particular the Backpropagation Through Time (BPTT) algorithm. I haven't seen any paper going into every computation of the derivatives, hence this is what I wanted to achieve. The paper builds the foundation for the further understanding of the LSTM model. Reading it does require prior knowledge of a feedforward Neural Network and the backpropagation in this setup (see my previous article in case you haven't familiarized yourself with it) and basic calculus. CLICK ON THE PICTURE BELOW TO READ THE ARTICLE

 
 
 

Seneste blogindlæg

Se alle
Recent Posts

© 2016 by Marco Singh. Proudly created with Wix.com 

  • Grey Twitter Icon
bottom of page