International Journal of Science and Research (IJSR)

International Journal of Science and Research (IJSR)
Call for Papers | Fully Refereed | Open Access | Double Blind Peer Reviewed

ISSN: 2319-7064

Downloads: 137 | Views: 292

Research Paper | Computer Science & Engineering | India | Volume 6 Issue 11, November 2017

Modified Long Short-Term Memory Recurrent Neural Network Architectures

Manish Rana [2] | Shubham Mishra [2]

Abstract: Long Short-Term Memory (LSTM) is a specific recurrent neural network (RNN) architecture that was designed to model temporal sequences and their long-range dependencies more accurately than conventional RNNs. In this paper, we explore LSTM RNN architectures and made some changes for its better performance. LSTM RNNs are more effective than DNNs. Here, we have changed the gates calculation and also have removed some unnecessary features of standard LSTM architecture. This architecture makes more effective use of model parameters than the others considered, converges quickly, and outperforms a deep feed forward neural network having an order of magnitude more parameters.

Keywords: Long Short-Term Memory, LSTM, recurrent neural network, RNN

Edition: Volume 6 Issue 11, November 2017,

Pages: 36 - 39

How to Download this Article?

Type Your Valid Email Address below to Receive the Article PDF Link

Verification Code will appear in 2 Seconds ... Wait