The website is in Maintenance mode. We are in the process of adding more features.
Any new bookmarks, comments, or user profiles made during this time will not be saved.

AIML.com

Machine Learning Resources

What is Long-Short Term Memory (LSTM)?

Bookmark this question

Long-Short Term Memory (LSTM) is an enhancement used to improve the performance of a Recurrent Neural Network (RNN) by allowing it to retain important information learned from earlier in the sequence at a later step. If a network is being trained on a large sequence, such as a paragraph of text, it might forget information from the first sentence that might be important to use in being able to generate text for the remainder of the paragraph.

The workings of an LSTM cell involve rigorous mathematics, but basically, an additional parameter called the cell state is introduced at each step of the sequence. At a given step, in addition to the input of the current step and the activation of the output from the previous step, the previous cell state is also passed in. Through a series of “gates”, the LSTM cell determines what information to retain at the next step as opposed to forgetting at that point of the sequence. Once it determines what to discard and what to carry through the sequence, it updates the cell state and passes it to the next step, continuing the sequential process.

Leave your Comments and Suggestions below:

Please Login or Sign Up to leave a comment

Partner Ad  

Find out all the ways
that you can

Explore Questions by Topics