What is GRU()?
The Gated Recurrent Unit (GRU) is a model that makes LSTM a little simpler. The input gate and the forgetting gate are integrated into one gate as an "update gate".
Similar to LSTM, introducing such an oblivion and update gate makes it easy to maintain the memory of the features of events before long steps.
That's because it can be said that shortcut paths that bypass between each time step are efficiently generated.
Because of this, errors can be easily back-propagated during learning, which reduces the problem of gradient loss.
When GRU become "a better choice" than LSTMs?
The biggest difference between GRU and LSTM is that GRU is faster and easier to execute (but it is not rich in expressiveness).
In practice, the advantages tend to offset the weaknesses, such as at the expense of performance, as large networks are needed, for example, to enrich the expressive power. GRU performs better than LSTM when it does not require expressive power.
Title | Detail |
---|---|
Environment | MacOS Mojave 10.14.3 |
Language | Python |
Library | Kras, scikit-learn, Numpy, matplotlib, Pandas, Seaborn |
Dataset | News Category Dataset |
Algorithm | GRU Network |
- Understanding GRU Networks
- Creating A Text Generator Using Recurrent Neural Network
- Text Generation With LSTM Recurrent Neural Networks in Python with Keras
- An applied introduction to LSTMs for text generation — using Keras and GPU-enabled Kaggle Kernels
- Word-level LSTM text generator. Creating automatic song lyrics with Neural Networks
- ニューラルネットワークの動物園 : ニューラルネットワーク・アーキテクチャのチートシート(後編)