- 【Updated on May 12, 2025】 Integration of CiNii Dissertations and CiNii Books into CiNii Research
- Trial version of CiNii Research Automatic Translation feature is available on CiNii Labs
- Suspension and deletion of data provided by Nikkei BP
- Regarding the recording of “Research Data” and “Evidence Data”
The Vanishing Gradient Problem During Learning Recurrent Neural Nets and Problem Solutions
-
- Sepp Hochreiter
- Institut für Informatik, Technische Universität München, München, D-80290, Germany
Description
<jats:p> Recurrent nets are in principle capable to store past inputs to produce the currently desired output. Because of this property recurrent nets are used in time series prediction and process control. Practical applications involve temporal dependencies spanning many time steps, e.g. between relevant inputs and desired outputs. In this case, however, gradient based learning methods take too much time. The extremely increased learning time arises because the error vanishes as it gets propagated back. In this article the de-caying error flow is theoretically analyzed. Then methods trying to overcome vanishing gradients are briefly discussed. Finally, experiments comparing conventional algorithms and alternative methods are presented. With advanced methods long time lag problems can be solved in reasonable time. </jats:p>
Journal
-
- International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems
-
International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 06 (02), 107-116, 1998-04
World Scientific Pub Co Pte Lt
- Tweet
Details 詳細情報について
-
- CRID
- 1360855569653404544
-
- ISSN
- 17936411
- 02184885
-
- Data Source
-
- Crossref