Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent

A. Yegenoglu, K. Krajsek, S.D. Pier, M. Herty, in: Lecture Notes in Computer Science, Springer International Publishing, Cham, 2021.

Download
No fulltext has been uploaded.
Book Chapter | Published | English
Author
Yegenoglu, AlperLibreCat ; Krajsek, Kai; Pier, Sandra Diaz; Herty, Michael
Abstract
The successful training of deep neural networks is dependent on initialization schemes and choice of activation functions. Non-optimally chosen parameter settings lead to the known problem of exploding or vanishing gradients. This issue occurs when gradient descent and backpropagation are applied. For this setting the Ensemble Kalman Filter (EnKF) can be used as an alternative optimizer when training neural networks. The EnKF does not require the explicit calculation of gradients or adjoints and we show this resolves the exploding and vanishing gradient problem. We analyze different parameter initializations, propose a dynamic change in ensembles and compare results to established methods.
Publishing Year
Book Title
Lecture Notes in Computer Science
LibreCat-ID

Cite this

Yegenoglu A, Krajsek K, Pier SD, Herty M. Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent. In: Lecture Notes in Computer Science. Springer International Publishing; 2021. doi:10.1007/978-3-030-64580-9_7
Yegenoglu, A., Krajsek, K., Pier, S. D., & Herty, M. (2021). Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent. In Lecture Notes in Computer Science. Springer International Publishing. https://doi.org/10.1007/978-3-030-64580-9_7
@inbook{Yegenoglu_Krajsek_Pier_Herty_2021, place={Cham}, title={Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent}, DOI={10.1007/978-3-030-64580-9_7}, booktitle={Lecture Notes in Computer Science}, publisher={Springer International Publishing}, author={Yegenoglu, Alper and Krajsek, Kai and Pier, Sandra Diaz and Herty, Michael}, year={2021} }
Yegenoglu, Alper, Kai Krajsek, Sandra Diaz Pier, and Michael Herty. “Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-Performing Gradient Descent.” In Lecture Notes in Computer Science. Cham: Springer International Publishing, 2021. https://doi.org/10.1007/978-3-030-64580-9_7.
A. Yegenoglu, K. Krajsek, S. D. Pier, and M. Herty, “Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-performing Gradient Descent,” in Lecture Notes in Computer Science, Cham: Springer International Publishing, 2021.
Yegenoglu, Alper, et al. “Ensemble Kalman Filter Optimizing Deep Neural Networks: An Alternative Approach to Non-Performing Gradient Descent.” Lecture Notes in Computer Science, Springer International Publishing, 2021, doi:10.1007/978-3-030-64580-9_7.

Export

Marked Publications

Open Data LibreCat

Search this title in

Google Scholar
ISBN Search