Date of Award:
Doctor of Philosophy (PhD)
Electrical and Computer Engineering
Jacob H. Gunther
Neural network training algorithms have always suffered from the problem of local minima. The advent of natural gradient algorithms promised to overcome this shortcoming by finding better local minima. However, they require additional training parameters and computational overhead. By using a new formulation for the natural gradient, an algorithm is described that uses less memory and processing time than previous algorithms with comparable performance.
Bastian, Michael R., "Neural Networks and the Natural Gradient" (2010). All Graduate Theses and Dissertations. 539.
Copyright for this work is retained by the student. If you have any questions regarding the inclusion of this work in the Digital Commons, please email us at .