hIPPYLearn : an inexact Newton-CG method for training neural networks with analysis of the Hessian

dc.contributor.advisorGhattas, Omar N.
dc.contributor.committeeMemberDawson, Clint
dc.creatorGao, Ge, 1993-
dc.creator.orcid0000-0001-5033-1279
dc.date.accessioned2017-11-02T13:30:03Z
dc.date.available2017-11-02T13:30:03Z
dc.date.created2017-05
dc.date.issued2017-05
dc.date.submittedMay 2017
dc.date.updated2017-11-02T13:30:03Z
dc.description.abstractNeural networks, as part of deep learning, have become extremely pop- ular due to their ability to extract information from data and to generalize it to new unseen inputs. Neural network has contributed to progress in many classic problems. For example, in natural language processing, utilization of neural network significantly improved the accuracy of parsing natural language sentences [11]. However, training complicated neural network is expensive and time-consuming. In this paper, we introduce more efficient methods to train neural network using Newton-type optimization algorithm. Specifically, we use TensorFlow, the powerful machine learning package developed by Google [2] to define the structure of the neural network and the loss function that we want to optimize. TensorFlow’s automatic differentiation capabilities allow us to efficiently compute gradient and Hessian of the loss function that are needed by the scalable numerical optimization algorithm implemented in hIPPYlib [12]. Numerical examples demonstrate the better performance of Newton method compared to Steepest Descent method, both in terms of number of iterations and computational time. Another important contribution of this work is the study of the spectral properties of the Hessian of the loss function. The distribution of the eigenvalues of the Hessian, in fact, provides extremely valuable information regarding which directions in parameter space are well informed by the data.
dc.description.departmentComputational Science, Engineering, and Mathematics
dc.format.mimetypeapplication/pdf
dc.identifierdoi:10.15781/T2599ZH9S
dc.identifier.urihttp://hdl.handle.net/2152/62383
dc.language.isoen
dc.subjectMachine learning
dc.subjectNeural networks
dc.subjectOptimization problem
dc.subjectMNIST
dc.titlehIPPYLearn : an inexact Newton-CG method for training neural networks with analysis of the Hessian
dc.typeThesis
dc.type.materialtext
thesis.degree.departmentComputational Science, Engineering, and Mathematics
thesis.degree.disciplineComputational Science, Engineering, and Mathematics
thesis.degree.grantorThe University of Texas at Austin
thesis.degree.levelMasters
thesis.degree.nameMaster of Science in Computational Science, Engineering, and Mathematics

Access full-text files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
GAO-MASTERSREPORT-2017.pdf
Size:
2.62 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 2 of 2
No Thumbnail Available
Name:
PROQUEST_LICENSE.txt
Size:
4.45 KB
Format:
Plain Text
Description:
No Thumbnail Available
Name:
LICENSE.txt
Size:
1.84 KB
Format:
Plain Text
Description: