Evolutionary neural architecture search for deep learning

dc.contributor.advisorMiikkulainen,, Risto
dc.contributor.committeeMemberStone, Peter
dc.contributor.committeeMemberBaldick, Ross
dc.contributor.committeeMemberHuang, Qixing
dc.creatorLiang, Jason Zhi
dc.creator.orcid0000-0002-7041-9136
dc.date.accessioned2019-04-11T17:07:11Z
dc.date.available2019-04-11T17:07:11Z
dc.date.created2018-12
dc.date.issued2019-02-08
dc.date.submittedDecember 2018
dc.date.updated2019-04-11T17:07:12Z
dc.description.abstractDeep neural networks (DNNs) have produced state-of-the-art results in many benchmarks and problem domains. However, the success of DNNs depends on the proper configuration of its architecture and hyperparameters. DNNs are often not used to their full potential because it is difficult to determine what architectures and hyperparameters should be used. While several approaches have been proposed, computational complexity of searching large design spaces makes them impractical for large modern DNNs. This dissertation introduces an efficient evolutionary algorithm (EA) for simultaneous optimization of DNN architecture and hyperparameters. It builds upon extensive past research of evolutionary optimization of neural network structure. Various improvements to the core algorithm are introduced, including: (1) discovering DNN architectures of arbitrary complexity; (1) generating modular, repetitive modules commonly seen in state-of-the-art DNNs; (3) extending to the multitask learning and multiobjective optimization domains; (4) maximizing performance and reducing wasted computation through asynchronous evaluations. Experimental results in image classification, image captioning, and multialphabet character recognition show that the approach is able to evolve networks that are competitive with or even exceed hand-designed networks. Thus, the method enables an automated and streamlined process to optimize DNN architectures for a given problem and can be widely applied to solve harder tasks.
dc.description.departmentComputer Science
dc.format.mimetypeapplication/pdf
dc.identifier.urihttps://hdl.handle.net/2152/74258
dc.identifier.urihttp://dx.doi.org/10.26153/tsw/1388
dc.language.isoen
dc.subjectNeural architecture search
dc.subjectDeep learning
dc.subjectNeuroevolution
dc.subjectEvolutionary computation
dc.subjectArtificial intelligence
dc.titleEvolutionary neural architecture search for deep learning
dc.typeThesis
dc.type.materialtext
thesis.degree.departmentComputer Sciences
thesis.degree.disciplineComputer Science
thesis.degree.grantorThe University of Texas at Austin
thesis.degree.levelDoctoral
thesis.degree.nameDoctor of Philosophy

Access full-text files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
LIANG-DISSERTATION-2018.pdf
Size:
5.74 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
LICENSE.txt
Size:
1.84 KB
Format:
Plain Text
Description: