Improving Deep Neural Networks: Optimization, Regularization, and Generative Modeling

dc.contributor.advisorLi, Zongpeng
dc.contributor.advisorSamavati, Faramarz
dc.contributor.authorZhang, Zijun
dc.contributor.committeememberLi, Zongpeng
dc.contributor.committeememberDenzinger, Jorg
dc.contributor.committeememberKrishnamurthy, Diwakar
dc.date2020-06
dc.date.accessioned2019-12-13T20:09:12Z
dc.date.available2019-12-13T20:09:12Z
dc.date.issued2019-12
dc.description.abstractIn the past decade, deep learning has revolutionized the fields of computer vision, speech recognition, natural language processing, and continues spreading to many other fields. Therefore, it is important to better understand and improve deep neural networks (DNNs), which serve as the backbone of deep learning. In this thesis, we approach this topic from three different perspectives: optimization, regularization, and generative modeling. Firstly, we address the generalization gap recently observed between adaptive optimization methods, such as Adam, and simple stochastic gradient descent (SGD). We develop a tailored version of Adam for training DNNs, which is shown to close the gap on image classification tasks. Secondly, we identify a side effect of a widely used regularization technique, dropout, and multiplicative noise in general. That is, multiplicative noise tends to increase the correlation between features. We then exploit batch normalization to efficiently remove the correlation effect. Finally, we focus on generative modeling, a fundamental application of DNNs. We propose a framework for training autoencoder-based generative models, with non-adversarial losses and unrestricted neural network architectures.en_US
dc.identifier.citationZhang, Z. (2019). Improving Deep Neural Networks: Optimization, Regularization, and Generative Modeling (Doctoral thesis, University of Calgary, Calgary, Canada). Retrieved from https://prism.ucalgary.ca.en_US
dc.identifier.doihttp://dx.doi.org/10.11575/PRISM/37338
dc.identifier.urihttp://hdl.handle.net/1880/111343
dc.language.isoengen_US
dc.publisher.facultyScienceen_US
dc.publisher.institutionUniversity of Calgaryen
dc.rightsUniversity of Calgary graduate students retain copyright ownership and moral rights for their thesis. You may use this material in any way that is permitted by the Copyright Act or through licensing that has been assigned to the document. For uses that are not allowable under copyright legislation or licensing, you are required to seek permission.en_US
dc.subject.classificationComputer Scienceen_US
dc.titleImproving Deep Neural Networks: Optimization, Regularization, and Generative Modelingen_US
dc.typedoctoral thesisen_US
thesis.degree.disciplineComputer Scienceen_US
thesis.degree.grantorUniversity of Calgaryen_US
thesis.degree.nameDoctor of Philosophy (PhD)en_US
ucalgary.item.requestcopytrueen_US
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ucalgary_2019_zhang_zijun.pdf
Size:
8.75 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.62 KB
Format:
Item-specific license agreed upon to submission
Description: