Automated deep learning architecture design using differentiable architecture search (DARTS)
dc.contributor.author | Sharma, Kartikay, author | |
dc.contributor.author | Anderson, Chuck, advisor | |
dc.contributor.author | Beveridge, Ross, committee member | |
dc.contributor.author | Kirby, Michael, committee member | |
dc.date.accessioned | 2020-01-13T16:42:13Z | |
dc.date.available | 2020-01-13T16:42:13Z | |
dc.date.issued | 2019 | |
dc.description.abstract | Creating neural networks by hand is a slow trial-and-error based process. Designing new architectures similar to GoogleNet or FractalNets, which use repeated tree-based structures, is highly likely to be inefficient and sub-optimal because of the large number of possibilities for composing such structures. Recently, neural architecture search algorithms have been able to automate the process of architecture design and have often attained state-of-the-art performances on CIFAR-10, ImageNet and Penn Tree Bank datasets. Even though the search time has been reduced to tens of GPU hours from tens of thousands of GPU hours, most search algorithms rely on additional controllers and hypernetworks to generate architecture encoding or predict weights for sampled architectures. These controllers and hypernetworks might require optimal structure when deployed on a new task on a new dataset. And since this is done by hand, the problem of architecture search is not really solved. Differentiable Architecture Search (DARTS) avoids this problem by using gradient descent methods. In this work, the DARTS algorithm is studied under various conditions and search hyperparameters. DARTS is applied to CIFAR-10 to check reproducibility of the original results. It is also tested in a new setting — on the CheXpert dataset — to discover new architectures and is compared to a baseline DenseNet121 model. The architectures searched using DARTS achieve better performance on the validation set than the baseline model. | |
dc.format.medium | born digital | |
dc.format.medium | masters theses | |
dc.identifier | Sharma_colostate_0053N_15836.pdf | |
dc.identifier.uri | https://hdl.handle.net/10217/199856 | |
dc.language | English | |
dc.language.iso | eng | |
dc.publisher | Colorado State University. Libraries | |
dc.relation.ispartof | 2000-2019 | |
dc.rights | Copyright and other restrictions may apply. User is responsible for compliance with all applicable laws. For information about copyright law, please see https://libguides.colostate.edu/copyright. | |
dc.subject | convolutional neural networks (CNNs) | |
dc.subject | differentiable architecture search | |
dc.subject | CheXpert dataset | |
dc.subject | neural architecture search | |
dc.subject | DARTS | |
dc.title | Automated deep learning architecture design using differentiable architecture search (DARTS) | |
dc.type | Text | |
dcterms.rights.dpla | This Item is protected by copyright and/or related rights (https://rightsstatements.org/vocab/InC/1.0/). You are free to use this Item in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s). | |
thesis.degree.discipline | Computer Science | |
thesis.degree.grantor | Colorado State University | |
thesis.degree.level | Masters | |
thesis.degree.name | Master of Science (M.S.) |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Sharma_colostate_0053N_15836.pdf
- Size:
- 1.48 MB
- Format:
- Adobe Portable Document Format