Pruning visual transformers to increase model compression and decrease inference time
dc.contributor.author | Yost, James E., author | |
dc.contributor.author | Whitley, Darrell, advisor | |
dc.contributor.author | Ghosh, Sudipto, committee member | |
dc.contributor.author | Betten, Anton, committee member | |
dc.date.accessioned | 2024-05-27T10:31:53Z | |
dc.date.available | 2024-05-27T10:31:53Z | |
dc.date.issued | 2024 | |
dc.description.abstract | We investigate the efficacy of pruning a visual transformer during training to reduce inference time while maintaining accuracy. Various training techniques were explored, including epoch-based training, fixed-time training, and training to achieve a specific accuracy threshold. Results indicate that pruning from the inception of training offers significant reductions in inference time without sacrificing model accuracy. Different pruning rates were evaluated, demonstrating a trade-off between training speed and model compression. Slower pruning rates allowed for better convergence to higher accuracy levels and more efficient model recovery. Furthermore, we examine the cost of pruning and the recovery time of pruned models. Overall, the findings suggest that early-stage pruning strategies can effectively produce smaller, more efficient models with comparable or improved performance compared to non-pruned counterparts, offering insights into optimizing model efficiency and resource utilization in AI applications. | |
dc.format.medium | born digital | |
dc.format.medium | masters theses | |
dc.identifier | Yost_colostate_0053N_18226.pdf | |
dc.identifier.uri | https://hdl.handle.net/10217/238378 | |
dc.language | English | |
dc.language.iso | eng | |
dc.publisher | Colorado State University. Libraries | |
dc.relation.ispartof | 2020- | |
dc.rights | Copyright and other restrictions may apply. User is responsible for compliance with all applicable laws. For information about copyright law, please see https://libguides.colostate.edu/copyright. | |
dc.subject | machine learning | |
dc.subject | transformer | |
dc.subject | pruning | |
dc.subject | artificial intelligence | |
dc.title | Pruning visual transformers to increase model compression and decrease inference time | |
dc.type | Text | |
dcterms.rights.dpla | This Item is protected by copyright and/or related rights (https://rightsstatements.org/vocab/InC/1.0/). You are free to use this Item in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s). | |
thesis.degree.discipline | Computer Science | |
thesis.degree.grantor | Colorado State University | |
thesis.degree.level | Masters | |
thesis.degree.name | Master of Science (M.S.) |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Yost_colostate_0053N_18226.pdf
- Size:
- 1.04 MB
- Format:
- Adobe Portable Document Format