Binarized activations for post-hoc interpretability and representation analysis in deep neural networks
| dc.contributor.author | Jamil, Huma, author | |
| dc.contributor.author | Blanchard, Nathaniel, advisor | |
| dc.contributor.author | Krishnaswamy, Nikhil, committee member | |
| dc.contributor.author | Kirby, Michael, committee member | |
| dc.contributor.author | Peterson, Christopher, committee member | |
| dc.date.accessioned | 2026-01-12T11:29:31Z | |
| dc.date.issued | 2025 | |
| dc.description.abstract | Interpretability and representation analysis have advanced rapidly, but most methods rely on the continuous activations of convolutional, attention, or pooled penultimate layers. While informative, magnitudes can blur interpretability by conflating whether a unit participates at all with how strongly it fires, and they are sensitive to rescaling. Moreover, analyses at scale can become computationally and memory intensive. To address these issues, I propose an alternative lens: bit vectors, the binarized activations of ReLU layers. While this representation achieves a 60× compression, we show that it can perform on par with established attribution methods by developing Quant-CAM, which performs competitively with gradient-based CAMs, outperforms the perturbation-based Score-CAM on ResNet-50 by 8% in deletion AUC and 2% in insertion AUC, and is over two orders of magnitude faster (minutes versus hours) to compute, while also extending naturally to Transformer architectures. Building on this lens, I investigate three aspects of bit vector representations. First, their ability to probe representation efficiency: bit vectors can rank channel importance and reveal differences in feature economy, robustness, and redundancy across supervised and self-supervised paradigms. Second, their comparability to continuous activations: bit vectors preserve interpretability quality while maintaining the 60× compression advantage. Third, their role in attribution: Quant-CAM demonstrates that binary activations are not merely a compressed signal but a foundation for effective and generalizable explanation techniques. Taken together, this work establishes bitvectors as a lightweight lens for analyzing and explaining deep networks, yielding insights into feature economy, robustness, and redundancy, and leading to Quant-CAM, an attribution method that demonstrates the strength of this binary perspective. | |
| dc.format.medium | born digital | |
| dc.format.medium | doctoral dissertations | |
| dc.identifier | Jamil_colostate_0053A_19298.pdf | |
| dc.identifier.uri | https://hdl.handle.net/10217/242757 | |
| dc.identifier.uri | https://doi.org/10.25675/3.025649 | |
| dc.language | English | |
| dc.language.iso | eng | |
| dc.publisher | Colorado State University. Libraries | |
| dc.relation.ispartof | 2020- | |
| dc.rights | Copyright and other restrictions may apply. User is responsible for compliance with all applicable laws. For information about copyright law, please see https://libguides.colostate.edu/copyright. | |
| dc.rights.access | Embargo expires: 01/07/2028. | |
| dc.title | Binarized activations for post-hoc interpretability and representation analysis in deep neural networks | |
| dc.type | Text | |
| dc.type | Image | |
| dcterms.embargo.expires | 2028-01-07 | |
| dcterms.embargo.terms | 2028-01-07 | |
| dcterms.rights.dpla | This Item is protected by copyright and/or related rights (https://rightsstatements.org/vocab/InC/1.0/). You are free to use this Item in any way that is permitted by the copyright and related rights legislation that applies to your use. For other uses you need to obtain permission from the rights-holder(s). | |
| thesis.degree.discipline | Computer Science | |
| thesis.degree.grantor | Colorado State University | |
| thesis.degree.level | Doctoral | |
| thesis.degree.name | Doctor of Philosophy (Ph.D.) |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Jamil_colostate_0053A_19298.pdf
- Size:
- 2.03 MB
- Format:
- Adobe Portable Document Format
