Volltext-Downloads (blau) und Frontdoor-Views (grau)

Should You Go Deeper? : Optimizing Convolutional Neural Network Architectures without Training

  • When optimizing convolutional neural networks (CNN) for a specific image-based task, specialists commonly overshoot the number of convolutional layers in their designs. By implication, these CNNs are unnecessarily resource intensive to train and deploy, with diminishing beneficial effects on the predictive performance.The features a convolutional layer can process are strictly limited by its receptive field. By layer-wise analyzing the size of the receptive fields, we can reliably predict sequences of layers that will not contribute qualitatively to the test accuracy in the given CNN architecture. Based on this analysis, we propose design strategies based on a so-called border layer. This layer allows to identify unproductive convolutional layers and hence to resolve these inefficiencies, optimize the explainability and the computational performance of CNNs. Since neither the strategies nor the analysis requires training of the actual model, these insights allow for a very efficient design process of CNN architectures, which might be automated in the future.

Export metadata

Additional Services

Search Google Scholar

Statistics

frontdoor_oas
Metadaten
Author:Mats L. RichterORCiD, Julius SchöningORCiD, Anna Wiedenroth, Ulf KrumnackORCiD
Title (English):Should You Go Deeper? : Optimizing Convolutional Neural Network Architectures without Training
DOI:https://doi.org/10.1109/ICMLA52953.2021.00159
ISBN:978-1-6654-4337-1
Parent Title (English):20th IEEE International Conference on Machine Learning and Applications : ICMLA 2021 : 13-16 December 2021, virtual event : proceedings
Publisher:IEEE
Place of publication:Piscataway, NJ
Document Type:Conference Proceeding
Language:English
Year of Completion:2021
electronic ID:Zur Anzeige in scinos
Release Date:2024/08/27
First Page:964
Last Page:971
Note:
20th IEEE International Conference on Machine Learning and Applications (ICMLA 2021), 13.12.-16.12.2021, Online
Note:
Zugriff im Hochschulnetz
Faculties:Fakultät IuI
DDC classes:000 Allgemeines, Informatik, Informationswissenschaft / 004 Informatik
Review Status:Peer Reviewed