In this paper, we propose a strategy for network simplification and acceleration. First, we propose to generate a suitable resized image using multiscale patching in the first convolutional layer, which can then be used for the rest of the network. We use p convolutional filters that operate on patches of size m × n, and we first select all the possible non-superposed m × n patches from the available images. If the number of such patches is not sufficient, the remaining ones are collected using scales a × b or c × d such that ab = cd = mn or a = 2 m or b = 2n. Patches generated from the former condition are directly extracted from the images, while the downsampled results are used in the latter case. We also introduce a 2-dimensional decomposition for patch compression, by stacking all the available image patches along the columns and applying a 2D PCA decomposition. Finally, a layer weight decomposition technique followed by module-based finetuning is adopted, for a new fast module-based CNN model. Extensive evaluations using public data sets like MNIST, Pascal VOC, and WebCASIA, and with different state-of-the art CNN architectures like Open-Face, VGG and DarkNet verify that our proposed model is able to accelerate the training process and even to provide higher classification accuracies for small-sized datasets. We obtain 8% increase in Top-1 and Top-5 recognition rates, and 5% increase in F1 score over general interpolation-based resizing.
Speeded-up Convolution Neural Network for classification tasks using multiscale 2-dimensional decomposition
Ramponi G.
2020-01-01
Abstract
In this paper, we propose a strategy for network simplification and acceleration. First, we propose to generate a suitable resized image using multiscale patching in the first convolutional layer, which can then be used for the rest of the network. We use p convolutional filters that operate on patches of size m × n, and we first select all the possible non-superposed m × n patches from the available images. If the number of such patches is not sufficient, the remaining ones are collected using scales a × b or c × d such that ab = cd = mn or a = 2 m or b = 2n. Patches generated from the former condition are directly extracted from the images, while the downsampled results are used in the latter case. We also introduce a 2-dimensional decomposition for patch compression, by stacking all the available image patches along the columns and applying a 2D PCA decomposition. Finally, a layer weight decomposition technique followed by module-based finetuning is adopted, for a new fast module-based CNN model. Extensive evaluations using public data sets like MNIST, Pascal VOC, and WebCASIA, and with different state-of-the art CNN architectures like Open-Face, VGG and DarkNet verify that our proposed model is able to accelerate the training process and even to provide higher classification accuracies for small-sized datasets. We obtain 8% increase in Top-1 and Top-5 recognition rates, and 5% increase in F1 score over general interpolation-based resizing.| File | Dimensione | Formato | |
|---|---|---|---|
|
Speeded_up_Convolution_Neural_Network_for_classification_tasks_using_multiscale_2_dimensional_decomposition_.pdf
Open Access dal 31/05/2022
Descrizione: Articolo principale
Tipologia:
Bozza finale post-referaggio (post-print)
Licenza:
Creative commons
Dimensione
896.16 kB
Formato
Adobe PDF
|
896.16 kB | Adobe PDF | Visualizza/Apri |
|
1-s2.0-S092523122030919X-main.pdf
Accesso chiuso
Tipologia:
Documento in Versione Editoriale
Licenza:
Copyright Editore
Dimensione
987.14 kB
Formato
Adobe PDF
|
987.14 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


