In this paper, we propose a strategy for network simplification and acceleration. First, we propose to generate a suitable resized image using multiscale patching in the first convolutional layer, which can then be used for the rest of the network. We use p convolutional filters that operate on patches of size m × n, and we first select all the possible non-superposed m × n patches from the available images. If the number of such patches is not sufficient, the remaining ones are collected using scales a × b or c × d such that ab = cd = mn or a = 2 m or b = 2n. Patches generated from the former condition are directly extracted from the images, while the downsampled results are used in the latter case. We also introduce a 2-dimensional decomposition for patch compression, by stacking all the available image patches along the columns and applying a 2D PCA decomposition. Finally, a layer weight decomposition technique followed by module-based finetuning is adopted, for a new fast module-based CNN model. Extensive evaluations using public data sets like MNIST, Pascal VOC, and WebCASIA, and with different state-of-the art CNN architectures like Open-Face, VGG and DarkNet verify that our proposed model is able to accelerate the training process and even to provide higher classification accuracies for small-sized datasets. We obtain 8% increase in Top-1 and Top-5 recognition rates, and 5% increase in F1 score over general interpolation-based resizing.

Speeded-up Convolution Neural Network for classification tasks using multiscale 2-dimensional decomposition

Ramponi G.
2020-01-01

Abstract

In this paper, we propose a strategy for network simplification and acceleration. First, we propose to generate a suitable resized image using multiscale patching in the first convolutional layer, which can then be used for the rest of the network. We use p convolutional filters that operate on patches of size m × n, and we first select all the possible non-superposed m × n patches from the available images. If the number of such patches is not sufficient, the remaining ones are collected using scales a × b or c × d such that ab = cd = mn or a = 2 m or b = 2n. Patches generated from the former condition are directly extracted from the images, while the downsampled results are used in the latter case. We also introduce a 2-dimensional decomposition for patch compression, by stacking all the available image patches along the columns and applying a 2D PCA decomposition. Finally, a layer weight decomposition technique followed by module-based finetuning is adopted, for a new fast module-based CNN model. Extensive evaluations using public data sets like MNIST, Pascal VOC, and WebCASIA, and with different state-of-the art CNN architectures like Open-Face, VGG and DarkNet verify that our proposed model is able to accelerate the training process and even to provide higher classification accuracies for small-sized datasets. We obtain 8% increase in Top-1 and Top-5 recognition rates, and 5% increase in F1 score over general interpolation-based resizing.
File in questo prodotto:
File Dimensione Formato  
Speeded_up_Convolution_Neural_Network_for_classification_tasks_using_multiscale_2_dimensional_decomposition_.pdf

Open Access dal 31/05/2022

Descrizione: Articolo principale
Tipologia: Bozza finale post-referaggio (post-print)
Licenza: Creative commons
Dimensione 896.16 kB
Formato Adobe PDF
896.16 kB Adobe PDF Visualizza/Apri
1-s2.0-S092523122030919X-main.pdf

Accesso chiuso

Tipologia: Documento in Versione Editoriale
Licenza: Copyright Editore
Dimensione 987.14 kB
Formato Adobe PDF
987.14 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11368/2969119
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact