DSConv: efficient convolution operator

<p>Quantization is a popular way of increasing the speed and lowering the memory usage of Convolution Neural Networks (CNNs). When labelled training data is available, network weights and activations have successfully been quantized down to 1-bit. The same cannot be said about the scenario whe...

Ausführliche Beschreibung

Bibliographische Detailangaben
Hauptverfasser: Gennari do Nascimento, M, Fawcett, R, Prisacariu, VA
Format: Internet publication
Sprache:English
Veröffentlicht: 2019
_version_ 1826313305552060416
author Gennari do Nascimento, M
Fawcett, R
Prisacariu, VA
author_facet Gennari do Nascimento, M
Fawcett, R
Prisacariu, VA
author_sort Gennari do Nascimento, M
collection OXFORD
description <p>Quantization is a popular way of increasing the speed and lowering the memory usage of Convolution Neural Networks (CNNs). When labelled training data is available, network weights and activations have successfully been quantized down to 1-bit. The same cannot be said about the scenario when labelled training data is not available, e.g. when quantizing a pre-trained model, where current approaches show, at best, no loss of accuracy at 8-bit quantizations. We introduce DSConv, a flexible quantized convolution operator that replaces single-precision operations with their far less expensive integer counterparts, while maintaining the probability distributions over both the kernel weights and the outputs. We test our model as a plug-and-play replacement for standard convolution on most popular neural network architectures, ResNet, DenseNet, GoogLeNet, AlexNet and VGG-Net and demonstrate state-of-the-art results, with less than 1% loss of accuracy, without retraining, using only 4-bit quantization. We also show how a distillation-based adaptation stage with unlabelled data can improve results even further.</p>
first_indexed 2024-09-25T04:10:56Z
format Internet publication
id oxford-uuid:a93ea289-a44e-4ce0-b450-8577c22296af
institution University of Oxford
language English
last_indexed 2024-09-25T04:10:56Z
publishDate 2019
record_format dspace
spelling oxford-uuid:a93ea289-a44e-4ce0-b450-8577c22296af2024-06-20T13:57:43ZDSConv: efficient convolution operatorInternet publicationhttp://purl.org/coar/resource_type/c_7ad9uuid:a93ea289-a44e-4ce0-b450-8577c22296afEnglishSymplectic Elements2019Gennari do Nascimento, MFawcett, RPrisacariu, VA<p>Quantization is a popular way of increasing the speed and lowering the memory usage of Convolution Neural Networks (CNNs). When labelled training data is available, network weights and activations have successfully been quantized down to 1-bit. The same cannot be said about the scenario when labelled training data is not available, e.g. when quantizing a pre-trained model, where current approaches show, at best, no loss of accuracy at 8-bit quantizations. We introduce DSConv, a flexible quantized convolution operator that replaces single-precision operations with their far less expensive integer counterparts, while maintaining the probability distributions over both the kernel weights and the outputs. We test our model as a plug-and-play replacement for standard convolution on most popular neural network architectures, ResNet, DenseNet, GoogLeNet, AlexNet and VGG-Net and demonstrate state-of-the-art results, with less than 1% loss of accuracy, without retraining, using only 4-bit quantization. We also show how a distillation-based adaptation stage with unlabelled data can improve results even further.</p>
spellingShingle Gennari do Nascimento, M
Fawcett, R
Prisacariu, VA
DSConv: efficient convolution operator
title DSConv: efficient convolution operator
title_full DSConv: efficient convolution operator
title_fullStr DSConv: efficient convolution operator
title_full_unstemmed DSConv: efficient convolution operator
title_short DSConv: efficient convolution operator
title_sort dsconv efficient convolution operator
work_keys_str_mv AT gennaridonascimentom dsconvefficientconvolutionoperator
AT fawcettr dsconvefficientconvolutionoperator
AT prisacariuva dsconvefficientconvolutionoperator