Zernike Pooling: Generalizing Average Pooling Using Zernike Moments

Most of the established neural network architectures in computer vision are essentially composed of the same building blocks (e.g., convolutional, normalization, regularization, pooling layers, etc.), with their main difference being the connectivity of these components within the architecture and n...

Full description

Bibliographic Details
Main Authors: Thomas Theodoridis, Kostas Loumponias, Nicholas Vretos, Petros Daras
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9524712/
Description
Summary:Most of the established neural network architectures in computer vision are essentially composed of the same building blocks (e.g., convolutional, normalization, regularization, pooling layers, etc.), with their main difference being the connectivity of these components within the architecture and not the components themselves. In this paper we propose a generalization of the traditional average pooling operator. Based on the requirements of <italic>efficiency</italic> (to provide information without repetition), <italic>equivalence</italic> (to be able to produce the same output as average pooling) and <italic>extendability</italic> (to provide a natural way of obtaining novel information), we arrive at a formulation that generalizes average pooling using the Zernike moments. Experimental results on <italic>Cifar 10</italic>, <italic>Cifar 100</italic> and <italic>Rotated MNIST</italic> data-sets showed that the proposed method was able to outperform the two baseline approaches, global average pooling and average pooling <inline-formula> <tex-math notation="LaTeX">$2 \times 2$ </tex-math></inline-formula>, as well as the two variants of Stochastic pooling and AlphaMEX in every case. A worst-case performance analysis on <italic>Cifar-100</italic> showed that significant gains in classification accuracy can be realised with only a modest 10&#x0025; increase in training time.
ISSN:2169-3536