Measuring (In)variances in convolutional networks

2Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Convolutional neural networks (CNN) offer state-of-the-art performance in various computer vision tasks such as activity recognition, face detection, medical image analysis, among others. Many of those tasks need invariance to image transformations (i.e., rotations, translations or scaling). This work proposes a versatile, straightforward and interpretable measure to quantify the (in)variance of CNN activations with respect to transformations of the input. Intermediate output values of feature maps and fully connected layers are also analyzed with respect to different input transformations. The technique is applicable to any type of neural network and/or transformation. Our technique is validated on rotation transformations and compared with the relative (in)variance of several networks. More specifically, ResNet, AllConvolutional and VGG architectures were trained on CIFAR10 and MNIST databases with and without rotational data augmentation. Experiments reveal that rotation (in)variance of CNN outputs is class conditional. A distribution analysis also shows that lower layers are the most invariant, which seems to go against previous guidelines that recommend placing invariances near the network output and equivariances near the input.

Cite

CITATION STYLE

APA

Quiroga, F., Torrents-Barrena, J., Lanzarini, L., & Puig, D. (2019). Measuring (In)variances in convolutional networks. In Communications in Computer and Information Science (Vol. 1050 CCIS, pp. 98–109). Springer. https://doi.org/10.1007/978-3-030-27713-0_9

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free