Projects per year
Abstract
Normalization as a layer within neural networks has over the years
demonstrated its effectiveness in neural network optimization across a
wide range of different tasks, with one of the most successful
approaches being that of batch normalization. The consensus is that
better estimates of the BatchNorm normalization statistics (μ and σ2)
in each mini-batch result in better optimization. In this work, we
challenge this belief and experiment with a new variant of BatchNorm
known as GhostNorm that, despite independently normalizing batches
within the mini-batches, i.e., μ and σ2
are independently computed and applied to groups of samples in each
mini-batch, outperforms BatchNorm consistently. Next, we introduce
sequential normalization (SeqNorm), the sequential application of the
above type of normalization across two dimensions of the input, and find
that models trained with SeqNorm consistently outperform models trained
with BatchNorm or GhostNorm on multiple image classification data sets.
Our contributions are as follows: (i) we uncover a source of
regularization that is unique to GhostNorm, and not simply an extension
from BatchNorm, and illustrate its effects on the loss landscape, (ii)
we introduce sequential normalization (SeqNorm) a new normalization
layer that improves the regularization effects of GhostNorm, (iii) we
compare both GhostNorm and SeqNorm against BatchNorm alone as well as
with other regularization techniques, (iv) for both GhostNorm and
SeqNorm models, we train models whose performance is consistently better
than our baselines, including ones with BatchNorm, on the standard
image classification data sets of CIFAR–10, CIFAR-100, and ImageNet ((+0.2%, +0.7%, +0.4%), and (+0.3%, +1.7%, +1.1%) for GhostNorm and SeqNorm, respectively).
Original language | English |
---|---|
Article number | 337 |
Number of pages | 14 |
Journal | Information |
Volume | 13 |
Issue number | 7 |
DOIs | |
Publication status | Published - 12 Jul 2022 |
Keywords
- Batch normalization
- CIFAR
- Computer vision
- Ghost normalization
- ImageNet
- Loss landscape
- Neural networks
Fingerprint
Dive into the research topics of 'Sequential normalization: embracing smaller sample sizes for normalization'. Together they form a unique fingerprint.Projects
- 1 Finished
-
ICAIRD: I-CAIRD: Industrial Centre for AI Research in Digital Diagnostics
Harris-Birtill, D. C. C. (PI) & Arandelovic, O. (CoI)
1/02/19 → 31/01/23
Project: Standard