Normalisation is dead, long live normalisation!
2022-01-17ICLR Track Blog 2022Unverified0· sign in to hype
Anonymous
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Since the advent of Batch Normalisation (BN) almost every state-of-the-art (SOTA) method uses some form of normalisation. After all, normalisation generally speeds up learning and leads to models that generalise better than their unnormalised counterparts. This turns out to be especially useful when using some form of skip connections, which are prominent in Residual Networks (ResNets), for example. However, Brock et al. (2021a) suggest that SOTA performance can also be achieved using ResNets without normalisation!