Numerical influence of ReLU’(0) on backpropagation - INSA Toulouse - Institut National des Sciences Appliquées de Toulouse Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2021

Numerical influence of ReLU’(0) on backpropagation

Influence de ReLU'(0) sur la backpropagation

Résumé

In theory, the choice of ReLU'(0) in [0, 1] for a neural network has a negligible influence both on backpropagation and training. Yet, in the real world, 32 bits default precision combined with the size of deep learning problems makes it a hyperparameter of training methods. We investigate the importance of the value of ReLU'(0) for several precision levels (16, 32, 64 bits), on various networks (fully connected, VGG, ResNet) and datasets (MNIST, CIFAR10, SVHN). We observe considerable variations of backpropagation outputs which occur around half of the time in 32 bits precision. The effect disappears with double precision, while it is systematic at 16 bits. For vanilla SGD training, the choice ReLU'(0) = 0 seems to be the most efficient. We also evidence that reconditioning approaches as batch-norm or ADAM tend to buffer the influence of ReLU'(0)'s value. Overall, the message we want to convey is that algorithmic differentiation of nonsmooth problems potentially hides parameters that could be tuned advantageously.
Fichier principal
Vignette du fichier
Impact_of_ReLU_Prime.pdf (1005.23 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03265059 , version 1 (22-06-2021)
hal-03265059 , version 2 (29-06-2021)
hal-03265059 , version 3 (18-10-2023)

Identifiants

  • HAL Id : hal-03265059 , version 1

Citer

David Bertoin, Jérôme Bolte, Sébastien Gerchinovitz, Edouard Pauwels. Numerical influence of ReLU’(0) on backpropagation. 2021. ⟨hal-03265059v1⟩
344 Consultations
993 Téléchargements

Partager

Gmail Facebook X LinkedIn More