Question : Explain the concept of batch normalization and its advantages in training deep neural networks.Reponse : Batch normalization normalizes the inputs of a layer within a mini-batch, reducing internal covariate shift. It stabilizes and accelerates the training process, enables the use of higher learning rates, and acts as a form of regularization, reducing the reliance on techniques like dropout. |
Enregistrer pour revision
Ajoutez cet element aux favoris, marquez-le comme difficile ou placez-le dans un ensemble de revision.
Connectez-vous pour enregistrer des favoris, des questions difficiles et des ensembles de revision.
Est-ce utile ? Oui Non
Les plus utiles selon les utilisateurs :
- Explain the purpose of an activation function in a neural network.
- What is transfer learning, and how is it used in deep learning?
- What is a convolutional neural network (CNN), and how is it different from a fully connected neural network?