Prepare Interview

Mock Exams

Make Homepage

Bookmark this page

Subscribe Email Address

Question: Explain the concept of batch normalization and its advantages in training deep neural networks.
Answer: Batch normalization normalizes the inputs of a layer within a mini-batch, reducing internal covariate shift. It stabilizes and accelerates the training process, enables the use of higher learning rates, and acts as a form of regularization, reducing the reliance on techniques like dropout.
Is it helpful? Yes No

Most helpful rated by users:

©2025 WithoutBook