Learning Rate Scheduling
Start with a large learning rate and reduce it during training (e.g., every epochs or when validation loss plateaus).
Importance of Normalisation
Network performs best when data distributions are consistent.
Input Normalisation Techniques
- Feed Scaling: Dividing pixel values by to get a range.
- Min-Max Norm: Scaling data specifically between its maximum and minimum values.
- Z-Score Norm: . This centres at with standard deviation of .
Batch Normalisation (BN):
- Concept: Normalising the output of every neuron or filter within a batch.
- Learnable Parameters: BN introduces (scaling) and (shifting) which the network learns to fund the optimal distribution.
- Test Time Difficulty: During testing, there is no “batch”.
- Solution: Precompute the mean and standard deviation during training and use those fixed values during testing
- Efficacy: BN significantly speeds up training and improves validation accuracy.