Invariance vs. Equivariance:
- Convolutional Layers achieve **shift equivariant (the output shifts as the input shifts).
- Pooling/Fully Connected Layers achieve shift invariance (the classification remains the same regardless of small shifts in input).
- Semantic Segmentation Idea: Using āFully Convolutionalā networks to make prediction for every pixel simultaneously
Understanding 4D Data
Deep learning frameworks typically process data in dimensions:
- Batch Index : which training example in the batch (the āArchiveā).
- Channel (): which feature map or colour channel (the āCabinetā).
- Row : vertical coordinate
- Column : Horizontal coordinate
Weight Initialisation
- Crucial Rule: Never set all weights to or . If weights are equal, all derivatives will be the same, making neurons āinterchangeableā and preventing learning.
- Random Initialisation: Generally centred around 0.
- He/Kaiming Initialisation: Best for ReLU activation functions.
- Xavier/Glorot Initialisation: Best for Sigmoid/Tanh functions.