Invariance vs. Equivariance:

  • Convolutional Layers achieve **shift equivariant (the output shifts as the input shifts).
  • Pooling/Fully Connected Layers achieve shift invariance (the classification remains the same regardless of small shifts in input).
  • Semantic Segmentation Idea: Using ā€œFully Convolutionalā€ networks to make prediction for every pixel simultaneously

Understanding 4D Data

Deep learning frameworks typically process data in dimensions:

  1. Batch Index : which training example in the batch (the ā€œArchiveā€).
  2. Channel (): which feature map or colour channel (the ā€œCabinetā€).
  3. Row : vertical coordinate
  4. Column : Horizontal coordinate

Weight Initialisation

  • Crucial Rule: Never set all weights to or . If weights are equal, all derivatives will be the same, making neurons ā€œinterchangeableā€ and preventing learning.
  • Random Initialisation: Generally centred around 0.
  • He/Kaiming Initialisation: Best for ReLU activation functions.
  • Xavier/Glorot Initialisation: Best for Sigmoid/Tanh functions.