Lecture 2.7

Harmonic Networks

We conclude Module 2 by analyzing the seminal Harmonic Networks (H-Nets) paper by Worrall et al. (2017). This work was one of the first to successfully train deep steerable CNNs using the complex circular harmonic basis.

1. H-Nets as Steerable CNNs

Although the paper predates the general theory of "Steerable CNNs" (Weiler et al.), H-Nets are exactly a specific instance of this framework:

  • Group: $SO(2)$ (2D discrete and continuous rotations).
  • Basis: Complex Circular Harmonics $\Psi_m(\theta) = e^{im\theta}$.
  • Features: Complex-valued feature maps. A complex number $z = r e^{i\phi}$ naturally encodes a magnitude $r$ (feature strength) and an orientation $\phi$.

2. The Equivariance Condition

The paper derives a constraint for rotation equivariance, which they call the Equivariance Condition. It states that if an input feature map has rotation frequency $m$ and the filter has rotation frequency $k$, the output feature map must have frequency $m+k$.

$$ \text{Frequency of Output} = \text{Frequency of Input} + \text{Frequency of Filter} $$

This is exactly the Clebsch-Gordan condition for the group $SO(2)$, where the tensor product of irreducible representations corresponds to addition of frequencies: $e^{im\theta} \cdot e^{ik\theta} = e^{i(m+k)\theta}$.

3. Significance and Results

Harmonic Networks demonstrated that it is possible to achieve state-of-the-art results on rotation-sensitive tasks (like Rotated MNIST) without data augmentation. By hard-coding the symmetry into the architecture, the network becomes extremely data-efficient and robust, generalizing to orientations it has never seen during training.

This concludes Module 2. In Module 3, we will generalize all these concepts to 3D and the group $SO(3)$.