Slow Feature Analysis

Setting

Inputs x(t)Rn, and output signals yj(t)=gj(x(t))

Objective

Minimize Δ(yj)=yj˙2t under the constraints

  1. yjt=0 (zero mean)
  2. yj2t=1 (unit variance)
  3. yiyjt=0:i<j (decorrelation and order)

Linear case (SFA)

g(x)=wTf(x) where f(x)=(f1(x),fM(x))T defines a non-linear basis. Then, we can write

Δ(yj)=yj˙(t)2t=wTz˙(t)z˙(t)Ttw=wTC˙wvar(yj)=yj(t)2t=wTz(t)z(t)Ttw=wTCw

Under the Lagrangian formulation, we get the objective function

L(w)=y˙2tλy2t=wTC˙wλwTCw

Taking the gradient and setting it to zero gives the generalized eigenvalue problem:

C˙w=λCw

The cool thing is that this shows that SFA isn't just about the correlation structure in the data but the joint correlation of the data and its dynamics.


Links

Sources