As we saw earlier, Hartley transform based fixed pre-processing is
inferior to that based on Fourier transform. An explanation was given
on the basis of symmetries and shift invariance in the section
. Therefore we expect improved performances from Fourier
transform even when the pre-processing is adaptive. However a training
procedure which preserves the symmetries of weight distributions must
be used. Main argument of the use of Hartley transform is to avoid the
complex weights. But as seen from fig., even Fourier
transform can be implemented as a neural network containing real
weights, but with a slightly modified network structure than the usual
MLP. We can easily derive the equations which give the forward and
backward pass.
Forward pass is given by,
where N denotes the window length, and .
If we use the notation
and error is denoted by J, then we can find simply by using the chain rule,
We assume that is known and can simply be found by differentiating eqn.2.1 wrt . Thus we get,
Eqns.2.2 and 2.3 define the backward pass. Note that can be further back propagated as usual.