As we saw earlier, Hartley transform based fixed pre-processing is
inferior to that based on Fourier transform. An explanation was given
on the basis of symmetries and shift invariance in the section
. Therefore we expect improved performances from Fourier
transform even when the pre-processing is adaptive. However a training
procedure which preserves the symmetries of weight distributions must
be used. Main argument of the use of Hartley transform is to avoid the
complex weights. But as seen from fig.
, even Fourier
transform can be implemented as a neural network containing real
weights, but with a slightly modified network structure than the usual
MLP. We can easily derive the equations which give the forward and
backward pass.
Forward pass is given by,
 
where N denotes the window length, and  
 .
If we use the notation
 
 
 and error is
denoted by J, then  we can find
 
  simply by using the
chain rule,
We assume that  
  is known
and  
  can simply
be found by differentiating eqn.2.1 wrt  
 . Thus we
get,
Eqns.2.2 and 2.3 define the backward pass. Note that
 
  can be further back propagated as usual.