Week 20, 21 & 22 - Training ConvLSTM-4 for raw samples

1 minute read

Increase the number of layers

Once the best recurring strategy is obtained, we try to improve the results by increasing the number of layers, according to the conclusions reached in the modeled images. I have repeated the experiment with the modeled images, this time with raw images, to analyze the effect of increasing the number of layers on the network performance.

  • Sinusoidal dataset, 4 DOF, 80000 training samples, 10000 test samples
More layers-Sinusoidal 4 DOF

The best results are obtained with 4 layers ConvLSTM so it will be the structure used.

Proposed net

The new defined network is as follows:

ConvLSTM-4 Net

You can see that I have also reduced the number of previous convolutional layers since the server where the networks are trained has enough memory to analyze more values. Furthermore, its use could be reducing the spatial information too much, worsening the results.

Linear dataset results

  • 2 DOF, 80000 training samples, 10000 test samples
ConvLSTM-4-Linear 2 DOF

Parabolic dataset results

ConvLSTM-4-Parabolic 3 DOF

Sinusoidal dataset results

  • 2 DOF, 80000 training samples, 10000 test samples
ConvLSTM-4-Sinusoidal 2 DOF
  • 3 DOF, 80000 training samples, 10000 test samples
ConvLSTM-4-Sinusoidal 3 DOF
  • 4 DOF, 80000 training samples, 10000 test samples
ConvLSTM-4-Sinusoidal 4 DOF

Combined dataset results

  • Combined 33%-33%-34%, 80000 training samples, 10000 test samples
ConvLSTM-4-Combined

Conclusions

Although the results have improved in terms of mean relative error, there is still a prediction limit in the parabolic 2 DOF and a new limit in the sinusoidal 4 DOF. In addition, there are still high maximum values and a large number of outliers.