Help needed: strange results with Google Colab
Closed this issue · 1 comments
Hi!
In our research project we started to work with the KalmanNet. As a first step, we only tried the simple linear case using the 2x2 linear system (2x2_020_T100.pt). For this, we imported the following files to Google Colab and ran a learning and testing sequence:
-
KalmanNet_nn.py
-
Linear_KF.py
-
Linear_sysmdl.py
-
Extended_data.py
-
Pipeline_KF.py
-
KalmanFilter_test.py
-
main_linear.py
-
Plot.py
We found the following behaviour: using the same code as in the GitHub repository and the same files, in some cases the results look rather promising:
But in other cases – without any modifications applied – the results look like this:
The strange thing is that the system, the data, the hyperpaparmeters and the length of the learning sequence are the same, however the results seem to converge to different values: sometimes to -7db/-8dB and sometimes to 0dB.
Unfortunately we haven’t come up with an explanation so far, so we would be much obliged if you could give us a helping hand what might went wrong in our side.
Thanks in advance!
This is possibly because that KalmanNet has been trained to converge to some local optimal. One solution to this is try to use larger weight decay and larger numbers of training epoches.