-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CNN-RNNs not learning anything #39
Comments
https://docs.google.com/spreadsheets/d/1ZWiciyp-fRBzmOj5dewdBBrCnttDqPFTo8_iuCU8VF0/edit#gid=0&vpid=A1 a google spreadsheet of some of the runs |
This is subject depedant case of training. Will update the table with full detials once the run is finnished.
It behaves exactly the same as pure RNN aproach, will get solid data for that as well at some point. |
Hi Timotej, Sorry for the slow reply but I’ve been sick. Without changing more parameters it is a bit hard. Usually this is key to working with NN, nas therefore my recurrent suggestion to be document all the tests and parameters used. The sequences that the LSTM received are quite short, and sometimes LSTM do not work well that way. Also, you are using 200 units, which may be a little to much considering that you only have 34 input features. Try using 20. Is the RMSE computed on the standardised outputs? If not it actually looks quite low (error of 1.5 BPM is very low). Can you plot the predictions for training, validation and test? Cheers, From: Timotej Kapus <[email protected]mailto:[email protected]> @eadwardhttps://github.com/eadward I've now done the CNN-RNN approach discussed in the meeting. It is behaving very similarly to the subject independent RNN aproach, meaning it is not learning anything. For example here is a training trace: Train on 10721 samples, validate on 10410 samples #Final result on validation set: #Final result on test set: The loss (root mean square error in this case) keeps decreasing, but the training has no effect on the validation set performance. This is exactly the same thing as it is happening with the direct RNN approach. The architecture of the network is (tinkered keras syntax for clarity), shape comments represent the shape of the network at that point:
— |
Please send as attachment or add to github. From: Timotej Kapus <[email protected]mailto:[email protected]> https://docs.google.com/spreadsheets/d/1ZWiciyp-fRBzmOj5dewdBBrCnttDqPFTo8_iuCU8VF0/edit#gid=0&vpid=A1 a google spreadsheet of some of the runs — |
So, both RNN and CC+RNN work similarly for subject dependant, right? Make sure that you report this. Make sure that you train a model on a single subject (1 male and 1 female). In the meantime let us try to solve the independent folds issue: are you sure that there is no error in the data preparation? E. |
@eadward They also work similarly (bad) for subject independent model. I'm not sure there is no error in the data preparation, but have no idea how to test it. |
I assume that there no errors then, as you are using the same scripts for other tests. Have you checked the if you have a gender balanced folds and also that the HR distributions for the various folds are similar? Male/female unbalance could be a problem, and if the HR disctributions in each fold are very different the event worst.. E. |
@eadward I've now done the CNN-RNN approach discussed in the meeting.
It is behaving very similarly to the subject independent RNN aproach, meaning it is not learning anything.
For example here is a training trace:
The loss (root mean square error in this case) keeps decreasing, but the training has no effect on the validation set performance. This is exactly the same thing as it is happening with the direct RNN approach.
The architecture of the network is (tinkered keras syntax for clarity), shape comments represent the shape of the network at that point:
The text was updated successfully, but these errors were encountered: