-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Print initial validation loss + final validation loss #1228
Conversation
I greatly simplified this @carmocca . Let me know if that's ok now. It should introduce any changes to the defaults. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM minus the adapter copy-paste accidents
I originally had that but others found it was misleading because it's calculated only on 2 batches and is not reliable. I then replaced it with the full loss calculation but then this was too slow. So the new argument is a compromise so that I can enable it for me but it doesn't slow it down for everyone. |
EDIT: After Adrian already added the final validation, I greatly simplified this PR. So this only adds the initial validation. Since there was not that much enthusiasm from others, I made it optional and disabled it by default. I think it's a super useful feature to be able to know what the initial val_loss of the model is so that one can compare it to the final one and be able to tell if/how much things improved after training.