-
Notifications
You must be signed in to change notification settings - Fork 3.5k
Overfit batches parameter gives a validation batch #15021
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I had the same problem. I was going crazy because in the documentation they supposed to be the same 😅. |
Same here |
@Borda |
@nilsleh Indeed the video snippet suggest that their identical which IMO should't, see #20731 (comment). With regards to the documentation (version # default used by the Trainer
trainer = Trainer(overfit_batches=0.0)
# use only 1% of the train & val set
trainer = Trainer(overfit_batches=0.01)
# overfit on 10 of the same batches <--- This seems confusing
trainer = Trainer(overfit_batches=10) I think the last comment is a little bit confusing since it gives the idea that the same 10 batches are used for both training and validation. Maybe it should read as |
Bug description
When overfitting on a single batch and defining dataloaders in class, the batch provided to the validation step is different from the batch on the training step. I was told in the slack community that this is NOT the intended behaviour.
How to reproduce the bug
The val batch is the [10,11] tensor, the train batch is the [0,1] tensor

The text was updated successfully, but these errors were encountered: