-
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training fixes #21
Training fixes #21
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good, nice work!
@@ -164,7 +163,7 @@ def _get_premade_batches_datapipe(self, subdir, shuffle=False, add_filename=Fals | |||
file_pipeline = FileLister(f"{self.batch_dir}/{subdir}", masks="*.pt", recursive=False) | |||
|
|||
if shuffle: | |||
file_pipeline = file_pipeline.shuffle(buffer_size=1000) | |||
file_pipeline = file_pipeline.shuffle(buffer_size=10_000) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Out of interest, what's the reason for the buffer size change here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No particular reason. Just that at this stage all we'll being doing is loading up to 1000 or 10,000 file names before shuffling them. I thought that since the filenames don't take up any amount of space why not use more and have a more complete shuffle.
This pull request includes:
relative_scale_pvnet_outputs
option, out input values to the network were of order 1E-4Note that this will break the compatibility with already trained summation models