-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Experiment with augmenting a higher percentage of the dataset #13
Comments
I have a suggestion in a related vein. I think we can simplify the structure of the code here and do away with the
The
(I hope I haven't got the math wrong --- please correct me if I did.) This structure of this code can thus be simplified as follows. Let |
The only thing to keep in mind here is that for non-augmented datapoints, the feature values are cached, so I believe your approach might result in a performance penalty. Tho tbh I'm guessing that doing more augmentation will outweigh the costs, but just a note |
I'm not sure if this helps with making sure the caching works for the non-agumented data points, but we could also set p to 1 for the transforms, and adjust the amount of augmentation by keeping My main thought is just that stacking a probabilistic thing on top of another probabilistic thing makes things harder to reason about --- it would be clearer if we either removed the random.uniform stuff or made the transforms deterministic by setting p = 1. |
Today, it appears that augmentation is only occurring 10% of the time
parrot.py/lib/audio_dataset.py
Lines 69 to 72 in 5b57d12
The text was updated successfully, but these errors were encountered: