Replies: 3 comments 4 replies
-
For numerous oversampling techniques it is certainly possible. There are a bunch of oversampling algorithms, e.g. SMOTE_PSO, etc., which do optimize the number of samples being generated. With these techniques it is up to the algorithm how many minority samples will be generated in the end. However, in many cases one can set the number of samples to be generated through the Namely, let N_min and N_maj denote the number of minority and majority samples, thus, the difference is N_maj - N_min. The proportion parameter specifies the number of samples to be generated in terms of this difference. Particularly, If you want to generate a certain number of samples, for example, 10 additional minority samples are desired, then you can set the proportion parameter to The |
Beta Was this translation helpful? Give feedback.
-
Sorry for asking, but in the API doc it is not noted which of the 80+ oversampling techniques supports
What if I want to 5x or 10x minority samples before using an undersampler? |
Beta Was this translation helpful? Give feedback.
-
That's a good point. I have just created a release (0.7.1) with an additional query function 'get_proportion_oversamplers' to get all oversampler classes with import smote_variants as sv
prop_oversamplers = sv.get_proportion_oversamplers() # list of all oversampler classes with proportion parameters Also please note that despite having a import smote_variants as sv
extensive_oversamplers = sv.get_multiclass_oversamplers() # the list of all oversamplers having a proportion parameter and only extending the set of minority samples (leaving the majority samples intact) Regarding the combination of oversamplers and filters, it is completely up to the user how he combines them. There are some oversampling techniques which inherently contain some noise filter (like To generate minority samples, say, M times the original N_min, one needs to set the proportion parameter to M*N_min/(N_maj - N_min). |
Beta Was this translation helpful? Give feedback.
-
SMOTE variants can be used with Undersamplers to speed up classification of imbalanced datasets. However oversampling normally precedes undersampling. Is it possible to generate minority samples that are less than the majority?
scikit-learn-contrib/imbalanced-learn#925
Beta Was this translation helpful? Give feedback.
All reactions