-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Method and Results #2
Comments
Hi, Thanks for reading this work. Yes, We apply Softmax on each row to normalize the learned noise matrix. Correct Labels or 0% noise, This model learns a very pessimistic noise model (an aggressive dropout). Therefore, we find this model not doing good on correctly labeled datasets. We have developed a new model to tackle all kinds and types of label noise and is under review. Hope this helps. |
In the paper you don't say that you apply the softmax to every row. You say quite the opposite, quote: |
You don't even say anything about the initialization of the weight matrix W. If you initialize it randomly, good luck with the convergence. |
Hi, I have read your paper with method description. I found this paper really interesting and have several theoretical questions. First of all, since the noise matrix is unconstrained (in our case it has usually negative values), how do you extract normalized (0,1) values that are described in figures. Do you apply softmax on each row? Secondly, how is your model behaving when applied to the correct labels without noise?
The text was updated successfully, but these errors were encountered: