-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Measuring and benchmarking the safety of the fine-tuned models #5
Comments
Thanks for bringing this up! I think this is a worthwhile exercise for us to try and evaluate this benchmark. Looks like the benchmark is still in POC, but they have a repo with steps outlined on how to test it out: https://github.com/mlcommons/modelbench |
Is this something you might have the bandwidth to try/look into @fcanogab? |
we might also look at unitxt (an ibm open source project) Jonathan Bnayahu has added some safety related benchmarks and others, see this search for list: |
@hemajv, yes, I would like to try to work on this myself. Thanks for the hint @erikerlandson. I'll take a look at it. |
There are different frameworks to measure and benchmark against other models the safety/harmfulness of a fine-tuned model. For example, MLCommons defines a framework that can be used for this.
The text was updated successfully, but these errors were encountered: