Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Measuring and benchmarking the safety of the fine-tuned models #5

Open
fcanogab opened this issue Jun 21, 2024 · 4 comments
Open

Measuring and benchmarking the safety of the fine-tuned models #5

fcanogab opened this issue Jun 21, 2024 · 4 comments

Comments

@fcanogab
Copy link

There are different frameworks to measure and benchmark against other models the safety/harmfulness of a fine-tuned model. For example, MLCommons defines a framework that can be used for this.

@hemajv
Copy link
Collaborator

hemajv commented Jul 9, 2024

Thanks for bringing this up! I think this is a worthwhile exercise for us to try and evaluate this benchmark. Looks like the benchmark is still in POC, but they have a repo with steps outlined on how to test it out: https://github.com/mlcommons/modelbench

@hemajv
Copy link
Collaborator

hemajv commented Jul 9, 2024

Is this something you might have the bandwidth to try/look into @fcanogab?

@hemajv hemajv moved this to Ready in Data Science WG Jul 12, 2024
@erikerlandson
Copy link
Collaborator

we might also look at unitxt (an ibm open source project)

Jonathan Bnayahu has added some safety related benchmarks and others, see this search for list:

https://github.com/IBM/unitxt/issues?q=author%3Abnayahu+

@fcanogab
Copy link
Author

fcanogab commented Jul 22, 2024

@hemajv, yes, I would like to try to work on this myself.

Thanks for the hint @erikerlandson. I'll take a look at it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: Ready
Development

No branches or pull requests

3 participants