Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Release MLGYM-Bench dataset on Hugging Face #5

Open
NielsRogge opened this issue Feb 22, 2025 · 0 comments
Open

Release MLGYM-Bench dataset on Hugging Face #5

NielsRogge opened this issue Feb 22, 2025 · 0 comments

Comments

@NielsRogge
Copy link

Hi @deepakn97 🤗

I'm Niels from the open-source team at Hugging Face. I discovered your work on MLGym through Hugging Face's daily papers: https://huggingface.co/papers/2502.14499. The paper page lets people discuss your work and find related artifacts. You can also claim the paper, which will show up on your public Hugging Face profile.

It'd be great to make the MLGYM-Bench benchmark available on the 🤗 hub, to improve its discoverability and visibility. We can add tags so that people can easily find it when searching datasets.

While MLGYM-Bench might not be a dataset in the traditional sense, it represents a curated set of tasks and evaluation metrics, which are essential for evaluating research agents. Hosting a structured description of MLGYM-Bench on Hugging Face, alongside the associated code on Github, would make it easier for others to understand, use, and contribute to the benchmark. This could include details on the tasks, evaluation metrics, and the datasets or models used within each task.

Would you be interested in hosting MLGYM-Bench on https://huggingface.co/datasets? Hosting on Hugging Face provides several benefits:

  • Increased visibility and discoverability: Make it easier for other researchers to find and use your benchmark.
  • Structured format: Provide a clear and organized way to present the benchmark details, including tasks, metrics, and associated resources.
  • Version control: Track changes and updates to the benchmark over time.
  • Community engagement: Foster discussion and collaboration around the benchmark.

Here's a guide on how to upload a dataset to Hugging Face: https://huggingface.co/docs/datasets/loading.

After uploading, we can link the dataset to the paper page (read here), further increasing its visibility.

Let me know if you're interested or need any help with this!

Cheers,

Niels
ML Engineer @ HF 🤗

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant