You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm Niels from the open-source team at Hugging Face. I discovered your work on MLGym through Hugging Face's daily papers: https://huggingface.co/papers/2502.14499. The paper page lets people discuss your work and find related artifacts. You can also claim the paper, which will show up on your public Hugging Face profile.
It'd be great to make the MLGYM-Bench benchmark available on the 🤗 hub, to improve its discoverability and visibility. We can add tags so that people can easily find it when searching datasets.
While MLGYM-Bench might not be a dataset in the traditional sense, it represents a curated set of tasks and evaluation metrics, which are essential for evaluating research agents. Hosting a structured description of MLGYM-Bench on Hugging Face, alongside the associated code on Github, would make it easier for others to understand, use, and contribute to the benchmark. This could include details on the tasks, evaluation metrics, and the datasets or models used within each task.
Would you be interested in hosting MLGYM-Bench on https://huggingface.co/datasets? Hosting on Hugging Face provides several benefits:
Increased visibility and discoverability: Make it easier for other researchers to find and use your benchmark.
Structured format: Provide a clear and organized way to present the benchmark details, including tasks, metrics, and associated resources.
Version control: Track changes and updates to the benchmark over time.
Community engagement: Foster discussion and collaboration around the benchmark.
Hi @deepakn97 🤗
I'm Niels from the open-source team at Hugging Face. I discovered your work on MLGym through Hugging Face's daily papers: https://huggingface.co/papers/2502.14499. The paper page lets people discuss your work and find related artifacts. You can also claim the paper, which will show up on your public Hugging Face profile.
It'd be great to make the MLGYM-Bench benchmark available on the 🤗 hub, to improve its discoverability and visibility. We can add tags so that people can easily find it when searching datasets.
While MLGYM-Bench might not be a dataset in the traditional sense, it represents a curated set of tasks and evaluation metrics, which are essential for evaluating research agents. Hosting a structured description of MLGYM-Bench on Hugging Face, alongside the associated code on Github, would make it easier for others to understand, use, and contribute to the benchmark. This could include details on the tasks, evaluation metrics, and the datasets or models used within each task.
Would you be interested in hosting MLGYM-Bench on https://huggingface.co/datasets? Hosting on Hugging Face provides several benefits:
Here's a guide on how to upload a dataset to Hugging Face: https://huggingface.co/docs/datasets/loading.
After uploading, we can link the dataset to the paper page (read here), further increasing its visibility.
Let me know if you're interested or need any help with this!
Cheers,
Niels
ML Engineer @ HF 🤗
The text was updated successfully, but these errors were encountered: