Skip to content

Different training size for benchmarking #69

Answered by janosh
knc6 asked this question in Q&A
Discussion options

You must be logged in to vote

We ultimately don't care about the typical strict architecture comparison found in other ML benchmarks. We care about measuring how good ML (any form of ML) is at OOD materials stability prediction. If some models (interatomic potentials) are trained on forces and therefore can leverage more of the maximum training set released with our benchmark (the entirety of the MP v2022.10.28 database version) then that's a genuine advantage of force-full models for the real-world application we care about and we want our benchmark to reflect that.

In short, we want to provide a walled garden for asking system-level questions which a traditional ML benchmark is too rigid to answer. I believe we succ…

Replies: 3 comments 14 replies

Comment options

You must be logged in to vote
0 replies
Answer selected by janosh
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
14 replies
@ml-evs
Comment options

@CompRhys
Comment options

@JonathanSchmidt1
Comment options

@JonathanSchmidt1
Comment options

@CompRhys
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
discussion Needs discussion help Q&A support issues
5 participants
Converted from issue

This discussion was converted from issue #68 on December 03, 2023 17:51.