Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

WIP: Granite model updates, and questions about best use of the MOF #66

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

obuzek
Copy link

@obuzek obuzek commented Jan 29, 2025

Hi 👋🏻 - thanks for all the work that's gone into this. There's clearly been a lot of thought about what makes a model open.

I'm trying to update the information regarding the IBM Granite models, and I'm finding that I have several thoughts along the way about how the MOT can most effectively represent the considerable work that we and others have put into openness.

This is a WIP PR, not intended to be committed as-is; I do not think it currently conforms to the schema.

(Let me know if this should be resolved as an issue / discussion forum before creating the PR. I thought it was best to share the questions alongside my attempts to answer within framework.)

High-level queries

  1. Many of the entries, such as those related to evaluation and datasets, have partial-to-full answers posted on the HuggingFace model card. In order to simplify updates, would it be possible to scrape this information from the HF model card? Training datasets and evaluation metrics are represented in the YAML at the top of the Granite HF README.md.

  2. How does the MOT want us to handle versioning? I see the existing Granite files are split based on number of parameters - but there are a plethora of models to date, with different architectural versions, fine-tuning targets, modalities, and number of parameters.

For at least the Granite family of models, the most appropriate distinction would be on architectural versions, fine-tuning targets, and modalities, and openness is likely not to vary according to the number of parameters.

  1. Several OSS frameworks exist to perform inference on models with open weights, and Granite is compatible with many of them. How can we best represent that information?

  2. How can we represent ongoing updates to transparency in the form of further research papers, technical reports, presentations, etc?


More specific questions are included in the PR itself.

Thanks! Excited to see this project continue.

@lehors
Copy link
Collaborator

lehors commented Feb 4, 2025

Thanks @obuzek for some thoughtful questions. I can't claim to have all the answers so I suggest you bring them up separately in the Discussions section so we can talk about them and figure out the best way to address them.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants