-
Notifications
You must be signed in to change notification settings - Fork 439
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Moe merging failed #452
Comments
It looks like you're using the In addition, this particular merge probably won't work - the two models you are looking at aren't the same size, so they will not be compatible. |
I hope this message finds you well. Specifically, I have the following models: Base Model: CMLM/ZhongJing-2-1_8b Current Challenge: When attempting to merge these models using a YML configuration, I continue encounter the error. Could you provide an example of a correctly structured YML file for merging these models? Despite following available guidelines, attempts to merge via your space result in errors. Attempted Configuration: Here's the YML configuration I used: yml
Thank you very much for your time and assistance. I look forward to your guidance to resolve this merging issue effectively. |
I encountered an error while trying to merge two Qwen-based lora models using a mixture of experts (MoE) configuration with qwen architecture. I’m working with a phi2_moe2.yml configuration file, but the system throws an error related to a missing field (merge_method).
Configuration and Setup
I am using the following configuration yml:
When I run this setup, I get the following error:
Attempted Solutions
I suspect adding merge_method might resolve the issue, but I’m not sure what options are available for this field. I would appreciate guidance on:
Complete yml file for qwen moe merge_method
Documentation or examples: Are there any detailed examples or documentation that explain each field in the YAML configuration for MoE?
Additional Context
First model: CMLM/ZhongJing-2-1_8b
Second model: Qwen2.5-1.5B-Instruct
Thank you for your assistance!
The text was updated successfully, but these errors were encountered: