Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Impact of batch size on training performance #55

Open
XingtongGe opened this issue Dec 18, 2024 · 1 comment
Open

Impact of batch size on training performance #55

XingtongGe opened this issue Dec 18, 2024 · 1 comment

Comments

@XingtongGe
Copy link

Hi, I would like to know how much the batch size affects training performance. For example, is there a significant performance gap between training DMD2-SDXL with 8 GPUs (total batch size of 16) and training with 64 GPUs (total batch size of 128)? Thanks!

@tzhu01
Copy link

tzhu01 commented Jan 16, 2025

I followed his ReadMe and trained using an 8 GPUs (A100, 80G), and the results were almost identical to the data in the paper.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants