Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

SVDquant W8A8 #45

Open
Manojbhat09 opened this issue Feb 8, 2025 · 1 comment
Open

SVDquant W8A8 #45

Manojbhat09 opened this issue Feb 8, 2025 · 1 comment

Comments

@Manojbhat09
Copy link

Hello!
I appreciate your great work on diffusion models.

For experimenting on W8A8 models, as I could see there is no config present in https://github.com/mit-han-lab/deepcompressor/tree/main/examples/diffusion/configs/svdquant
whereas on the results its reported with this performance:

Precision | Method | FID (↓) | IR (↑) | LPIPS (↓) | PSNR( ↑)
INT W8A8 | Ours | 16.3 | 0.955 | 0.109 | 23.7

here: https://github.com/mit-han-lab/deepcompressor/tree/main/examples/diffusion
Could you please elaborate how to generate the DiT for W8A8 precision as its not published in huggingface?

@synxlin
Copy link
Contributor

synxlin commented Feb 14, 2025

Hi,

For W8A8 configuration, you may directly change the quantization configuration to:

    dtype: sint8
    group_shapes:
    - - 1
      - -1
      - 1
      - 1
      - 1
    scale_dtypes:
    - null

for both wgts (weights) and ipts (input activations).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants