-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
n:m:g sparse format #3
Comments
We have just released our implementation, and you can view an example of how to use it here: Lines 6 to 30 in f2a5aa0
|
It seems that ur code is baed on CPU. I tried to time line 30 and benchmark it against output = model(input), the dense version. It seems a lot slower. In addition, how can I make it to GPU? calling .cuda() on model and input? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hi, thanks for open-sourcing the great work, which is very helpful for sparse deep learning workloads. I notice there is a
𝑛:𝑚:𝑔
sparsity layout in your paper, but I could not find theGroupedNMSparsifier
class in this repository. Could you kindly point me to that implementation?You also mentioned "CPU implementations for
𝑛:𝑚:𝑔
sparsity were compiled with GCC 8.4", but it seems this repository only contains the Python code. Will you release the kernel implementation later?The text was updated successfully, but these errors were encountered: