We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Speeding up Inference with Quantized models
The text was updated successfully, but these errors were encountered:
Can I work on it? and we need to use dynamic quantization of pytorch right?
Sorry, something went wrong.
Of course! Thanks a lot! You could submit a PR!
PR created #21
What's the status of this? want to use quantized model on Windows
I am currently adding a feature so that you can use -q and you'll be able to use a quantised model for inference
No branches or pull requests
Speeding up Inference with Quantized models
The text was updated successfully, but these errors were encountered: