Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

about QAT model #42

Open
lxzatwowone1 opened this issue Sep 25, 2024 · 1 comment
Open

about QAT model #42

lxzatwowone1 opened this issue Sep 25, 2024 · 1 comment

Comments

@lxzatwowone1
Copy link

I have a question, Why need qat model to convert dla model, can I use original onnx model?
Thanks! I am DL beginner

@liuanqi-libra7
Copy link
Collaborator

liuanqi-libra7 commented Nov 29, 2024

Hi @lxzatwowone1

  • DLA's INT8 compute capability is much higher than float, we recommend that the model be inferred with INT8 precision on DLA.
  • The accuracy of implicit quantization using TensorRT will be lower than PyTorch Quantization, so we recommend using PyTorch Quantization to quantize the model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants