Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about int4 quantization #179

Open
gaikwadrahul8 opened this issue Nov 28, 2024 · 2 comments
Open

Questions about int4 quantization #179

gaikwadrahul8 opened this issue Nov 28, 2024 · 2 comments

Comments

@gaikwadrahul8
Copy link
Contributor

Hi, all.
Since it is an inquiry rather than an issue, I will not write a template.

Looking at the kernel side code of tflite, I saw that int4 for filter is supported in several op kernels; conv2d, depthwise-conv2d, fully-connected.

Could you tell me if there are plans to support int4 quantization in the tflite converter or support int4 for each op's input as well as filter, and if so, what milestones do you have?

Thank you :)

@gaikwadrahul8
Copy link
Contributor Author

This issue originally reported by @0-chan-kor has been moved to this dedicated repository for LiteRT to enhance issue tracking and prioritization. To ensure continuity, we have created this new issue on your behalf.

We appreciate your understanding and look forward to your continued involvement.

@pkgoogle
Copy link

pkgoogle commented Dec 2, 2024

Original Issue: tensorflow/tensorflow#60125

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants