Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[QST] int8 Conv2D for volta V100 #1964

Open
sycz00 opened this issue Nov 26, 2024 · 0 comments
Open

[QST] int8 Conv2D for volta V100 #1964

sycz00 opened this issue Nov 26, 2024 · 0 comments

Comments

@sycz00
Copy link

sycz00 commented Nov 26, 2024

What is your question?

Hey I use following script to emit the cuda kernel for the Nvidia V100:

import torch
import random

import cutlass

dtype = torch.int32
type_A = torch.int8
type_B = torch.int8
type_C = torch.int32
type_D = torch.int32


plan = cutlass.Conv2dFprop(element=dtype, element_input=type_A, element_weight=type_B, element_C=type_C ,element_output=type_D, element_accumulator=type_D)
op = plan.construct()
conv_layer = cutlass.emit.pytorch(op, name='conv_layer', cc=plan.cc, sourcedir='conv', jit=True)

I tried a couple of configuration datatypes. Float works just fine. But I want them to be integer-8 Input and weights. And it should be possible with idp4a instruction set, right ?
Any ideas ? - thanks in advance !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant