-
Notifications
You must be signed in to change notification settings - Fork 221
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ResNet50 inference model seemed not support FP16 #173
Comments
@chenming22 we will work on adding |
@chenming22 we currently have AMP FP16 TF support based on the PR upstream here: tensorflow/tensorflow#62817. Since the script already supports FP16 input, installing |
thanks @sramakintel , i will have a try. |
hi, @sramakintel , noticed above pr been merged to tf2.16.1, so i install tf2.16.1 from pip and then ran inference.sh in ResNet50 V1.5 inference with precision=fp16, ONEDNN_MAX_CPU_ISA=AVX512_CORE_AMX_FP16 and model: resnet50_v1.pb, but log and metric on emon showed no AMX instructions used, any other environment variables or configurations needed? |
I want to evaluate inference performance using AMX-FP16 on new CPU while default docker or AI tool package not supported AMX-FP16 because of old version of onednn(3.2.0), so building onednn-3.3.0 and intel extensions for tensorflow-2.15.0 from source and using following model:
FP32, FP16 and BFloat32 Pretrained model:
howerver, log from onednn showed precision of this model was FP32 not FP16.
The text was updated successfully, but these errors were encountered: