Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
[CPU] Fixed BF16 Matmul inference precision (openvinotoolkit#22994)
### Details: CPU plugin uses EnforceInferencePrecision routine for BF16 precision mark-up. Its logic assumes only activations precision is changed before Matmul op, while weights precision keeps w/o any changes. Since dnnlFCTypeMapping misses BF16 activation, FP32 weights optimized configuration for bf16, execution always happens in FP32 precision even user manually set infer_precision=bf16. This bug is not visible on FP16 IRs (since BF16+FP16 config is present), so only FP32 IRs affected. SInce `save_model` and ovc apply FP16 compression be default, the issue mostly applicable for pipelines which use a model directly after `convert_model` call.
- Loading branch information