Remove unnecessary _prepare_decoder_attention_mask
patching
#1461
Labels
onnx
Related to the ONNX export
_prepare_decoder_attention_mask
patching
#1461
Feature request
Some patching of transformers
_prepare_decoder_attention_mask
was introduced in #1257, which can be avoided simply by exporting with a sequence length > 1.See
optimum/optimum/exporters/onnx/model_patcher.py
Line 405 in e7bd60d
Motivation
Code simplification
Your contribution
/
The text was updated successfully, but these errors were encountered: