XLMRoberta with Flash Attention 2 #27957
Labels
Feature request
Request for a new feature
Good Second Issue
Issues that are more difficult to do than "Good First" issues - give it a try if you want!
System Info
Who can help?
@ArthurZucker @younesbelkada
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
from transformers import AutoModelForSequenceClassification
model = AutoModelForSequenceClassification.from_pretrained("my_model/", attn_implementation="flash_attention_2")
Expected behavior
Ability to use flash attention 2 for inference. Is it possible to add support of flash attention 2 for XLMRoberta model?
The text was updated successfully, but these errors were encountered: