You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Repro here: pytorch/pytorch#134560
It is mentioned in the original flex attention blog that dynamic shapes work should I be implementing it differently? My usecase is for document packing so varying batch size, block mask is recomputed at each step. Blocked by this for switching to flex attention unfortunately, thanks in advance.
The text was updated successfully, but these errors were encountered:
Repro here: pytorch/pytorch#134560
It is mentioned in the original flex attention blog that dynamic shapes work should I be implementing it differently? My usecase is for document packing so varying batch size, block mask is recomputed at each step. Blocked by this for switching to flex attention unfortunately, thanks in advance.
The text was updated successfully, but these errors were encountered: