We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 5c076fb commit bffeefeCopy full SHA for bffeefe
src/transformers/integrations/flex_attention.py
@@ -66,7 +66,7 @@ def __init__(self, training):
66
# cause errors. The suggested fix is to compile with "max-autotune-no-cudagraphs"
67
# see https://github.com/pytorch/pytorch/issues/146260 for training
68
self.training = training
69
- if _torch_version == "2.6.0" and training:
+ if _torch_version.split("+")[0] == "2.6.0" and training:
70
self._compiled_flex_attention = torch.compile(
71
flex_attention, dynamic=False, mode="max-autotune-no-cudagraphs"
72
)
0 commit comments