Skip to content

Commit 6065db8

Browse files
Update
[ghstack-poisoned]
2 parents 4fc8c25 + d740be0 commit 6065db8

File tree

1 file changed

+13
-13
lines changed

1 file changed

+13
-13
lines changed

test/prototype/attention/test_fp8_attention.py

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -20,20 +20,20 @@
2020
if torch_version_at_least("2.11.0"):
2121
from torchao.prototype.attention.utils import _is_fa3_available, _is_hopper
2222

23-
if _is_hopper() and _is_fa3_available():
24-
from torch.nn.attention import (
25-
activate_flash_attention_impl,
26-
restore_flash_attention_impl,
27-
)
23+
if _is_hopper() and _is_fa3_available():
24+
from torch.nn.attention import (
25+
activate_flash_attention_impl,
26+
restore_flash_attention_impl,
27+
)
2828

29-
from torchao.prototype.attention import (
30-
AttentionBackend,
31-
apply_low_precision_attention,
32-
)
33-
from torchao.prototype.attention.fp8_fa3.attention import (
34-
fp8_fa3_rope_sdpa,
35-
fp8_fa3_sdpa,
36-
)
29+
from torchao.prototype.attention import (
30+
AttentionBackend,
31+
apply_low_precision_attention,
32+
)
33+
from torchao.prototype.attention.fp8_fa3.attention import (
34+
fp8_fa3_rope_sdpa,
35+
fp8_fa3_sdpa,
36+
)
3737

3838

3939
def _rope_cos_sin(S, D, device):

0 commit comments

Comments
 (0)