Skip to content

Commit

Permalink
ruff
Browse files Browse the repository at this point in the history
  • Loading branch information
jerryzh168 committed Dec 7, 2024
1 parent f3a8e8d commit 2895626
Showing 1 changed file with 4 additions and 2 deletions.
6 changes: 4 additions & 2 deletions test/dtypes/test_affine_quantized_float.py
Original file line number Diff line number Diff line change
Expand Up @@ -288,8 +288,10 @@ def test_fp8_weight_dimension_warning(self):
)


@unittest.skip("Only running locally so we don't need to add installation of fa3 "
"hopper kernels to CI, we'll probably copy paste kernel in the future")
@unittest.skip(
"Only running locally so we don't need to add installation of fa3 "
"hopper kernels to CI, we'll probably copy paste kernel in the future"
)
class TestAffineQuantizedFloat8Attention(common_utils.TestCase):
@unittest.skipIf(not torch.cuda.is_available(), "Need CUDA available")
def test_float8_attention(self):
Expand Down

0 comments on commit 2895626

Please sign in to comment.