Skip to content

Commit

Permalink
refine pytest.skip message
Browse files Browse the repository at this point in the history
  • Loading branch information
Xia-Weiwen committed Apr 19, 2024
1 parent 92900f6 commit 717245d
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions tests/test_functional.py
Original file line number Diff line number Diff line change
Expand Up @@ -932,7 +932,7 @@ def test_colrow_absmax(dim1, dim2, dims):
@pytest.mark.parametrize("dtype", [torch.half, torch.bfloat16], ids=id_formatter("dtype"))
def test_double_quant(dim1, dim2, device, dtype):
if device == "cuda" and dtype == torch.bfloat16:
pytest.skip("BFloat16 not supported on CUDA")
pytest.skip("bfloat16 is not implemented for this operation on CUDA backend")
for i in range(k):
A = torch.randn(dim1, dim2, device=device).to(dtype)
out_col1, Scol = F.vectorwise_quant(A, dim=0)
Expand Down Expand Up @@ -1212,7 +1212,7 @@ def test_overflow():
@pytest.mark.parametrize("dtype", [torch.half, torch.bfloat16], ids=id_formatter("dtype"))
def test_coo_double_quant(dim1, dim2, device, dtype):
if device == "cuda" and dtype == torch.bfloat16:
pytest.skip("BFloat16 not supported on CUDA")
pytest.skip("bfloat16 is not implemented for this operation on CUDA backend")
threshold = 3.00
for i in range(k):
A = torch.randn(dim1, dim2, device=device).to(dtype)
Expand Down

0 comments on commit 717245d

Please sign in to comment.