Skip to content

Commit 136cacb

Browse files
authored
Remove tensor parallel test for v1 of Int8DynamicActivationInt8WeightConfig (#4169)
Summary: CI is failiing: ``` The error occurs because torch.compile with DTensor generates wrapper code that tries to access tensor_impl on the outer LinearActivationQuantizedTensor instead of the inner AffineQuantizedTensor. This is a torch.compile subclass handling issue with the deeply nested subclass hierarchy in version 1: DTensor(LinearActivationQuantizedTensor(AffineQuantizedTensor(...))). The test uses Int8DynamicActivationInt8WeightConfig() with default version=1, which creates this nested structure. Version 2 uses Int8Tensor which avoids the problematic nesting. ``` v2 doesn't support view op so it won't work either. We haven't heard of a use case for this actually, it's used in some demos but didn't hear about real users yet. Therefore we delete the test in the PR, we'll be deleting v1 configs in the future as well. Test Plan: CI Reviewers: Subscribers: Tasks: Tags: [ghstack-poisoned]
1 parent 8fca033 commit 136cacb

1 file changed

Lines changed: 0 additions & 13 deletions

File tree

test/dtypes/test_affine_quantized_tensor_parallel.py

Lines changed: 0 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,6 @@
1818
Float8DynamicActivationFloat8WeightConfig,
1919
Float8WeightOnlyConfig,
2020
Int4WeightOnlyConfig,
21-
Int8DynamicActivationInt8WeightConfig,
2221
Int8WeightOnlyConfig,
2322
PerRow,
2423
PerTensor,
@@ -149,20 +148,8 @@ def test_tp(self, dtype):
149148
return self._test_tp(dtype)
150149

151150

152-
class TestInt8dqAffineQuantizedTensorParallel(TestAffineQuantizedTensorParallel):
153-
QUANT_METHOD_FN = staticmethod(Int8DynamicActivationInt8WeightConfig)
154-
COMMON_DTYPES = [torch.bfloat16]
155-
156-
@common_utils.parametrize("dtype", COMMON_DTYPES)
157-
@with_comms
158-
@unittest.skipIf(not torch.cuda.is_available(), "Need CUDA available")
159-
def test_tp(self, dtype):
160-
return self._test_tp(dtype)
161-
162-
163151
common_utils.instantiate_parametrized_tests(TestInt8woAffineQuantizedTensorParallel)
164152
common_utils.instantiate_parametrized_tests(TestInt4woAffineQuantizedTensorParallel)
165-
common_utils.instantiate_parametrized_tests(TestInt8dqAffineQuantizedTensorParallel)
166153

167154
# Float8 TP requires FP8-capable hardware (H100+ on CUDA, MI300+ on ROCm)
168155
from torchao.utils import is_MI300, is_MI350, is_sm_at_least_90

0 commit comments

Comments
 (0)