We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b58520d commit 28194a1Copy full SHA for 28194a1
test/unit/test_pytorch_xla.py
@@ -58,7 +58,7 @@ def num_gpus(instance_type):
58
@pytest.mark.parametrize("cluster_size", [1, 4])
59
class TestPyTorchXLARunner:
60
@patch.object(PyTorchXLARunner, "_check_compatibility")
61
- def test_setup(self, *patches):
+ def test_setup(self, cluster, cluster_size, master, instance_type, num_gpus, *patches):
62
for current_host in cluster:
63
rank = cluster.index(current_host)
64
print(f"Testing as host {rank+1}/{cluster_size}")
0 commit comments