From 1bc2fc13114e8a2869f5564034a10f74c1a1e949 Mon Sep 17 00:00:00 2001 From: Jack Zhang Date: Tue, 24 Jun 2025 17:03:42 -0700 Subject: [PATCH] Fix test_llm_config Reviewed By: GregoryComer Differential Revision: D77265031 --- examples/models/llama/config/test_llm_config.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/models/llama/config/test_llm_config.py b/examples/models/llama/config/test_llm_config.py index 0853e9dbbd8..2e8e0591184 100644 --- a/examples/models/llama/config/test_llm_config.py +++ b/examples/models/llama/config/test_llm_config.py @@ -94,7 +94,7 @@ def test_valid_llm_config(self): backend=BackendConfig( xnnpack=XNNPackConfig(enabled=False), coreml=CoreMLConfig( - enabled=True, ios=17, compute_units=CoreMLComputeUnit.ALL + enabled=True, ios=17, compute_units=CoreMLComputeUnit.cpu_only ), ), )