Skip to content

Commit 89eef8c

Browse files
authored
Merge pull request #158 from codelion/codelion-patch-1
Update README.md
2 parents cf9523a + 1868b11 commit 89eef8c

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -216,7 +216,7 @@ response = client.chat.completions.create(
216216
- e.g. for llama.cpp, run `python3 optillm.py --base_url http://localhost:8080/v1`
217217

218218
> [!WARNING]
219-
> Note that the Anthropic API, llama-server (and ollama) currently does not support sampling multiple responses from a model, which limits the available approaches to the following:
219+
> The Anthropic API, llama.cpp-server, and ollama currently do not support sampling multiple responses from a model, which limits the available approaches to the following:
220220
> `cot_reflection`, `leap`, `plansearch`, `rstar`, `rto`, `self_consistency`, `re2`, and `z3`. For models on HuggingFace, you can use the built-in local inference server as it supports multiple responses.
221221
222222
## Implemented techniques

0 commit comments

Comments
 (0)