Skip to content

Commit f7ad745

Browse files
authored
Removed llama-cpp-python as it still doesn't support sampling multiple responses
1 parent e703e04 commit f7ad745

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ python optillm.py
6969
> [!WARNING]
7070
> Note that llama-server currently does not support sampling multiple responses from a model, which limits the available approaches to the following:
7171
> `cot_reflection`, `leap`, `plansearch`, `rstar`, `rto`, `self_consistency`, and `z3`.
72-
> In order to use other approaches, consider using an alternative compatible server such as [ollama](https://github.com/ollama/ollama) or [llama-cpp-python](https://github.com/abetlen/llama-cpp-python).
72+
> In order to use other approaches, consider using an alternative compatible server such as [ollama](https://github.com/ollama/ollama).
7373
7474
> [!NOTE]
7575
> You'll later need to specify a model name in the OpenAI client configuration. Since llama-server was started with a single model, you can choose any name you want.

0 commit comments

Comments
 (0)