From 11a650e0da59b6b3e55defbcbba7ade7b8d59ea2 Mon Sep 17 00:00:00 2001 From: helunwencser Date: Tue, 29 Oct 2024 12:35:22 -0700 Subject: [PATCH] Update [ghstack-poisoned] --- examples/models/llama/evaluate/eager_eval.py | 6 +++++- examples/models/llama/install_requirements.sh | 2 +- 2 files changed, 6 insertions(+), 2 deletions(-) diff --git a/examples/models/llama/evaluate/eager_eval.py b/examples/models/llama/evaluate/eager_eval.py index e931c60c184..7d155f5f0e7 100644 --- a/examples/models/llama/evaluate/eager_eval.py +++ b/examples/models/llama/evaluate/eager_eval.py @@ -31,7 +31,7 @@ def __init__( use_kv_cache: bool = False, ): device = "cuda" if torch.cuda.is_available() else "cpu" - super().__init__(device=device) + super().__init__(device=device, pretrained="gpt2") self._model = model self._tokenizer = tokenizer self._device = torch.device(device) @@ -47,6 +47,10 @@ def eot_token_id(self): return self._tokenizer.eot_id return self._tokenizer.eos_id + @property + def prefix_token_id(self): + return self._tokenizer.eos_id + @property def max_length(self): return self._max_seq_length diff --git a/examples/models/llama/install_requirements.sh b/examples/models/llama/install_requirements.sh index 470e804c2d4..3103daeb7d2 100755 --- a/examples/models/llama/install_requirements.sh +++ b/examples/models/llama/install_requirements.sh @@ -15,7 +15,7 @@ pip install --no-use-pep517 "git+https://github.com/pytorch/ao.git@${TORCHAO_VER # Install lm-eval for Model Evaluation with lm-evalution-harness # Install tiktoken for tokenizer -pip install lm_eval==0.4.2 +pip install lm_eval==0.4.5 pip install tiktoken blobfile # Call the install helper for further setup