-
Notifications
You must be signed in to change notification settings - Fork 13.7k
Closed
Labels
bug-unconfirmedhigh severityUsed to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)stale
Description
What happened?
Hi, Im trying to load GGUF model into transformers and I'm doing it like it is in example here https://huggingface.co/docs/transformers/en/gguf
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "thirteenbit/madlad400-10b-mt-gguf"
filename = "model-q8_0.gguf"
tokenizer = AutoTokenizer.from_pretrained(model_id, gguf_file=filename)
model = AutoModelForCausalLM.from_pretrained(model_id, gguf_file=filename)
inputs = tokenizer("<2pt> I love pizza!", return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
But I'm getting this error
Traceback (most recent call last):
File "....path.../translate-run-madlad.py", line 14, in <module>
tokenizer = AutoTokenizer.from_pretrained(model_id, gguf_file=filename)
File "......path....../lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 851, in from_pretrained
config_dict = load_gguf_checkpoint(gguf_path, return_tensors=False)["config"]
File "......path....../lib/python3.10/site-packages/transformers/modeling_gguf_pytorch_utils.py", line 83, in load_gguf_checkpoint
reader = GGUFReader(gguf_checkpoint_path)
File "......path....../lib/python3.10/site-packages/gguf/gguf_reader.py", line 94, in __init__
if self._get(offs, np.uint32, override_order = '<')[0] != GGUF_MAGIC:
File "......path....../lib/python3.10/site-packages/gguf/gguf_reader.py", line 151, in _get
.newbyteorder(override_order or self.byte_order)
AttributeError: `newbyteorder` was removed from the ndarray class in NumPy 2.0. Use `arr.view(arr.dtype.newbyteorder(order))` instead.
It looks like you just need to substitute newbyteorder by arr.view(arr.dtype.newbyteorder(order)) , thx
Name and Version
It's i think on all versions https://github.com/ggerganov/llama.cpp/blob/master/gguf-py/gguf/gguf_reader.py
What operating system are you seeing the problem on?
No response
Relevant log output
No response
MusaKavakcompilade and leszekhanusz
Metadata
Metadata
Assignees
Labels
bug-unconfirmedhigh severityUsed to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)stale