Spaces:
Running
on
A10G
Unable to convert Phi 3 small to GGUF
I get the following error when running this space -
Error: Error converting to fp16: b'INFO:hf-to-gguf:Loading model: Phi-3-small-8k-instruct\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2585, in \n main()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2563, in main\n model_class = Model.from_model_architecture(hparams["architectures"][0])\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 370, in from_model_architecture\n raise NotImplementedError(f'Architecture {arch!r} not supported!') from None\nNotImplementedError: Architecture 'Phi3SmallForCausalLM' not supported!\n'
I got the same (similar) error today on a different model.
Hi @aayushg159 - I just tried it, it works for me: https://huggingface.co/reach-vb/Phi-3-mini-4k-instruct-Q8_0-GGUF
Closing this as fixed! (feel free to re-open a new issue)
Uh, I mentioned Phi-3 small model (link) not mini. Phi-3 mini works as expected. I just ran it again and got the following error -
Error: Error converting to fp16: b'INFO:hf-to-gguf:Loading model: Phi-3-small-8k-instruct\nTraceback (most recent call last):\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2777, in \n main()\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 2755, in main\n model_class = Model.from_model_architecture(hparams["architectures"][0])\n File "/home/user/app/llama.cpp/convert-hf-to-gguf.py", line 369, in from_model_architecture\n raise NotImplementedError(f'Architecture {arch!r} not supported!') from None\nNotImplementedError: Architecture 'Phi3SmallForCausalLM' not supported!\n'