Set up
- Huggingface CLI and access token
pip install huggingface-hub
- Install ollama
Steps
Let’s use TAIDE for example.
- Go to model page
https://huggingface.co/taide/TAIDE-LX-7B-Chat-4bit - Pull the model
huggingface-cli download taide/TAIDE-LX-7B-Chat-4bit --local-dir downloads --local-dir-use-symlinks False
- Create a
Modelfile
with the following contentModelfile FROM ~/Downloads/taide-7b-a.2-q4_k_m.gguf
- run
ollama create <your-model-name> -f Modelfile
- run the model using ollama as usual
Reference
- https://www.markhneedham.com/blog/2023/10/18/ollama-hugging-face-gguf-models/
- https://www.reddit.com/r/ollama/comments/1al30ut/how_to_run_gguf_model_in_ollama/