Replies: 1 comment
-
you just use the -hf switch and it will check in cache and load downloaded file. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello 👋🏻 Just started with llama.cpp and have a couple of questions. I downloaded a model using the
llama-cli -hf
like this according the instructions on the unsloth site.It runs, I can interact with it. But now I want to run it using
llama-server
to use it over http requests. I tried a couple of options with the model name but non of the worked.llama-server -m unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF:Q4_K_XL llama-server -m unsloth_Qwen3-Coder-30B-A3B-Instruct-GGUF_Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf # model name from the .cache/llama.cpp folder
Getting this error:
.cache/llama.cpp/
contentsHow do I find proper model name/path for
llama-server -m
command?Can I change default model folder?
Can I list model names using
llama-cli
to use the name for thellama-server -m
command?Beta Was this translation helpful? Give feedback.
All reactions