how to avoid online downing while i have got the offline wight files #2660
Unanswered
arctanbell
asked this question in
Q&A
Replies: 2 comments 1 reply
-
You can also use the local path of a model file, which can be ran by llama-cpp-python. chat = ops.LLM.Llama_2({YOUR_LOCAL_FILE_PATH}, n_ctx=4096, max_tokens=200) |
Beta Was this translation helpful? Give feedback.
1 reply
-
Install the previous version before this one: ggml-org/llama.cpp@dadbed9 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I want to] use the llm.llama2, but downloading files named like


so I downloaded the files llama-2-7b-chat.ggmlv3.q4_0 and llama-2-7b-chat.ggmlv3.q4_1 and placed them at the same dir, however the download still continued
Beta Was this translation helpful? Give feedback.
All reactions