You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+3-3
Original file line number
Diff line number
Diff line change
@@ -248,7 +248,7 @@ public class MyScript : MonoBehaviour
248
248
// Otherwise the model file can be copied directly inside the StreamingAssets folder.
249
249
llm.SetModel("Phi-3-mini-4k-instruct-q4.gguf");
250
250
// optional: you can also set a lora in a similar fashion
251
-
llm.SetLora("my-lora.bin");
251
+
llm.SetLora("my-lora.gguf");
252
252
// optional: you can set the chat template of the model if it is not correctly identified
253
253
// You can find a list of chat templates in the ChatTemplate.templates.Keys
254
254
llm.SetTemplate("phi-3");
@@ -374,8 +374,8 @@ If the user's GPU is not supported, the LLM will fall back to the CPU
374
374
375
375
- <details><summary>Advanced options</summary>
376
376
377
-
-`Download lora` click to download a LoRA model in .bin format
378
-
-`Load lora` click to load a LoRA model in .bin format
377
+
-`Download lora` click to download a LoRA model in .gguf format
378
+
-`Load lora` click to load a LoRA model in .gguf format
379
379
- <details><summary><code>Context Size</code> size of the prompt context (0 = context size of the model)</summary> This is the number of tokens the model can take as input when generating responses. Higher values use more RAM or VRAM (if using GPU). </details>
380
380
-`Batch Size` batch size for prompt processing (default: 512)
381
381
-`Model` the path of the model being used (relative to the Assets/StreamingAssets folder)
0 commit comments