You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Sep 12, 2024. It is now read-only.
const model =path.resolve(process.cwd(), "../ggml-vicuna-7b-4bit-rev1.bin");
406
+
const model =path.resolve(process.cwd(), "../ggml-vicuna-7b-1.1-q4_1.bin");
370
407
371
408
const llama =newLLama(LLamaCpp);
372
409
@@ -452,5 +489,5 @@ The following steps will allow you to compile the binary with best quality on yo
452
489
- [ ] more platforms and cross compile (performance related)
453
490
- [ ] tweak embedding API, make end token configurable
454
491
- [ ] cli and interactive
455
-
- [ ] support more open source models as llama-rs planned https://github.com/rustformers/llama-rs/pull/85 https://github.com/rustformers/llama-rs/issues/75
492
+
- [ ] support more open source models as llama-rs planned https://github.com/rustformers/llama-rs/pull/141
0 commit comments