Skip to content

Commit 8f5679d

Browse files
committed
add extra and flash attention options to readme
1 parent b37684a commit 8f5679d

File tree

1 file changed

+3
-0
lines changed

1 file changed

+3
-0
lines changed

README.md

+3
Original file line numberDiff line numberDiff line change
@@ -345,6 +345,7 @@ If you have loaded a model locally you need to set its URL through the expanded
345345

346346
- `Show/Hide Advanced Options` Toggle to show/hide advanced options from below
347347
- `Log Level` select how verbose the log messages are
348+
- `Use extras` select to install and allow the use of extra features (flash attention and IQ quants)
348349

349350
#### 💻 Setup Settings
350351

@@ -381,6 +382,7 @@ If the user's GPU is not supported, the LLM will fall back to the CPU
381382
- `Model` the path of the model being used (relative to the Assets/StreamingAssets folder)
382383
- `Chat Template` the chat template being used for the LLM
383384
- `Lora` the path of the LoRA being used (relative to the Assets/StreamingAssets folder)
385+
- `Flash Attention` click to use flash attention in the model (if `Use extras` is enabled)
384386

385387
</details>
386388

@@ -395,6 +397,7 @@ If the user's GPU is not supported, the LLM will fall back to the CPU
395397

396398
- `Show/Hide Advanced Options` Toggle to show/hide advanced options from below
397399
- `Log Level` select how verbose the log messages are
400+
- `Use extras` select to install and allow the use of extra features (flash attention and IQ quants)
398401

399402
#### 💻 Setup Settings
400403
<div>

0 commit comments

Comments
 (0)