You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+66-6
Original file line number
Diff line number
Diff line change
@@ -35,11 +35,71 @@ With `xTuring` you can,
35
35
36
36
## 🌟 What's new?
37
37
We are excited to announce the latest enhancements to our `xTuring` library:
38
-
1.__`Falcon LLM` integration__ - You can use and fine-tune the _`Falcon-7B`_ model in different configurations: _off-the-shelf_, _off-the-shelf with INT8 precision_, _LoRA fine-tuning_, and _LoRA fine-tuning with INT8 precision_.
39
-
2.__`GenericModel` wrapper__ - This new integration allows you to test and fine-tune any new model on `xTuring` without waiting for it to be integrated using class _`GenericModel`_.
38
+
1.__`LLaMA 2` integration__ - You can use and fine-tune the _`LLaMA 2`_ model in different configurations: _off-the-shelf_, _off-the-shelf with INT8 precision_, _LoRA fine-tuning_, _LoRA fine-tuning with INT8 precision_ and _LoRA fine-tuning with INT4 precision_ using the `GenericModel` wrapper and/or you can use the `Llama2` class from `xturing.models` to test and finetune the model.
39
+
```python
40
+
from xturing.models import Llama2
41
+
model = Llama2()
42
+
43
+
## or
44
+
from xturing.models import BaseModel
45
+
model = BaseModel.create('llama2')
46
+
47
+
```
48
+
2.__`Evaluation`__ - Now you can evaluate any `Causal Language Model` on any dataset. The metrics currently supported is [`perplexity`](https://towardsdatascience.com/perplexity-in-language-models-87a196019a94).
4.__CPU inference__ - Now you can use just your CPU for inference of any LLM. _CAUTION : The inference process may be sluggish because CPUs lack the required computational capacity for efficient inference_.
83
+
5.__Batch integration__ - By tweaking the 'batch_size' in the .generate() and .evaluate() functions, you can expedite results. Using a 'batch_size' greater than 1 typically enhances processing efficiency.
An exploration of the [Llama LoRA INT4 working example](examples/int4_finetuning/LLaMA_lora_int4.ipynb) is recommended for an understanding of its application.
40
101
41
-
You can check the [Falcon LoRA INT8 working example](examples/falcon/falcon_lora_int8.py) repository to see how it works.
42
-
Also, you can check the [GenericModel working example](examples/generic/generic_model.py) repository to see how it works.
102
+
For an extended insight, consider examining the [GenericModel working example](examples/generic/generic_model.py) available in the repository.
43
103
44
104
<br>
45
105
@@ -170,8 +230,8 @@ model = BaseModel.load("x/distilgpt2_lora_finetuned_alpaca")
170
230
-[x] INT4 LLaMA LoRA fine-tuning with INT4 generation
171
231
-[x] Support for a `Generic model` wrapper
172
232
-[x] Support for `Falcon-7B` model
173
-
-[X] INT4 low-precision fine-tuning support
174
-
-[] Evaluation of LLM models
233
+
-[x] INT4 low-precision fine-tuning support
234
+
-[x] Evaluation of LLM models
175
235
-[ ] INT3, INT2, INT1 low-precision fine-tuning support
Copy file name to clipboardExpand all lines: docs/docs/intro.md
+7-4
Original file line number
Diff line number
Diff line change
@@ -39,13 +39,16 @@ You can quickly get started with xTuring by following the [Quickstart](/quicksta
39
39
40
40
| Model | Examples |
41
41
| --- | --- |
42
-
| LLaMA |[LLaMA 7B fine-tuning on Alpaca dataset with/without LoRA and with/without INT8](https://github.com/stochasticai/xturing/tree/main/examples/llama)|
42
+
| Bloom |[Bloom fine-tuning on Alpaca dataset with/without LoRA and with/without INT8](https://github.com/stochasticai/xturing/tree/main/examples/bloom)|
43
+
| Cerebras-GPT |[Cerebras-GPT fine-tuning on Alpaca dataset with/without LoRA and with/without INT8](https://github.com/stochasticai/xturing/tree/main/examples/cerebras)|
44
+
| Falcon |[Falcon 7B fine-tuning on Alpaca dataset with/without LoRA and with/without INT8](https://github.com/stochasticai/xturing/tree/main/examples/falcon)|
45
+
| Galactica |[Galactica fine-tuning on Alpaca dataset with/without LoRA and with/without INT8](https://github.com/stochasticai/xturing/tree/main/examples/galactica)|
46
+
| Generic Wrapper |[Any large language model fine-tuning on Alpaca dataset with/without LoRA and with/without INT8](https://github.com/stochasticai/xturing/tree/main/examples/generic)|
0 commit comments