Skip to content

Commit 954fecd

Browse files
authored
Merge pull request #19 from souradipp76/hellokayas-patch-1
Update paper.md
2 parents 737fb22 + 15f5c01 commit 954fecd

File tree

1 file changed

+5
-25
lines changed

1 file changed

+5
-25
lines changed

paper/paper.md

Lines changed: 5 additions & 25 deletions
Original file line numberDiff line numberDiff line change
@@ -84,33 +84,13 @@ We conducted the fine-tuning experiment on a small dataset consisting of randoml
8484

8585
## Before Fine-tuning
8686

87-
We conducted a series of experiments utilizing the `TheBloke/Llama-2-7B-Chat-GPTQ` model [@llama-2-7b-chat-gptq] to demonstrate the functionality and efficacy of our proposed pipeline. The accompanying codebase is designed to be flexible, allowing the user to easily switch between different large language models (LLMs) by simply modifying the configuration file. Given the characteristics of LLMs, models with a greater number of parameters are generally expected to deliver enhanced performance. The BLEU and BERT scores for the `TheBloke/Llama-2-7B-Chat-GPTQ` model are reported in Table 1 and Table 2, under the "W/O FT" or "W/O Finetuning" columns.
87+
We conducted a series of experiments utilizing the `TheBloke/Llama-2-7B-Chat-GPTQ` model [@llama-2-7b-chat-gptq] to demonstrate the functionality and efficacy of our proposed pipeline. The accompanying codebase is designed to be flexible, allowing the user to easily switch between different large language models (LLMs) by simply modifying the configuration file. Given the characteristics of LLMs, models with a greater number of parameters are generally expected to deliver enhanced performance.
8888

8989
## After Fine-tuning
9090

91-
We utilized the PEFT library from Hugging Face, which supports several Parameter Efficient Fine-Tuning (PEFT) methods. This approach is cost-effective for fine-tuning large language models (LLMs), particularly on lightweight hardware. The training configuration and hyperparameters are detailed in Table 3 and Table 4 respectively. The results are reported in Table 1 and Table 2, under the "With FT" or "With Finetuning" columns where the contents are compared with each repository's original README file. It is observed that BLEU scores range from 15 to 30, averaging 20, indicating that the generated text is understandable but requires substantial editing to be acceptable. Conversely, BERT scores reveal a high semantic similarity to the original README content, with an average F1 score of ~85%.
91+
We utilized the PEFT library from Hugging Face, which supports several Parameter Efficient Fine-Tuning (PEFT) methods. This approach is cost-effective for fine-tuning large language models (LLMs), particularly on lightweight hardware. The training configuration and hyperparameters are detailed in Table 1 and Table 2 respectively.
9292

93-
### Table 1: BLEU Scores
94-
95-
| Repository | W/O FT | With FT |
96-
|------------|--------|---------|
97-
| allennlp | 32.09 | 16.38 |
98-
| autojump | 25.29 | 18.73 |
99-
| numpy-ml | 16.61 | 19.02 |
100-
| Spleeter | 18.33 | 19.47 |
101-
| TouchPose | 17.04 | 8.05 |
102-
103-
### Table 2: BERT Scores
104-
105-
| Repository | P (W/O FT) | R (W/O FT) | F1 (W/O FT) | P (With FT) | R (With FT) | F1 (With FT) |
106-
|------------|------------|------------|-------------|-------------|-------------|--------------|
107-
| allennlp | 0.904 | 0.8861 | 0.895 | 0.862 | 0.869 | 0.865 |
108-
| autojump | 0.907 | 0.86 | 0.883 | 0.846 | 0.87 | 0.858 |
109-
| numpy-ml | 0.89 | 0.881 | 0.885 | 0.854 | 0.846 | 0.85 |
110-
| Spleeter | 0.86 | 0.845 | 0.852 | 0.865 | 0.866 | 0.865 |
111-
| TouchPose | 0.87 | 0.841 | 0.856 | 0.831 | 0.809 | 0.82 |
112-
113-
### Table 3: QLoRA Configuration
93+
### Table 1: QLoRA Configuration
11494

11595
| Parameter | Value |
11696
|---------------|-------|
@@ -120,7 +100,7 @@ We utilized the PEFT library from Hugging Face, which supports several Parameter
120100
| `bias` | None |
121101
| `task_type` | CAUSAL_LM |
122102

123-
### Table 4: Training Hyper-parameters
103+
### Table 2: Training Hyper-parameters
124104

125105
| Parameter | Value |
126106
|-----------------------------------|----------|
@@ -133,4 +113,4 @@ We utilized the PEFT library from Hugging Face, which supports several Parameter
133113
| `lr_scheduler_type` | cosine |
134114
| `warmup_ratio` | 0.01 |
135115

136-
# References
116+
# References

0 commit comments

Comments
 (0)