Skip to content

Commit 2abdef7

Browse files
committed
Fix top k tuning list
1 parent 83aa07b commit 2abdef7

File tree

2 files changed

+4
-0
lines changed

2 files changed

+4
-0
lines changed

docs/language_model/llm/llm_intro_hyperparameter_tuning.ipynb

+2
Original file line numberDiff line numberDiff line change
@@ -364,7 +364,9 @@
364364
"metadata": {},
365365
"source": [
366366
"In LLMs, the `top_k` hyperparameter is a key factor that influences the unpredictability of the generated output.\n",
367+
"\n",
367368
"- **For smaller `top_k` values**: The model behaves in a more predictable manner. It only takes into account a limited set of the most probable next tokens at each step of the generation process. This can result in responses that are more concise and consistent, but there’s a possibility that the output may be too restricted or repetitive.\n",
369+
"\n",
368370
"- **For larger `top_k` values**: The model takes into consideration a broader set of potential next tokens. This infuses more variety and randomness into the generated output. However, the responses can become less consistent and may occasionally be less coherent or pertinent.\n",
369371
"Therefore, the selection of the top_k value can be viewed as a balance between consistency and variety in the model’s responses. It’s crucial to adjust this parameter based on the specific needs of your task. "
370372
]

docs/language_model/llm/llm_intro_hyperparameter_tuning.md

+2
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,9 @@ In the Python code above leveraging on `numpy` library, you can see that
207207
### Top-K Tuning
208208

209209
In LLMs, the `top_k` hyperparameter is a key factor that influences the unpredictability of the generated output.
210+
210211
- **For smaller `top_k` values**: The model behaves in a more predictable manner. It only takes into account a limited set of the most probable next tokens at each step of the generation process. This can result in responses that are more concise and consistent, but there’s a possibility that the output may be too restricted or repetitive.
212+
211213
- **For larger `top_k` values**: The model takes into consideration a broader set of potential next tokens. This infuses more variety and randomness into the generated output. However, the responses can become less consistent and may occasionally be less coherent or pertinent.
212214
Therefore, the selection of the top_k value can be viewed as a balance between consistency and variety in the model’s responses. It’s crucial to adjust this parameter based on the specific needs of your task.
213215

0 commit comments

Comments
 (0)