diff --git a/notebooks/en/fine_tuning_llm_grpo_trl.ipynb b/notebooks/en/fine_tuning_llm_grpo_trl.ipynb index dd7a3a01..67505e4d 100644 --- a/notebooks/en/fine_tuning_llm_grpo_trl.ipynb +++ b/notebooks/en/fine_tuning_llm_grpo_trl.ipynb @@ -470,9 +470,9 @@ "source": [ "### 3.4 Configuring GRPO Training Parameters\n", "\n", - "Next, let's configure the training parameters for GRPO. We recommend experimenting with the `max_completion_length`, `num_generations`, and `max_prompt_length` parameters (refer to the image at the beginning for details about each of them).\n", + "Next, let's configure the training parameters for GRPO. We recommend experimenting with the `max_completion_length`and `num_generations` parameters (refer to the image at the beginning for details about each of them).\n", "\n", - "To keep things simple, we’ll start by training for just one epoch and reducing the `max_completion_length`, `num_generations`, and `max_prompt_length` from their default values." + "To keep things simple, we’ll start by training for just one epoch and reducing the `max_completion_length`and `num_generations`from their default values." ], "metadata": { "id": "qW_3r8T1EtNg" @@ -500,7 +500,6 @@ " # Parameters that control de data preprocessing\n", " max_completion_length=64, # default: 256\n", " num_generations=4, # default: 8\n", - " max_prompt_length=128, # default: 512\n", "\n", " # Parameters related to reporting and saving\n", " report_to=[\"tensorboard\"],\n", @@ -888,4 +887,4 @@ }, "nbformat": 4, "nbformat_minor": 0 -} \ No newline at end of file +}