Skip to content

Commit e4efd3c

Browse files
caitlinwheelesscaitlinwheeless
and
caitlinwheeless
authored
docs: Small template fixes (#6666)
Co-authored-by: caitlinwheeless <[email protected]>
1 parent 98dbd7d commit e4efd3c

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

docs/source/templates/generative-pairwise-human-preference.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ This project will help you to get up your LLM to the ChatGPT quality level throu
1515

1616
Through ranking multiple responses based on quality, you can train a reward model that effectively captures human preferences. This reward model plays a crucial role in Reinforcement Learning, optimizing the performance of the fine-tuned foundational model.
1717

18-
### Further Reading and Resources
18+
#### Further Reading and Resources
1919

2020
- [Gathering Human Feedback Tutorial](https://github.com/heartexlabs/RLHF/blob/master/tutorials/RLHF_with_Custom_Datasets.ipynb) A Jupyter Notebook tutorial that will guide you through the step-by-step process of collecting comparison data, establishing human preferences, and incorporating this feedback into the reward model training.
2121
- [RLHF Resources](https://github.com/heartexlabs/RLHF): A collection of links, tutorials and best practices on how collect data and build an end-to-end Reinforcement Learning from Human Feedback (RLHF) system to fine-tune Generative AI models.

docs/source/templates/index.ejs

+1-1
Original file line numberDiff line numberDiff line change
@@ -51,6 +51,6 @@ cards:
5151
url: "/templates/gallery_generative_ai.html"
5252
- title: LLM Evaluations
5353
categories: generative ai, llm
54-
image: "/images/templates/generative-pairwise-human-preference.png"
54+
image: "/images/templates/response-moderation.png"
5555
url: "/templates/gallery_llm_evals.html"
5656
---

0 commit comments

Comments
 (0)