How should I structure data for fine-tuning a model in a groupchat setting? #729
Unanswered
PyroGenesis
asked this question in
Q&A
Replies: 2 comments
-
@afourney fyi |
Beta Was this translation helpful? Give feedback.
0 replies
-
I am also looking for an answer to this. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm trying to finetune some llm agents such that they perform some specialized task. However, I am unsure of how to best convert an Autogen groupchat conversation into a training format.
For example, let's say I'm working with the following conversation:
Now let's say I'm trying to fine-tune Agent 3 here, which is backed by a Llama 2 model
Llama 2 model format:
How should I go about structuring this into training data? The 3 ways I've thought about are:
Option 1: Treat other Agent responses as separate user responses
Option 2: Combine other Agent responses into single User response
Option 3: Treat other Agent responses as AI responses
Do any of these look viable? Or should I try and roll my own new format which would be more compatible with groupchat conversations?
Beta Was this translation helpful? Give feedback.
All reactions