Skip to content

step3_rlhf_finetuning and two tokenizers #577

Open
@GenVr

Description

@GenVr

Hello.
I'm trying to train a GPT-J 6B, and as a critical model I have trained several networks of different/similar families (gpt2, gpt-neo, bloom, ...)
I know that in step 3 only a tokenizer is used, so with gpt-j I get this error #512.
What critical model can I use with gpt-j?
Thanks.

Metadata

Metadata

Labels

deespeed chatDeepSpeed ChatmodelingRelated to modeling questions.new-configA modified config from the given examplequestionFurther information is requested

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions