Open
Description
hi, i'm doing some reaserch on mutiagent and big2 is a really good plateform to test multiagent algorithm!
but i got a problem about your code that after a long time update, the parameters of the nerual netowrk is Nan. Then i dump latest parameters which have not been nan and train the neural network again. After a long time update, parameters became Nan again. i debug it and i think it caused by Gradient explosion or gradien vanish. so Do you meet the problom just same as me? how do you fix it?
Yours Sincerely!
Metadata
Metadata
Assignees
Labels
No labels