hi, i'm doing some reaserch on mutiagent and big2 is a really good plateform to test multiagent algorithm!
but i got a problem about your code that after a long time update, the parameters of the nerual netowrk is Nan. Then i dump latest parameters which have not been nan and train the neural network again. After a long time update, parameters became Nan again. i debug it and i think it caused by Gradient explosion or gradien vanish. so Do you meet the problom just same as me? how do you fix it?
Yours Sincerely!
hi, i'm doing some reaserch on mutiagent and big2 is a really good plateform to test multiagent algorithm!
but i got a problem about your code that after a long time update, the parameters of the nerual netowrk is Nan. Then i dump latest parameters which have not been nan and train the neural network again. After a long time update, parameters became Nan again. i debug it and i think it caused by Gradient explosion or gradien vanish. so Do you meet the problom just same as me? how do you fix it?
Yours Sincerely!