Hi @xbpeng , I have been training AMP and ADD on some of my hand animated reference motions ( dynamically infeasible) for Go2 and Go2w. It seems like AMP and ADD returns are zero consistently throughout the iterations, my guess is the discriminator is trained on one cycle of reference motion hence it is over fitting, to find any state transition generated by policy to be fake. But, when I tried runnning policies from ADD, it is somewhat following the motions, but is way worse than Deepmimic trained policy. Do you have any suggestions or thoughts on this?
This is log from training AMP on one cycle of trot motion which is hand animated.

Hi @xbpeng , I have been training AMP and ADD on some of my hand animated reference motions ( dynamically infeasible) for Go2 and Go2w. It seems like AMP and ADD returns are zero consistently throughout the iterations, my guess is the discriminator is trained on one cycle of reference motion hence it is over fitting, to find any state transition generated by policy to be fake. But, when I tried runnning policies from ADD, it is somewhat following the motions, but is way worse than Deepmimic trained policy. Do you have any suggestions or thoughts on this?
This is log from training AMP on one cycle of trot motion which is hand animated.
