You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Oct 31, 2023. It is now read-only.
Is the eval_success being logged at each iteration according to the code and the plots which are reported in the paper simply have their axis scaled by an appropriate factor, in our case something like 2 x 10eX (X depends on the answer to the above point) - so essentially the total number of logged eval_success is 200?
Please do let me know if my understanding is right in this setting.
Thanks,
Megh
Hi Rutav,

The plots that are provided in the paper - example given below plot the number of samples vs the success rate
I also see that the code logs the success rate after every iteration as mentioned here - https://github.com/ShahRutav/mjrl/blob/6cdb8b8c72279abe8d9d8b8a800f8ac396413e42/mjrl/utils/train_agent.py#L119 and according to the default configuration file here - https://github.com/facebookresearch/RRL/blob/main/examples/config/hammer_dapg.yaml#L40 - the code is being run for 200 iterations, I also see here https://github.com/facebookresearch/RRL/blob/main/examples/config/hammer_dapg.yaml#L40 that the number of trajectories is 200, so I think the horizon length, say h, has to be 100, assuming that the training is happening for 4 x 10e6 if I am not wrong - I have the following doubts -
Please do let me know if my understanding is right in this setting.
Thanks,
Megh