You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This commit was created on GitHub.com and signed with GitHub’s verified signature.
The key has expired.
Refine RL workflow
Add **kwargs to support more problem setting (e.g., Graph based ones) (#589)
add **kwargs to RL models' forward funcs and _shape_check()
add **kwargs to RL policies' get_action related funcs and _post_check()
add **kwargs to choose_actions of AbsEnvSampler; remain it None in current sample() and eval()
Add detached loss to the return value of update_critic() and update_actor() of current TrainOps; add default Falseearly_stop to update_actor() of current TrainOps (#589)
Refine random seed setting logic in RL workflow (#584)
Replace the numpy data type with python common data type in whole project (#571)
Add RL benchmark on Mujoco as a module to tests/, compared with spinning up benchmark, performance results can be found in tests/rl/performance.md (#575, #577, #583, #584)