Skip to content

Conversation

@NeoZng
Copy link

@NeoZng NeoZng commented Nov 4, 2025

The original impl of foot_clearance_reward() encourages the agent to stand still, which could maximize the reward (since exp(-0)=1).

This PR add command and current velocity weights to tell whether the robot is staying still or moving.

Wrong implementation of the function is the main reason why the examples in this repo require a huge amount of iterations to just learn to walk. On contrast, examples in Isaaclab official impl takes only about 500 iters with 4096 envs to walk in complex terrain (though it takes base_lin_vel in policy observation)

@NeoZng
Copy link
Author

NeoZng commented Nov 4, 2025

BTW, it is not recommend to use this reward for terrains including pyramid/slope, better to modify the implementation of feet_height_body() .

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant