AirGym is an open souce Python quadrotor simulator based on IsaacGym, a part of AirGym series Sim-to-Real workding flow. It provides a high-fidelity dynamics and Deep Reinforcement Learning (DRL) framework for quadrotor robot learning research. Furthermore, we also provide toolkits for transferring policy from AirGym simulator to the real quadrotor emNavi-X152b, making Sim-to-Real possible. The documentation website is at https://emnavi.tech/AirGym/.
The Sim-to-Real working flow of AirGym series is broken down into four parts:
- AirGym: the quadrotor simulation platform described in this repository, providing environments and basic training algorithm.
- rlPx4Controller: a flight geometric controller that maintains strict control logic alignment with the open-source PX4, for a better Sim-to-Real.
- AirGym-Real: an onboard Sim-to-Real module compatible with AirGym, enabling direct loading of pretrained models, supporting onboard visual-inertial pose estimation, ROS topic publishing, and one-shot scripted deployment.
- control_for_gym: a middleware layer based on MAVROS for forwarding control commands at various levels to PX4 autopilot. It includes a finite state machine to facilitate switching between DRL models and traditional control algorithms.
AirGym is more lightweight and has a clearer file structure compared to other simulators, because it was designed from the beginning with the goal of achieving Sim-to-Real transfer.
- Lightweight & Customizable: AirGym is extremely lightweight yet highly extensible, allowing you to quickly set up your own customized training task.
- Strict Alignment with PX4 Logic: Flight control in AirGym is supported by rlPx4Controller. It maintains strict control logic alignment with the open-source PX4, for a better Sim-to-Real.
- Multiple Control Modes: AirGym provides various control modes including PY (position & yaw), LV (linear velocity & yaw), CTA (collective thrust & attitude angle), CTBR (collective thrust & body rate), SRT (single-rotor thrust).
- Sim-to-Real Toolkits: AirGym series have complete flow of robot learning Sim-to-Real and provide a potential to transfer well-trained policies to a physical device.
AirGym provides four basic tasks Hovering, Balloon, Tracking, Avoid, and a higher level task Planning. All tasks are implemented on X152b quadrotor frame since this is our Sim-to-Real device.
Task Hovering: the quadrotor is expected to be initialized randomly inside a cube with a side length of 2 meters, then converge into the center and hover until the end of the episode. Also, this task can be used as "go to waypoint" task if specify a target waypoint.
Task Balloon: also called target reaching. It is essentially a variation of the hovering task, but with a key difference: the quadrotor moves at a higher speed, rapidly dashing toward the balloon (the target).
Task Tracking: tracking a sequance of waypoints which is played as a trajectory. The tracking speed is effected by the trajectory playing speed.
Task Avoid: hover and try to avoid a cube or a ball with random throwing velocity and angle. This task provides depth image as input.
Task Planning: a drone navigates and flies through a random generated woods like area, using only depth information as input. No global information is utilized in this task which means a better adaptation in a GNSS-deny environment and without a VIO.
Furthermore, you can build a customized environment and even task upon AirGym. Here is a simple demo of random assets generation.
Note: this repository has been tested on Ubuntu 20.04 PyTorch 2.0.0 + CUDA11.8 with RTX4090.
Run configuration.sh
to download and install NVIDIA Isaac Gym Preview 4 and rlPx4Controller at ~
, and create a new conda environment named airgym
. Note that sudo
is required to install apt
package:
chmod +x configuration.sh
./configuration.sh
You can run the example script which is a quadrotor position control illustration of 4096 parallel environments:
conda activate airgym
python airgym/scripts/example.py --task hovering --ctl_mode pos --num_envs 4
The default ctl_mode
is position control.
AirGym/
│
├── airgym/ # source code
│ ├── envs/ # envs definition
│ │ ├── base/ # base envs
│ │ └── task/ # tasks
│ │ ├── planning.py # task env file
│ │ ├── planning_config.py # task env config file
│ │ ├── ...
│ │ ...
│ │
│ ├── assets/ # assets
│ │ ├── env_assets/ # env assets
│ │ └── robots/ # robot assets
│ ├── scripts/ # debug scripts
│ └── utils/ # others
│
├── doc/ # doc assets
│
├── lib/ # RL algorithm
│
├── scripts/ # program scipts
│ ├── config/ # config file for training
│ ├── runner.py # program entry
│ └── ...
│
└── README.md # introduction
We train the model by a rl_games
liked customed PPO (rl_games
was discarded after the version 0.0.1beta because of its complexity for use). The training algorithm is at airgym/lib
modified from rl_games-1.6.1
. Of course you can use any other RL libs for training.
Training from scratch can be down within minites. Using Hovering as an example:
python scripts/runner.py --ctl_mode rate --headless --task hovering
Algorithm related parameters can be edited in .yaml
files. Environment and simulator related parameters are located in ENV_config files like hovering_config.py
. The ctl_mode
must be spicified.
The input arguments can be overrided:
args | explanation |
---|---|
--play | Play(test) a trained network. |
--task | Spicify a task. 'hovering', 'avoid', 'balloon', 'planning', 'tracking' |
--num_envs | Spicify the number of parallel running environments. |
--headless | Run in headless mode or not. |
--ctl_mode | Spicify a control level mode from 'pos', 'vel', 'atti', 'rate', 'prop'. |
--checkpoint | Load a trained checkpoint model from path. |
--file | Spicify an algorithm config file for training. |
|
Load a trained checkpoint is quite easy:
python scripts/runner.py --play --task hovering --num_envs 64 --ctl_mode rate --checkpoint <path-to-ckpt>
Important: emNavi provide a general quadrotor sim2real approach, please refer to AirGym-Real @https://github.com/emNavi/AirGym-Real.
- Create a new
.py
file in directoryairgym/envs/task
. Usually, we use<task_name>.py
to describe an environment and a<task_name>_config.py
to describe basic environment settings. You can follow this rule. - Register new task in
airgym/envs/task/__init__.py
. This is registration for AirGym environment. Note that the registration for vectorized parallel environments and training agent have been implemented automatically. Using task Planning as an example:
TASK_CONFIGS = [
{
'name': 'planning',
'config_module': 'task.planning_config',
'config_class': 'PlanningConfig',
'task_module': 'task.planning',
'task_class': 'Planning'
}
]
We recommend users inherit from task hovering
(no camera) and customized
(depth).
We use an asset register to manage assets, and an asset manager to realize quickly assets loading and environment creating.
Before adding assets into the environment, users should register their own assets at airgym/assets/__init__.py
by using function registry.register_asset(asset_name, override_params, asset_type)
.
params | explanation |
---|---|
asset_name | The identified name of asset in airgym. |
override_params | Customized assets parameters that can override the default settings in airgym/assets/asset_register.py . |
asset_type | Asset type: "single" or "group". single : One asset defined in a .urdf file. group : A kind of assets that have similar features, like cubes with multiple shapes of cube1x1 , cube1x4 , et al. If asset type is "group", the key "path" in the override_params should be a folder. |
|
Here we use robot X152b
as an example:
registry.register_asset(
"X152b",
override_params={
"path": f"{AIRGYM_ROOT_DIR}/airgym/assets/robots/X152b/model.urdf",
"name": "X152b",
"base_link_name": "base_link",
"foot_name": None,
"penalize_contacts_on": [],
"terminate_after_contacts_on": [],
"disable_gravity": False,
"collapse_fixed_joints": True,
"fix_base_link": False,
"collision_mask": 1,
"density": -1,
"angular_damping": 0.0,
"linear_damping": 0.0,
"replace_cylinder_with_capsule": False,
"flip_visual_attachments": False,
"max_angular_velocity": 100.,
"max_linear_velocity": 100.,
"armature": 0.001,
},
asset_type="single"
)
Considering frequently changes of assets settings, we use a simple but very clear method to include your customed assets in the task, defining it in <task_name>_config.py
file. Here is an example:
class asset_config:
include_robot = {
"X152b": {
"num_assets": 1,
"enable_onboard_cameras": True,
"collision_mask": 1,
}
}
include_single_asset = {
"cubes/1x1": {
"collision_mask": 0,
"num_assets": 1,
},
}
include_group_asset = {
"cubes": {
"num_assets": 10,
"collision_mask": 0,
},
}
include_boundary = {
"18x18ground": {
"num_assets": 1,
},
}
To define four dictionary, you can add pre-registered assets into task. include_single_asset
indicates to include a specified type of asset. include_group_asset
indicatde to randomly include a group of a type of assets. include_boundary
means to include a ground, or any boundaries like a wall, et al. Robots are also included by include_robot
.
IMPORTANT: considering a fast debug of assets features, you can add and edit params of asset in the dictionaries, to override the original settings, which is extremly efficient during development.
Python file asset_manager.py
provide clear but powerful functions to load assets, create assets, and achieving quick params overriding. Here are key functions in class AssetManager
:
func load_asset(self: AssetManager, gym: gymapi.acquire_gym(), sim: gym.create_sim()):
Load and prepare assets from ENV_config.py
. Overriding and saving parameters into lists, to avoid repetitive loading of assets.
func create_asset(self: AssetManager, env_handle: gym.create_env(), start_pose: gymapi.Transform(), env_id: int):
Create isaacgym actors according to prepared lists, to add assets into task.
Add multiple drone training tasks.
TBC
LICENSE FILE This project is inspired and greatly improved from project aerial_gym_simulator that adheres to the BSD 3-Clause License. In consequence, this work will continuously develope under BSD 3-Clause License. Currently, we still use assets management part from aerial_gym_simulator although we have made a great changes, but we are rewriting and optimizing this part. In the future, we will continue to build upon BSD 3-Clause and claim the contributions of those who came before us.
We thanks colleagues @emNavi Tech (https://emnavi.tech/) who develope and maintain the outstanding hardware device emNavi-X152b for Sim-to-Real task.
Also, we thanks the excellent work by Aerial Gym Simulator licensed(https://github.com/ntnu-arl/aerial_gym_simulator) under the BSD 3-Clause License. AirGym is modified and greatly improved upon Aerial Gym Simulator especially for Sim-to-Real task.
TBC