The PARL team gets the first place in NeurIPS reinforcement learning competition, again! This folder contains our final submitted model and the code relative to the training process.
-
How to Run
- Enter the sub-folder
final_submit
- Download the model file from online storage service: Baidu Pan (password:
b5ck
) or Google Drive - Unpack the file:
tar zxvf saved_models.tar.gz
- Launch the test script:
python test.py
- Enter the sub-folder
The curriculum learning pipeline to get a walking slowly model is the same pipeline in our winning solution in NeurIPS 2018: AI for Prosthetics Challenge. You can get a walking slowly model by following the guide.
We also provide a pre-trained model that walk naturally at ~1.3m/s. You can download the model file (naming low_speed_model
) from online storage service: Baidu Pan (password: q9vj
) or Google Drive.
We built our distributed training agent based on PARL cluster. To start a PARL cluster, we can execute the following two xparl commands:
# starts a master node to manage computation resources and adds the local CPUs to the cluster. xparl start --port 8010
# if necessary, adds more CPUs (computation resources) in other machine to the cluster. xparl connect --address [CLUSTER_IP]:8010
For more information of xparl, please visit the documentation.
In this example, we can start a local cluster with 300 CPUs by running:
xparl start --port 8010 --cpu_num 300
Then, we can start the distributed training by running:
# NOTE: You need provide a self-trained model, or download the `low_speed_model` as mentioned above.
sh scripts/train_difficulty1.sh ./low_speed_model
Optionally, you can start the distributed evaluating by running:
sh scripts/eval_difficulty1.sh
sh scripts/train_difficulty2.sh [TRAINED DIFFICULTY=1 MODEL]
sh scripts/train_difficulty3_first_target.sh [TRAINED DIFFICULTY=2 MODEL]
sh scripts/train_difficulty3.sh [TRAINED DIFFICULTY=3 FIRST TARGET MODEL]