This project contains three environments of Toyota HSR for RL research. Three environments are corresponding to food-searching, maze and grasping tasks.
Here is the code installing testing in Windows 10.
- Downloading the codes moving to its content
git clone https://github.com/CROBOT975/HSR_RL-environment.git
cd HSR_RL-enviroment/PyHSR
- Activating the venv
cd venv/Scripts
activate
cd ../..
- Installing the required packages
pip install numpy
pip install gym
pip install pybullet
pip install stable-baselines3
You can download the Human Support Robot model, and yet the grippers of HSR model is not appropriate for grasping. Therefore, you'd better make some change in the original code of HSR urdf model,like following:
Download the model:
git clone https://github.com/ToyotaResearchInstitute/hsr_meshes.git
cp -rp hsr_meshes venv/lib/python3.9/site-packages/pybullet_data
cp -p HSR_RL_env/model/hsrb4s.urdf venv/lib/python3.9/site-packages/pybullet_data
Copy the object files to pybullet_data,and change the original code of hsrb4s.urdf like follwing image(from '.stl' to '.obj' for two fingers and two finger tips):
cp -p HSR_RL_env/model/l_finger.obj venv/lib/python3.9/site-packages/pybullet_data
cp -p HSR_RL_env/model/l_finger_tip.obj venv/lib/python3.9/site-packages/pybullet_data
Downloading some environment models
cp -p HSR_RL_env/model/food_cube.urdf venv/lib/python3.9/site-packages/pybullet_data
cp -p HSR_RL_env/model/food_cube_for_grasping.urdf venv/lib/python3.9/site-packages/pybullet_data
cp -p HSR_RL_env/model/food_sphere.urdf venv/lib/python3.9/site-packages/pybullet_data
cp -p HSR_RL_env/model/maze.urdf venv/lib/python3.9/site-packages/pybullet_data
cp -p HSR_RL_env/model/wall.urdf venv/lib/python3.9/site-packages/pybullet_data
cd HSR_RL_env
pip install -e .
cd ..
Following codes are only involved with PPO algorithm as example.
- Foodsearhing model
python examples/rl.py --env_name="FoodHuntingHSR-v0" --total_timesteps=206083 --filename="saved_model_F"
- Mazenavigating model
python examples/rl.py --env_name="MazeNavigatingHSR-v0" --total_timesteps=206083 --filename="saved_model_M"
- Grasping model
python examples/rl.py --env_name="GraspingHSR-v0" --total_timesteps=206083 --filename="saved_model_G"
After taining the models, you can demonstrate them due to following codes:
python examples/rl.py --env_name="FoodHuntingHSRGUI-v0" --total_timesteps=10000 --filename="saved_model_F" --play
python examples/rl.py --env_name="MazeNavigatingHSRGUI-v0" --total_timesteps=10000 --filename="saved_model_M" --play
python examples/rl.py --env_name="GraspingHSRGUI-v0" --total_timesteps=1000 --filename="saved_model_G" --play
You can see the demonstration which denotes the dramatic performance of RL algorithm due to the difference between 2e3 timesteps traning and 2e5 timesteps training as following(Click on thumbnail to download video.):
- Foodsearhing
- Mazenavigating
- Grasping