GithubHelp home page GithubHelp logo

abs's Introduction

ABS

Official Implementation for Agile But Safe: Learning Collision-Free High-Speed Legged Locomotion.

Robotics: Science and Systems (RSS) 2024

Tairan He*, Chong Zhang*, Wenli Xiao, Guanqi He, Changliu Liu, Guanya Shi

This codebase is under CC BY-NC 4.0 license, with inherited license in Legged Gym and RSL RL from ETH Zurich, Nikita Rudin and NVIDIA CORPORATION & AFFILIATES. You may not use the material for commercial purposes, e.g., to make demos to advertise your commercial products.

Please read through the whole README.md before cloning the repo.

Training in Simulation

Pipeline to Install and Train ABS

Note: Before running our code, it's highly recommended to first play with RSL's Legged Gym version to get a basic understanding of the Isaac-LeggedGym-RslRL framework.

  1. Create environment and install torch

    conda create -n xxx python=3.8  # or use virtual environment/docker
    
    pip3 install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu116  
    # used version during this work: torch==2.0.1 torchvision==0.15.2 torchaudio==2.0.2
    # for older cuda ver:
    pip3 install torch==1.10.0+cu113 torchvision==0.11.1+cu113 torchaudio==0.10.0+cu113 -f https://download.pytorch.org/whl/cu113/torch_stable.html
    
  2. Install Isaac Gym preview 4 release https://developer.nvidia.com/isaac-gym

    unzip files to a folder, then install with pip:

    cd isaacgym/python && pip install -e .

    check it is correctly installed by playing:

    cd examples && python 1080_balls_of_solitude.py
  3. Clone this codebase and install our rsl_rl in the training folder

    pip install -e rsl_rl
  4. Install our legged_gym

    pip install -e legged_gym

    Ensure you have installed the following packages:

    • pip install numpy==1.20 (must < 1.24)
    • pip install tensorboard
    • pip install setuptools==59.5.0
  5. Try training.

    can use "--headless" to disable gui, press "v" to pause/resume gui play.

    for go1, in legged_gym/legged_gym,

    # agile policy
    python scripts/train.py --task=go1_pos_rough --max_iterations=4000 
    
    # agile policy, lagrangian ver
    python scripts/train.py --task=go1_pos_rough_ppo_lagrangian --max_iterations=4000
    
    # recovery policy
    python scripts/train.py --task=go1_rec_rough --max_iterations=1000
    
  6. Play the trained policy

    python scripts/play.py --task=go1_pos_rough
    python scripts/play.py --task=go1_rec_rough
  7. Use the testbed, and train/test Reach-Avoid network:

    # try testbed
    python scripts/testbed.py --task=go1_pos_rough [--load_run=xxx] --num_envs=1
    
    # train RA (be patient it will take time to converge) 
    # make sure you have at least exported one policy by play.py so the exported folder exists
    python scripts/testbed.py --task=go1_pos_rough --num_envs=1000 --headless --trainRA
    
    # test RA (only when you have trained one RA)
    python scripts/testbed.py --task=go1_pos_rough --num_envs=1 --testRA
    
    # evaluate
    python scripts/testbed.py --task=go1_pos_rough --num_envs=1000 --headless [--load_run=xxx] [--testRA]
    
  8. Sample dataset for ray-prediction network training

    python scripts/camrec.py --task=go1_pos_rough --num_envs=3
    • Tips 1: You can edit the shift value in Line 93 and the log_root in Line 87 to collect different dataset files in parallel (so you can merge them by simply moving the files), and manually change the obstacles in env_cfg.asset.object_files in Line 63.
    • Tips 2: After collecting the data, there's a template code in train_depth_resnet.py to train the ray-prediction network, but using what you like for training CV models is highly encouraged!
    • Tips 3: You may change camera configs of resolution, position, FOV, and depth range in the config file Line 151.

Hardware Deployment

System Overview

3D Print Mounts

Deployment Code Installation

Deployment Setup

  • Low Level Control Mode for Unitree Go1: L2+A -> L2+B -> L1+L2+Start

  • Network Configuration for Orin NX:

    • IP: 192.168.123.15
    • Netmask: 255.255.255.0
  • Convert the .pt files of agile/recovery policy, RA value to .onnx files using src/abs_src/onnx_model_converter.py

  • Modify the path of (.onnx or .pt) models in publisher_depthimg_linvel.py and depth_obstacle_depth_goal_ros.py

Deployment Scripts

  1. roscore: Activate ROS Noetic Envrioment
  2. cd src/abs_src/: Enter the ABS scripts file
  3. python publisher_depthimg_linvel.py: Publish ray prediction results and odometry results for navigation goals
  4. python led_control_ros.py: Control the two LED lights based on RA values
  5. python depth_obstacle_depth_goal_ros.py: Activate the Go1 using the agile policy and the recovery policy

Deployment Controllers

  • B: Emergence stop
  • Default: Go1 Running ABS based on goal command
  • L2: Turn left
  • R2: Turn right
  • Down: Back
  • Up: Stand
  • A: Turn around
  • X: Back to initial position

Troubleshooting:

Contact

Issues

You can create an issue if you meet any bugs, except:

  • If you cannot run the vanilla RSL's Legged Gym, it is expected that you first go to the vanilla Legged Gym repo for help.
  • There can be CUDA-related errors when there are too many parallel environments on certain PC+GPU+driver combination: we cannot solve thiss, you can try to reduce num_envs.
  • Our codebase is only for our hardware system showcased above. We are happy to make it serve as a reference for the community, but we won't tune it for your own robots.

Credit

If our work does help you, please consider citing us and the following works:

@inproceedings{AgileButSafe,
  author    = {He, Tairan and Zhang, Chong and Xiao, Wenli and He, Guanqi and Liu, Changliu and Shi, Guanya},
  title     = {Agile But Safe: Learning Collision-Free High-Speed Legged Locomotion},
  booktitle = {Robotics: Science and Systems (RSS)},,
  year      = {2024},
}

We used codes in Legged Gym and RSL RL, based on the paper:

  • Rudin, Nikita, et al. "Learning to walk in minutes using massively parallel deep reinforcement learning." CoRL 2022.

Previsou works that heavily inspired the policy training designs:

  • Rudin, Nikita, et al. "Advanced skills by learning locomotion and local navigation end-to-end." 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2022.
  • Zhang, Chong, et al. "Learning Agile Locomotion on Risky Terrains." arXiv preprint arXiv:2311.10484 (2023).
  • Zhang, Chong, et al. "Resilient Legged Local Navigation: Learning to Traverse with Compromised Perception End-to-End." ICRA 2024.

Previsou works that heavily inspired the RA value design:

  • Hsu, Kai-Chieh, et al. "Safety and liveness guarantees through reach-avoid reinforcement learning." RSS 2021.

Previsou works that heavily inspired the perception design:

  • Hoeller, David, et al. "Anymal parkour: Learning agile navigation for quadrupedal robots." Science Robotics 9.88 (2024): eadi7566.
  • Acero, F., K. Yuan, and Z. Li. "Learning Perceptual Locomotion on Uneven Terrains using Sparse Visual Observations." IEEE Robotics and Automation Letters 7.4 (2022): 8611-8618.

abs's People

Contributors

tairanhe avatar guanyashi avatar zita-ch avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.