GithubHelp home page GithubHelp logo

qduoliver / disconet Goto Github PK

View Code? Open in Web Editor NEW

This project forked from ai4ce/disconet

0.0 0.0 0.0 1.41 MB

[NeurIPS2021] Learning Distilled Collaboration Graph for Multi-Agent Perception

License: MIT License

disconet's Introduction

DiscoNet: Learning Distilled Collaboration Graph for Multi-Agent Perception [NeurIPS 2021]

Yiming Li, Shunli Ren, Pengxiang Wu, Siheng Chen, Chen Feng, Wenjun Zhang

"Learn a digraph with matrix-valued edge weight for multi-agent perception."

Documentation Status PyTorch Linux GitLab issues total GitHub stars

News

[2022-07] We updated the codebase to coperception library and dataset to V2X-Sim 2.0. New tasks including segmentation and tracking are included.

[2021-11] Our paper is available on arxiv.

[2021-10] Our dataset V2X-Sim 1.0 is available here.

[2021-09] 🔥 DiscoNet is accepted at NeurIPS 2021.

Abstract

To promote better performance-bandwidth trade-off for multi-agent perception, we propose a novel distilled collaboration graph (DiscoGraph) to model trainable, pose-aware, and adaptive collaboration among agents. Our key novelties lie in two aspects. First, we propose a teacher-student framework to train DiscoGraph via knowledge distillation. The teacher model employs an early collaboration with holistic-view inputs; the student model is based on intermediate collaboration with single-view inputs. Our framework trains DiscoGraph by constraining post-collaboration feature maps in the student model to match the correspondences in the teacher model. Second, we propose a matrix-valued edge weight in DiscoGraph. In such a matrix, each element reflects the inter-agent attention at a specific spatial region, allowing an agent to adaptively highlight the informative regions. During inference, we only need to use the student model named as the distilled collaboration network (DiscoNet). Attributed to the teacher-student framework, multiple agents with the shared DiscoNet could collaboratively approach the performance of a hypothetical teacher model with a holistic view. Our approach is validated on V2X-Sim 1.0, a large-scale multi-agent perception dataset that we synthesized using CARLA and SUMO co-simulation. Our quantitative and qualitative experiments in multi-agent 3D object detection show that DiscoNet could not only achieve a better performance-bandwidth trade-off than the state-of-the-art collaborative perception methods, but also bring more straightforward design rationale. Our code is available on https://github.com/ai4ce/DiscoNet.

Getting started

Please refer to the docs website of coperception for detailed documentations: https://coperception.readthedocs.io/en/latest/
Installation:

Download dataset:

How to create datasets & run tasks:

Example arguments to run DiscoNet:
(assume dataset created properly according to the documentations)

  • Detection / segmentation:
    Inside coperception codebase:

    cd tools/det
    

    or

    cd tools/seg
    

    Training:

    python train_codet.py \
        --data  /path/to/training/dataset \
        --com disco \
        --log --batch 4 \
        --kd_flag 1 \
        --resume_teacher /path/to/teacher/checkpoint.pth \
        --auto_resume_path logs \
        --logpath logs \
        --nepoch 100 \
        -- rsu [0/1]

    Testing:

    python test_codet.py \
        --data /path/to/testing/dataset \
        --com disco \
        --resume /path/to/teacher/checkpoint.pth \
        --tracking \
        --logpath logs \
        --visualization 1 \
        --rsu 1
  • Tracking: Inside coperception codebase:

    cd tools/track
    

    Run tracker:

    make sort \
        mode=disco/[no_rsu/with_rsu] \
        split [test/val] \
        from_agent 1 \
        to_agent 6 \
        det_logs_path /path/to/detection/checkpoints.pth 

    See performance:

    make eval
        mode=disco \
        rsu=[0/1] \
        from_agent=1 \
        to_agent=6 \
        split=test

Acknowledgment

This project is not possible without the following great codebases.

Citation

If you find V2X-Sim 1.0 or DiscoNet useful in your research, please cite our paper.

@InProceedings{Li_2021_NeurIPS,
    title = {Learning Distilled Collaboration Graph for Multi-Agent Perception},
    author = {Li, Yiming and Ren, Shunli and Wu, Pengxiang and Chen, Siheng and Feng, Chen and Zhang, Wenjun},
    booktitle = {Thirty-fifth Conference on Neural Information Processing Systems (NeurIPS 2021)},
    year = {2021}
}

@article{Li_2021_RAL,
    title = {V2X-Sim: Multi-Agent Collaborative Perception Dataset and Benchmark for Autonomous Driving},
    author = {Li, Yiming and Ma, Dekun and An, Ziyan and Wang, Zixun and Zhong, Yiqi and Chen, Siheng and Feng, Chen},
    booktitle = {IEEE Robotics and Automation Letters},
    year = {2022}
}

disconet's People

Contributors

dekunma avatar roboticsyimingli avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.