star
star copied to clipboard
[CoRL 2022] Multi Robot Scene Completion
Multi Robot Scene Completion: Towards Task-agnostic Collaborative Perception
Yiming Li, Juexiao Zhang, Dekun Ma, Yue Wang, Chen Feng
See our paper on OpenReview.
News:
[2022-11] Our paper is camera-ready!
[2022-10] The project website is online.
[2022-09] Our work is accepted at the 6th Conference on Robot Learning (CoRL 2022).
Abstract:
Collaborative perception learns how to share information among multiple robots to perceive the environment better than individually done. Past research on this has been task-specific, such as detection or segmentation. Yet this leads to different information sharing for different tasks, hindering the large-scale deployment of collaborative perception. We propose the first task-agnostic collaborative perception paradigm that learns a single collaboration module in a self-supervised manner for different downstream tasks. This is done by a novel task termed multi-robot scene completion, where each robot learns to effectively share information for reconstructing a complete scene viewed by all robots. Moreover, we propose a spatiotemporal autoencoder (STAR) that amortizes over time the communication cost by spatial sub-sampling and temporal mixing. Extensive experiments validate our method's effectiveness on scene completion and collaborative perception in autonomous driving scenarios.
Installation
The work is tested with:
- python 3.7
- pytorch 1.8.1
- torchvision 0.9.1
- timm 0.3.2
Download the GitHub repository:
git clone https://github.com/coperception/star.git
cd star
Create a conda environment with the dependencies:
conda env create -f environment.yml
conda activate star
If conda installation failed, install the dependencies through pip:
(Make sure your Python version is 3.7
)
pip install -r requirements.txt
Usage:
To train, run:
cd completion/
make train_completion
To test the trained model on scene completion:
cd completion/
make test_completion
More commands and experiment settings are included in the Makefile.
You can find the training and test scripts at: completion.
Dataset:
Our experiments are conducted on the V2X-Sim[1] simulated dataset. Find more about the dataset on the website.
[1] Li, Yiming, et al. "V2X-Sim: Multi-agent collaborative perception dataset and benchmark for autonomous driving." IEEE Robotics and Automation Letters 7.4 (2022): 10914-10921.
Citation:
@inproceedings{li2022multi,
title={Multi-Robot Scene Completion: Towards Task-Agnostic Collaborative Perception},
author={Li, Yiming and Zhang, Juexiao and Ma, Dekun and Wang, Yue and Feng, Chen},
booktitle={6th Annual Conference on Robot Learning}
}