awesome-neural-planner
awesome-neural-planner copied to clipboard
personal paper reading on neural motion planner and controller
Papers on neural motion planner and controller
Autonomous Driving
arXiv: A Survey of End-to-End Driving:Architectures and Training Methods. paper
comment: a recent survey on end-to-end driving pipeline, compared to conventional modular pipeline.
ECCV20: Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations. paper
comment: A follow-up for the CVPR19 work below, a semantic occupancy representation is introduced between perception and motion planning. Cost function incorporates comfort driving.
CVPR19: End-to-end Interpretable Neural Motion Planner. paper
comment: Like inverseRL, a cost function is learned from human demonstration when input LiDAR and roadmap, output perception detection result (bounding boxes) as well as cost volume. Then learned cost function is used to select the sampled candidate trajectories. Pay attention to its trajectory parameterization and the trajectory sampler for plausible trajectories. Planning horization is of 3 seconds. Interpretability cames from the joint output of 3D LiDAR detection result.
arXiv: End-to-end Contextual Perception and Prediction with Interaction Transformer. paper
comment: marrying 3D object detection and motion forecasting. A transformer module is introduced for motion forecasting task.
coRL20sub: Super-Human Performance in Gran Turismo Sport Using Deep Reinforcement Learning. paper video!
comment: pure deep RL approach for car-racing. Pay attention to its reward design and policy input and output. the side-by-side comparison between human expert and RL algorithm is interesting. The system architecture is also interesting, by using 4 PlayState 4 game consoles and a single desktop PC through ethernet connection.
Indoor navigation
coRL19: Combining Optimal Control and Learning for Visual Navigation in Novel Environments. paper page
comment: learning-based perception + model-based control. waypoints are output from the perception, then fed into the model-based planner to generate smoonth feasible trajectory. strong assumption that map is known and odometry is perfect and environment is static
Mobile manipulation
coRL20sub: ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation. paper page
comment: a hybrid system integrating traditional motion planner and RL learning. Rapidly-exploring random tree (RRT) and probabilistic raod-maps (PRM) have been adopted in the motion planner. RL policy output a subgoal (desired 2D loction/end-effector 3D location). The policy has different parameterization. Experimental platform iGibson
Robot manipulation
Algorithms
IROS20: Learning High-Level Policies for Model Predictive Control. paper code
comment: loop high-level policy learning into MPC
Relevant Researchers
Davide Scaramuzza: ETH badass researcher.
Silvio Savarese: Stanford faculty, focusing on indoor navigation
Raquel Urtasun: Utoronto, Uber ATG toronto head, autonomous driving leading researcher.
Saurabh Gupta: UIUC, vision+navigation.