zhoubolei / awesome-neural-planner

personal paper reading on neural motion planner and controller

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Papers on neural motion planner and controller

Autonomous Driving

arXiv: A Survey of End-to-End Driving:Architectures and Training Methods. paper
comment: a recent survey on end-to-end driving pipeline, compared to conventional modular pipeline.

ECCV20: Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations. paper
comment: A follow-up for the CVPR19 work below, a semantic occupancy representation is introduced between perception and motion planning. Cost function incorporates comfort driving.

CVPR19: End-to-end Interpretable Neural Motion Planner. paper
comment: Like inverseRL, a cost function is learned from human demonstration when input LiDAR and roadmap, output perception detection result (bounding boxes) as well as cost volume. Then learned cost function is used to select the sampled candidate trajectories. Pay attention to its trajectory parameterization and the trajectory sampler for plausible trajectories. Planning horization is of 3 seconds. Interpretability cames from the joint output of 3D LiDAR detection result.

arXiv: End-to-end Contextual Perception and Prediction with Interaction Transformer. paper
comment: marrying 3D object detection and motion forecasting. A transformer module is introduced for motion forecasting task.

coRL20sub: Super-Human Performance in Gran Turismo Sport Using Deep Reinforcement Learning. paper video!
comment: pure deep RL approach for car-racing. Pay attention to its reward design and policy input and output. the side-by-side comparison between human expert and RL algorithm is interesting. The system architecture is also interesting, by using 4 PlayState 4 game consoles and a single desktop PC through ethernet connection.

Indoor navigation

coRL19: Combining Optimal Control and Learning for Visual Navigation in Novel Environments. paper page
comment: learning-based perception + model-based control. waypoints are output from the perception, then fed into the model-based planner to generate smoonth feasible trajectory. strong assumption that map is known and odometry is perfect and environment is static

Mobile manipulation

coRL20sub: ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation. paper page
comment: a hybrid system integrating traditional motion planner and RL learning. Rapidly-exploring random tree (RRT) and probabilistic raod-maps (PRM) have been adopted in the motion planner. RL policy output a subgoal (desired 2D loction/end-effector 3D location). The policy has different parameterization. Experimental platform iGibson

Robot manipulation

Algorithms

IROS20: Learning High-Level Policies for Model Predictive Control. paper code
comment: loop high-level policy learning into MPC

Relevant Researchers

Davide Scaramuzza: ETH badass researcher.

Silvio Savarese: Stanford faculty, focusing on indoor navigation

Raquel Urtasun: Utoronto, Uber ATG toronto head, autonomous driving leading researcher.

Saurabh Gupta: UIUC, vision+navigation.

About

personal paper reading on neural motion planner and controller