Skip to content
/ MTR Public
forked from sshaoshuai/MTR

MTR adaption: Motion Transformer with Global Intention Localization and Local Movement Refinement

Notifications You must be signed in to change notification settings

kanikel/MTR

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Motion Transformer (MTR): A Strong Baseline for Multimodal Motion Prediction in Autonomous Driving

teaser

This repository is the official implementation of the NeurIPS 2022 paper (oral presentation) "Motion Transformer with Global Intention Localization and Local Movement Refinement".

Authors: Shaoshuai Shi, Li Jiang, Dengxin Dai, Bernt Schiele

[arXiv]

Abstract

Predicting multimodal future behavior of traffic participants is essential for robotic vehicles to make safe decisions. Existing works explore to directly predict future trajectories based on latent features or utilize dense goal candidates to identify agent's destinations, where the former strategy converges slowly since all motion modes are derived from the same feature while the latter strategy has efficiency issue since its performance highly relies on the density of goal candidates. In this paper, we propose the Motion TRansformer (MTR) framework that models motion prediction as the joint optimization of global intention localization and local movement refinement. Instead of using goal candidates, MTR incorporates spatial intention priors by adopting a small set of learnable motion query pairs. Each motion query pair takes charge of trajectory prediction and refinement for a specific motion mode, which stabilizes the training process and facilitates better multimodal predictions. Experiments show that MTR achieves state-of-the-art performance on both the marginal and joint motion prediction challenges, ranking $1^{st}$ on the leaderbaords of Waymo Open Motion Dataset.

Notably, its variant MTR-A also won the Championship of Motion Prediction Challenge of Waymo Open Dataset Challenge 2022, see the official post here.

Highlights

MTR Codebase

  • State-of-the-art performance with clear code structure, easy to be extended
  • A very simple context encoder for modeling agent/map relations
  • Motion decoder with learnable queries on intention points
  • Loss with Gaussian Mixture Model for multimodal motion prediction
  • Clear data processing and organization on Waymo Open Motion Dataset
  • Local evaluation tool with official Waymo Motion Evaluation API

Method

  • Simple: pure transformer-based context encoder and motion decoder
  • Efficient: model multimodal future prediction with a small number of learnable intention queries
  • Accurate: rank 1st place at Waymo Motion Prediction leaderboard (last update: Feb 2023)

teaser

Getting Started

Main Results

Performance on the validation set of Waymo Open Motion Dataset

Model Training Set minADE minFDE Miss Rate mAP
MTR 20% 0.6697 1.3712 0.1668 0.3437
MTR 100% 0.6046 1.2251 0.1366 0.4164
MTR-e2e 100% 0.5160 1.0404 0.1234 0.3245

Performance on the testing set of Waymo Open Motion Dataset

Model Training Set minADE minFDE Miss Rate mAP
MTR 100% 0.6050 1.2207 0.1351 0.4129
MTR-A (ens) 100% 0.5640 1.1344 0.1160 0.4492

Citation

If you find this work useful in your research, please consider cite:

@article{shi2022mtr,
  title={Motion Transformer with Global Intention Localization and Local Movement Refinement},
  author={Shi, Shaoshuai and Jiang, Li and Dai, Dengxin and Schiele, Bernt},
  journal={Advances in Neural Information Processing Systems},
  year={2022}
}

@article{shi2022mtra,
  title={MTR-A: 1st Place Solution for 2022 Waymo Open Dataset Challenge--Motion Prediction},
  author={Shi, Shaoshuai and Jiang, Li and Dai, Dengxin and Schiele, Bernt},
  journal={arXiv preprint arXiv:2209.10033},
  year={2022}
}

About

MTR adaption: Motion Transformer with Global Intention Localization and Local Movement Refinement

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 76.9%
  • Cuda 14.1%
  • C++ 8.4%
  • Other 0.6%