Just Label What You Need: Fine-Grained Active Selection for P&P Through Partially Labeled Scenes

Abstract

Self-driving vehicles must perceive and predict the future positions of nearby actors to avoid collisions and drive safely. A deep learning module is often responsible for this task, requiring large-scale, high-quality training datasets. Due to high labeling costs, active learning approaches are an appealing solution to maximizing model performance for a given labeling budget. However, despite its appeal, there has been little scientific analysis of active learning approaches for the perception and prediction (P&P) problem. In this work, we study active learning techniques for P&P and find that the traditional active learning formulation is ill-suited. We thus introduce generalizations that ensure that our approach is both cost-aware and allows for fine-grained selection of examples through partially labeled scenes. Extensive experiments on a real-world dataset suggest significant improvements across perception, prediction, and downstream planning tasks.

Cite

Text

Segal et al. "Just Label What You Need: Fine-Grained Active Selection for P&P Through Partially Labeled Scenes." Conference on Robot Learning, 2021.

Markdown

[Segal et al. "Just Label What You Need: Fine-Grained Active Selection for P&P Through Partially Labeled Scenes." Conference on Robot Learning, 2021.](https://mlanthology.org/corl/2021/segal2021corl-just/)

BibTeX

@inproceedings{segal2021corl-just,
  title     = {{Just Label What You Need: Fine-Grained Active Selection for P&P Through Partially Labeled Scenes}},
  author    = {Segal, Sean and Kumar, Nishanth and Casas, Sergio and Zeng, Wenyuan and Ren, Mengye and Wang, Jingkang and Urtasun, Raquel},
  booktitle = {Conference on Robot Learning},
  year      = {2021},
  pages     = {816-826},
  volume    = {164},
  url       = {https://mlanthology.org/corl/2021/segal2021corl-just/}
}