Abstract

Datasets drive vision progress, yet existing driving datasets are impoverished in terms of visual content and supported tasks to study multitask learning for autonomous driving. Researchers are usually constrained to study a small set of problems on one dataset, while real-world computer vision applications require performing tasks of various complexities. We construct BDD100K, the largest driving video dataset with 100K videos and 10 tasks to evaluate the exciting progress of image recognition algorithms on autonomous driving. The dataset possesses geographic, environmental, and weather diversity, which is useful for training models that are less likely to be surprised by new conditions. Based on this diverse dataset, we build a benchmark for heterogeneous multitask learning and study how to solve the tasks together. Our experiments show that special training strategies are needed for existing models to perform such heterogeneous tasks. BDD100K opens the door for future studies in this important venue.

Keywords

Computer scienceBenchmark (surveying)Machine learningArtificial intelligenceMulti-task learningConstruct (python library)Set (abstract data type)Training setTask analysisTask (project management)

Affiliated Institutions

Related Publications

Publication Info

Year
2020
Type
article
Pages
2633-2642
Citations
1978
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

1978
OpenAlex
300
Influential
1754
CrossRef

Cite This

Fisher Yu, Haofeng Chen, Xin Wang et al. (2020). BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 2633-2642. https://doi.org/10.1109/cvpr42600.2020.00271

Identifiers

DOI
10.1109/cvpr42600.2020.00271
arXiv
1805.04687

Data Quality

Data completeness: 84%