Berkeley DeepDrive is hosting three challenge tasks for CVPR 2018 Workshop on Autonomous Driving: Road Object Detection, Drivable Area Segmentation, and Domain Adaptation of Semantic Segmentation.
Submission Deadline | June 11, 2018 |
Result Announcement | June 16, 2018 |
Team | Members | Organization | Score |
IBN_PSA/P | Xingang Pan*, Hengshuang Zhao*, Jianping Shi, Jiaya Jia | CUHK, SenseTime, Tencent | 86.18 |
Mapillary Research | Lorenzo Porzi, Samuel Rota Bulò, Peter Kontschieder | Mapillary Research | 86.04 |
DiDi AI Labs | Qiaoyi Li, Zhengping Che, Guangyu Li, Ying Lu, Bo Jiang, Yan Liu | DiDi AI Labs | 84.01 |
Team | Members | Organization | Score |
Sogou_MM | Tao Wei, Bin Li, Xiaoyun Liang | Sogou | 33.10 |
ICST & Ali | Qijie Zhao, Tao Sheng, Yongtao Wang, Feng Ni, Ling Cai | Institute of Computer Science & Technology of Peking University, Ali AI Lab | 29.69 |
seb | Sebastian Bayer | Karlsruhe Institute of Technology | 20.66 |
Team | Members | Organization | Score |
NvDA | Aysegul Dundar, Zhiding Yu, Ming-Yu Liu, Ting-Chun Wang, John Zedlewski, Jan Kautz | Nvidia | 62.40 |
DiDi AI Labs | Zhengping Che, Qiaoyi Li, Ying Lu, Guangyu Li, Bo Jiang, Yan Liu | DiDi AI Labs | 57.67 |
CMU-GM | Yang Zou, Vijayakumar Bhagavatula, Jinsong Wang | Carnegie Mellon University and General Motors | 54.59 |
The competition is based on BDD100K, the largest driving video dataset to date. It contains 100,000 videos representing more than 1000 hours of driving experience with more than 100 million frames. The videos comes with GPU/IMU data for trajectory information. They are manually tagged with weather, time of the day, and scene types. We also labeled bounding boxes of all the objects on the road, lane markings, drivable areas and detailed full-frame instance segmentation. To obtain the dataset, please log in and go to the download pages. The challenges include three tasks based on BDD100K: road object detection, drivable area segmentation and full-frame semantic segmentationThere are 70,000 training and 10,000 validation images for the first two tasks. 7,000 training and 1,000 validation images are provided for the third task.
In the our portal, you can upload your results of challenges in Submission tab in the same portal of data downloading. The evaluation metric is documented in our github repo. You can use our evaluation code to check your results on validation set. We will put up the leaderboard when the challenges conclude. For all tasks, it is a fair game to pre-train your network with ImageNet, but if other datasets are used, please note in the submission description. We will rank the methods without using external datasets except ImageNet.
All the bounding boxes information is stored in the label files in .json format.
Training images: bdd100k/images/100k/train
Training labels: bdd100k/labels/100k/train
Validation images: bdd100k/images/100k/val
Validation labels: bdd100k/labels/100k/val
The evaluation will be based on testing results of the images in bdd100k/images/100k/test.
All the drivable area annotations are also stored in the label files. But we also provide “Drivable Maps”, which
provide the drivable area annotation in label map format. They can be used directly in semantic segmentation
training.
Training images: bdd100k/images/100k/train
Training labels: bdd100k/drivable_maps/100k/train
Validation images: bdd100k/images/100k/val
Validation labels: bdd100k/drivable_maps/100k/val
The evaluation will be based on testing results of the images in bdd100k/images/100k/test.
The training annotations for semantic segmentation is provided in label map format. The data name in the portal is Segmentation under BDD100K. We currently follow the
same “train_id” with Cityscapes dataset. That is, each pixel
has ids from 0-18 for training categories or 255 for ignored regions.
Training images: bdd100k/seg/images/train & Apollo training images.
Training labels: bdd100k/seg/labels/train
Validation: There is no official validation set for the video segmentation challenge.
Testing: Testing set of Video Segmentation
Challenge.
Note: We evaluate the label maps based on Cityscapes train_id, even for the Apollo data. The mapping from Apollo categories to train ids is provided. To relieve your burden for evaluation and submission, we subsample the testing images for Camera_5 by every 5th frame. Also, the testing labels for road01_ins is problematic and they won't be used for evaluation in any of the challenges. Therefore, we evaluate on 179 images in this list. The list in md5 coding is also available. Please note that our evaluation server expects the file names to be in md5 coding.