Switch branches/tags
Nothing to show
Find file History


The Online Self-localization Challenge of Apolloscape

This repository contains the evaluation scripts for the online self-localization challenge of the ApolloScapes dataset, Where we extended the dataset with more scenes and 100x large data including recorded videos under different lighting conditions, i.e. morning, noon and night, with stereo pair of images. A test dataset for each new scene will be withheld for benchmark. (Notice we will not have point cloud for the very large data due to size of dataset)

Details and download of data from different roads are available, please follow the most updated content in this page rather than apolloscape.auto homepage which is outdated. Here are some interesting facts:

For each road, we record it by driving from start-to-end and then end-to-start at different day times, which means at each site along the road, a scene will be looked at from two opposit directions. We provide the set of record id recorded from start-to-end and the set of record id from end-to-start in training set for each road at LoopDirection. One may discover the corresponding images from the camera pose we provided.

In this challenge, we recard records from forward (start-to-end) and inverse (end-to-start) driving as records from two different roads, which means we will not have forward videos as training while have inverse driving as testing videos. However, it could be interesting to do that in your research as showed in the work of Semantic Visual Localization.

Dataset Structure

The dataset has the following structure

{split}/{scene_names}/{data_type}/{record time}/{record id}/{camera id}/{image_name}

split is the split of the dataset:

  • Train: include the data for training, including image, pose, split.
  • Test: include the image for testing, including image without pose ground truth.

scene_names include a sample scene:

  • Road11: which is road id where the dataset is collected. Due to large amount of images, you need to download train and test image sets for each Roadxx separately.

data_type includes:

  • image: the RGB image from the dataset
  • pose: the abosolute pose (roll,pitch,yall,x,y,z) of each image related to a map (Notice this is converted from the 4x4 pose matrix from Apolloscape dataset)
  • split: train val split

record time: the time stamp for recorded videos, i.e. BJxxxB is one set of collections at site BJ

  • Important message: due to different starting points when recording at different day times, so the relative offset of pose at same road under different time stamps are different, yielding non comparable poses. Thus we release the relative offset between different recording times at LoopDirection.
  • For example, for Road14, one may transfer the pose in time stamp BJxx1B to the pose in time stamp BJxx1D by: trans(BJxx1B) + trans_off(BJxx1B -> BJxx1D) = trans(BJxx1D)

record id: each sequence, i.e. Recordxxx is a video sequence of the corresponding scene and the images are sorted numerically.

camera_id: each scene we provide images recorede by two camera facing front side, e.g. Camera_5 and Camera_6 There is a camera-name in-consistency of the device between the two scene, which will be fixed for the larger dataset later. The camera parameters is consistent with that in other datasets released in https://Apolloscape.auto. ([fx, fy, cx, cy] are also available at data.py

Here split include the train and val image names for each scene, where val images are recorded at different period with training images.

image_name: the name of the file, images will be {timestamp}_{camera id}.jpg

There is only one pose file (i.e., pose.txt) for each camera and each record. This pose file contains all the extrinsic parameters for all the images of the corresponding camera and record. The format of each line in the pose file is as follows:

image_name row,pitch,yall,x,y,z

Similar data structure is described in apolloscape.auto/scene.html, while having the pose saved in a 4x4 matrix, a conversion code from 4x4 matrix to 6 DOF is provided in utils of this toolkit. Later we will also release semantic labels, and semantic 3d point cloud python toolkit to render 3d point to 2d image for visualizing the semantic points. You may download the dataset from self-localization. The sample data is used for paper

   title={DeLS-3D: Deep Localization and Segmentation with a 3D Semantic Map},
   author={Wang, Peng and Yang, Ruigang and Cao, Binbin and Xu, Wei and Lin, Yuanqing},

Background Pointcloud

At download page, we also provide backgroud point cloud for supporting the localization task for each road.

You may download the point cloud to point_cloud under each road directory {scene_names} as described in above data structure.
Since the point cloud for each road could be very large, we partition it to several parts, and they are saved following the dir structure:

Train/{scene_names}/point_cloud/{record time}/Part{id}.pcd

id here is the index of partition for each set of point cloud. One may perform projection from point cloud to a depth map based on the given pose for corresponding camera and road.

  • proj_point_cloud.py is an example opengl rendering program that is wrapped by cython to demonstrate the projection/rendering process.

For rendering a depth map for each image, if your computer memory is large enough > 20G, you may load all the parts of each road to be a whole set of point cloud for rendering. Otherwise, you may render depth map from each part of point cloud and merge all the rendered depth maps by taking care of the zbuffer, i.e. select the minimum depth value at each corresponding pixel.


There are several scripts included with the dataset in a folder named scripts

  • eval_pose.py Code for evalution pose accuracy based the commonly used eval metric of meidian translation and rotation error.

Code for test evaluation:

python eval_pose.py --test_dir='./test_eval_data/Test' --gt_dir='./test_eval_data/Test_gt' --res_file='./test_eval_data/res.txt'

Metric formula

For each image, given the predicted rotation and translation of image , and the ground truth and , the metric for evaluation is defined as:

where is the quaternions representation of the Euler angle row, pitch, yall

Rules of ranking

Result benchmark will be:

Method mean scene1 scene2 scene3
Deepxxx xx , xx xx , xx xx , xx xx , xx

Our ranking will determined by number of winning metrics from all scenes.

Submission of data format

Please follow the data format under test_eval_data/ for example.

The submission structure for test folder is:

Test/{scene_names}/pose/{record time}/{record id}/{camera id}.txt

We only ask for the pose of a single camera, i.e. Camera_5 in this case

  • Example dir tree of submitted zip file
├── Test
│   ├── Road1x
│   │   ├── pose
│   │   │   ├── xxxxx
│   │   │   │   ├── Recordxxx
│   │   │   │   │   ├── Camera_5.txt
│   │    ...
│   ├── Road1x
│   │   ├── pose
│   │   │   ├── xxxxx
│   │   │   │   ├── Recordxxx
│   │   │   │   │   ├── Camera_5.txt
  • Example format of {camera id}.txt
image_name1 roll,pitch,yaw,x,y,z
image_name2 roll,pitch,yaw,x,y,z
image_name3 roll,pitch,yaw,x,y,z
image_name4 roll,pitch,yaw,x,y,z
image_name5 roll,pitch,yaw,x,y,z

Here roll,pitch,yaw,x,y,z are float32 numbers


Please feel free to contact us, or raise an issue with any questions, suggestions or comments: