We present the most diverse in number of images/labels/classes fisheye synthetic dataset with source codes and models. As well as a benchmarking testing real dataset.
COHI-O365 is a benchmark testing dataset for object detection in hemispherical/fisheye for field of view invariant applications. It contains 1,000 real fisheye images of 74 classes sampled from the Objects365 dataset with 20.798 object instances per image on average. The images were captured using a hemispherical camera ELP-USB8MP02G-L180 with the 2,448 by 3,264 pixel resolution and manually annotated with standard axis-aligned bounding boxes afterward. The samples of raw images from the COHI-O365 dataset are shown below
The names of sampled classes and the number of bounding boxes for each class are presented in the next figure (to be inserted).
To train object detection models for the COHI-O365 dataset, the Fisheye versions of Objects 365 dataset was generated by applying a non-linear mapping to obtain fisheye-looking images. We called it the RMFV365 dataset, comprising 5.1 million images that encompass a broad spectrum of perspectives and distortions. This dataset serves as a valuable resource for training the model in generalized object recognition tasks. The samples of raw and annotated images are illustrated below.
One can access the COHI-O365 dataset using the following Google Drive link: COHI-O365
The RMFV365 dataset can be downloaded with a link: (Link is coming!!!). Alternatively it can be generated from Objects365 dataset using the python scripts in transformations directory for image transformation and coordinates directory for coordinate mapping
- numpy
- PIL
- pandas
We used YOLOv7 to train and evaluate object detection models. All needed information can be found on their official GitHub page YOLOv7.
We trained the YOLOv7 model with 36.9 M parameters on three datasets, namely Objects365, RMFV365, and a variant of RMFV365 codenamed RMFV365-v1 and evaluated the performance of models with our benchmark testing dataset - COHI-O365.
- YOLOv7-0: trained on the Objects365 dataset
- YOLOv7-T1: trained on Objects365 dataset and fisheye images transformed using lens and camera independent fisheye transformation with parameter n = 4, codenamed RMFV365-v1
- YOLOv7-T2: trained on RMFV365
mAP50 results are summarized in the table below.
S/N | Model | Test Results (%) | |||||||
---|---|---|---|---|---|---|---|---|---|
Objects365 | RMFV365-v1 | RMFV365 | COHI-365 | ||||||
mAP50 | mAP50:95 | mAP50 | mAP50:95 | mAP50 | mAP50:95 | mAP50 | mAP50:95 | ||
1 | FPN | 35.5 | 22.5 | N/A | N/A | N/A | N/A | N/A | N/A |
2 | RetinaNet | 27.3 | 18.7 | N/A | N/A | N/A | N/A | N/A | N/A |
3 | YOLOv5m | 27.3 | 18.8 | 22.6 | 14.1 | 18.7 | 10.1 | 40.4 | 28.0 |
4 | YOLOv7-0 | 34.97 | 24.57 | 29.1 | 18.3 | 24.2 | 13.0 | 47.5 | 33.5 |
5 | YOLOv7-T1 | 34.3 | 24.0 | 32.7 | 22.7 | 32.0 | 22.0 | 49.1 | 34.6 |
6 | YOLOv7-T2 | 34 | 23.1 | 32.9 | 23 | 33 | 22.8 | 49.9 | 34.9 |
Table: Objects recognition results on Objects365, RMFV365-v1, RMFV365, and COHI-365 Testing Sets
Pre-trained model weights can be downloaded using a Google Drive link: (Link is Coming)
.........................