EPOSH - Honda Research Institute USA
Introduction
Most video clips are between 15 − 60 sec long and are recorded around construction zones. For each video about 10 frames are manually selected and annotated. We annotate a total of 5, 630 perspective images.
Using COLMAP, we reconstruct a 3D dense point cloud given a video clip. We then annotate semantic labels of 3D points manually. A total of about 70, 000 BEV image / ground truth pairs are constructed.
The below showing distribution of classes and corresponding attributes in the perspective EPOSH dataset. The left subplot shows classes and the right subplot shows the corresponding attributes and affordance classes in the dataset.
Dataset Visualization
Data Format
DATA: pers # perspective annotations | video_id | img/frame_n.png # images viz-img # color coded annotations label-topo/frame_n.png # topology related annotations label-plan/frame_n.png # planning related annotations label-lane-markings/frame_n.png # lane marking annotations label-affordance/frame_n.png # affordance annotations bev # bev annotations | video_id | rectified_resized_img/frame_n.png # rectified resized images rectified_resized_img-viz/frame_n.png # color coded annotations label-topo/frame_n.png # topology related annotations label-plan/frame_n.png # planning related annotations resized_rectified_vid.mp4 # rectified resized video vid.mp4 # original resized video vids | video_id.mp4 # set of videos splits | bev_train.txt bev_val.txt pers_train.txt pers_val.txt
Please refer to the dataset readme for more details.
Download the dataset
The dataset is available for non-commercial usage. You must be affiliated with a university and use your university email address to make the request. Use this link to make the download request.
Citation
This dataset corresponds to the paper, Bird’s Eye View Segmentation Using Lifted 2D Semantic Features", as it appears in the proceedings of British Machine Vision Conference (BMVC) 2021". In the current release, the data is available for researchers from universities.
Please cite the following paper if you find the dataset useful in your work:
@inproceedings{dwivedi2021bird, title={Bird’s eye view segmentation using lifted 2D semantic features}, author={Dwivedi, Isht and Malla, Srikanth and Chen, Yi-Ting and Dariush, Behzad}, booktitle={British Machine Vision Conference (BMVC)}, pages={6985--6994}, year={2021} }