Welcome to the SeePerSea Dataset from the Dartmouth Reality and Robotics Lab! Our paper is currently under review for IEEE Transaction on Field Robotics (T-FR)!
our dataset website: https://seepersea.github.io/
This paper introduces the first publicly accessible labeled multi-modal perception dataset for autonomous maritime navigation, focusing on in-water obstacles within the aquatic environment to enhance situational awareness for Autonomous Surface Vehicles (ASVs). This dataset, collected over 4 years and consisting of diverse objects encountered under varying environmental conditions, aims to bridge the research gap in ASVs by providing a multi-modal, annotated, and ego-centric perception dataset, for object detection and classification. We also show the applicability of the proposed dataset by training deep learning-based open-source perception algorithms that have shown success in the autonomous ground vehicle domain. We expect that our dataset will contribute to the development of future marine autonomy pipelines and marine (field) robotics.
Jeong, Mingi* and Chadda, Arihant* and Ren, Ziang and Zhao, Luyang and Liu, Haowen and Roznere, Monika and Zhang, Aiwei and Jiang, Yitao and Achong, Sabriel and Lensgraf, Samuel and Alberto Quattrini Li
* these authors equally contributed.
- Mingi Jeong: mingi.jeong.gr@dartmouth.edu
- Arihant Chadda: ari.chadda@gmail.com
- SeePerSea, the first LiDAR-camera dataset in aquatic environments with object labels across two modalities, will foster the development of robust fusion perception pipelines for ASV autonomy.
- SeePerSea, covering various environments and day conditions, will help ensure that developed perception pipelines are increasingly generalizable.
The following environment is tested for the GUI viewer for the sample dataset.
- Ubuntu >= 20.04 (tested environment)
- PyQt5 5.14.1
- Open3D 0.13.0
- Numpy 1.24.4
- openCV-python 4.10.0.84
- python 3.8.10
- camera_label (.json): object labels based on camera ( AnyLabeling ) -- 6 digit file name per frame number
- annotations (.json): object labels based on LiDAR ( 3d-bat ) -- 6 digit file name per frame number
- calib (.txt): camera intrinsic and lidar extrinsic calibbration info -- 6 digit file name per frame number
- P2: camera matrix (3 by 3)
- Tr_velo_to_cam: extrinsic from LiDAR to camera (3 by 4)
- timestamp (.txt): camera and LiDAR data's timestamp as synched. each frame's timestamp
- point_clouds (.pcd): point clouds data -- 6 digit file name per frame number
- undist_images (.png): undistorted camera images -- 6 digit file name per frame number
- motion_data (.csv): gps and imu data, respectively. Each sensor contains its own timestamp and correspondingly recorded data.
Note: labels for testsplit are empty.
- Please download sample data from the following link: ( Google-drive )
- git clone the current repo
- install dependencies (we recommend Poetry and have included a
pyproject.toml
for convenience; runpoetry shell
and thenpoetry install
). python view_images.py
- Click the
Load Folder
button to load a sample sequence (mascoma-20221007-1560) - You can see a sample image with groudtruth bounding boxes overlaid.
- Click
Show 3D View
to visualize point clouds also with groudtruth bounding boxes. Open3D window will pop-up. - If you want to move to the next frame, close the Open3D window and click the
Up
button. - You are welcome to explore the dataset using the
Up
,Down
, andShow 3D View
as well as load other sequences to explore the dataset.
- mascoma-20221007-1560: many boats and buoys
- sunapee-20210916-1440: 3 kayaks, near and moving, overlapping
- barbados-202302220530-200: lots of buoys in line (more than 20)
All datasets and benchmarks on this page are copyright by us. https://creativecommons.org/licenses/by-nc-sa/4.0/