This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Time and Cost-Efficient Bathymetric Mapping System using Sparse Point Cloud Generation and Automatic Object Detection

Andres Pulido Mechanical and Aerospace Engineering
University of Florida
Gainesville, FL, United States
[email protected]
   Ruoyao Qin Mechanical and Aerospace Engineering
University of Florida
Gainesville, FL, United States
[email protected]
   Antonio Diaz Mechanical and Aerospace Engineering
University of Florida
Gainesville, FL, United States
[email protected]
   Andrew Ortega Geomatics Program
University of Florida
Gainesville, FL, United States
[email protected]
   Peter Ifju Mechanical and Aerospace Engineering
University of Florida
Gainesville, FL, United States
[email protected]
   Jaejeong (Jane) Shin Mechanical and Aerospace Engineering
University of Florida
Gainesville, FL, United States
[email protected]
Abstract

Generating 3D point cloud (PC) data from noisy sonar measurements is a problem that has potential applications for bathymetry mapping, artificial object inspection, mapping of aquatic plants and fauna as well as underwater navigation and localization of vehicles such as submarines. Side-scan sonar sensors are available in inexpensive cost ranges, especially in fish-finders, where the transducers are usually mounted to the bottom of a boat and can approach shallower depths than the ones attached to an Uncrewed Underwater Vehicle (UUV) can. However, extracting 3D information from side-scan sonar imagery is a difficult task because of its low signal-to-noise ratio and missing angle and depth information in the imagery. Since most algorithms that generate a 3D point cloud from side-scan sonar imagery use Shape from Shading (SFS) techniques, extracting 3D information is especially difficult when the seafloor is smooth, is slowly changing in depth, or does not have identifiable objects that make acoustic shadows. This paper introduces an efficient algorithm that generates a sparse 3D point cloud from side-scan sonar images. This computation is done in a computationally efficient manner by leveraging the geometry of the first sonar return combined with known positions provided by GPS and down-scan sonar depth measurement at each data point. Additionally, this paper implements another algorithm that uses a Convolutional Neural Network (CNN) using transfer learning to perform object detection on side-scan sonar images collected in real life and generated with a simulation. The algorithm was tested on both real and synthetic images to show reasonably accurate anomaly detection and classification.

Index Terms:
Bathymetry, Drone, Sonar, Autonomous System, 3D Point Cloud, Object Detection

I Introduction

Bathymetric maps take an important role in many applications by providing information of the underwater environment including water depth and geometry of the seafloor [1, 2]. For example, in [3], the authors obtained bathymetric mapping using LIDAR to determine nearshore benthic habitat complexity, which can assist ocean activity conservation and management. In [4], the authors introduces bathymetric and gravity technologies to reduce submarine fleets’ dependence on GPS navigation. In [5], the authors present a terrain-based underwater navigation using sonar bathymetric profiles. The natural change of seafloor can also be discovered by bathymetric mapping. An example work is shown in [6], where the authors used a bathymetric map to determine the relationship between natural change of seafloor to long-term shoreline change in century-scale. From this finding, shoreline change could be predicted by monitoring the change of seafloor, and this finding can help prevent erosion and reduce the risk of flooding. Bathymetric data also plays an important role in generating an ocean circulation model [7], as the data can help researchers generate climate models and predict some global phenomena.

Although many applications require and benefit from bathymetric map data, obtaining such maps is often very time-consuming and costly. The most commonly-used sensor for bathymetric mapping is multi-beam echo-sounders, which are very costly because the sensors can send out multiple sound waves in a fan-shaped pattern at once and interpret the reflected signals. Moreover, multi-beam echo-sounders and other underwater sonar methods are generally done by a vessel operating on the surface of the sea [7]. One limitation of these methods is that the vessel must travel in a certain path which can only scan a limited region and also makes the job time-consuming. Another common way to obtain bathymetry in shallow clear water is to use lidar sensors with remote sensing technology. Lidar sensors transmit lasers from airborne platforms and measures their return [7]. This operation is usually performed using helicopters, satellites, and airplanes, which are not only expensive but also not able to map the sea floor with high resolution when the water is too deep or the water visibility is low.

Moreover, most existing bathymetric platforms focus on investigating a large scale area, such as ocean regions [8]. In order to provide quality bathymetry data for local reservoirs and retention ponds and overcome the aforementioned limitations, the authors in this paper developed a novel platform, named Bathy-drone, that can conduct bathymetric mapping for pond-scale bodies of water [9]. The proposed platform uses low-cost side-scan sonar sensors that are used in fish finders, which are generally available in affordable price ranges. Although the cost of these sensors is relatively lower than other methods, the sensors can still provide high-resolution bottom surface imagery [10, 11]. These type of sensors are proven to be useful in related applications, including marine life habitat tracking research [12]. The Bathy-drone system is also designed to be lightweight such that a large transporting platform like a boat or an airplane is not required, and thus, the system requires reduced cost of operation. Autonomous or semi-autonomous systems can further reduce the number of people needed for operating the system and processing the sonar data. Furthermore, automation of the process and remote controlling of the system can access some region that is inaccessible or dangerous to operators [7].

The Bathy-drone platform requires an efficient and automatic mapping software, because the platform needs to be lightweight. The software system must have a very efficient sonar image processing algorithm given limited computation power and memory. Therefore, this paper proposes a computationally efficient algorithm that can map bathymetry using point cloud (PC). This paper presents our work on a bathymetric mapping system using sparse point cloud generation and automatic object detection with side-scan sonar. The proposed bathymetric mapping system consists of two algorithms. The first algorithm generates a sparse 3D point cloud data using side-scan sonar images, down-scan sonar depth readings, and GPS measurements from the sonar sensor. The second algorithm detects objects from the side-scan sonar images and outputs bounding boxes around the detected objects using YOLOv4 architecture and transfer learning [13]. In order to overcome the lack of existing sonar images for training, we also generated some synthetic sonar data using Unreal Engine as well as real-life sonar data to perform fast and online side-scan sonar object detection. Real-life sonar data was acquired by the use of the Bathy-drone, a platform designed and manufactured by the University of Florida [9].

This paper is organized in four sections. In Section II, the Bathy-drone platform is introduced with its specifications, such as the sensors equipped in the platform and the sensing resolutions. In Section III, the efficient point cloud generation for bathymetric mapping is presented. The detail of the developed algorithms for object detection and sparse point cloud generation is included. In Section IV, the field experiment results are presented. The authors performed the field experimentation at a retention pond at Citra, FL. The generated point cloud data of the pond is presented in this section.

II Bathy-drone: Time- and Cost-Efficient Bathymetric Mapping Platform

The Bathy-drone (Fig. 1) is a semi-autonomous system that incorporates a drone towing a tethered vessel that can be equipped with a variety of sensors such as sonar or underwater cameras [9]. It has advantages over USVs since the system can be flown to the survey location; thus, many surveys can be initiated from a land-based ground station, and no boats or boat ramps are needed if the location is within the FAA’s required visual line-of-sight. Additionally, since the vessel has no propulsion system (propellers) entanglement from flotsam, underwater or floating vegetation, does not hamper its operation. This allows it to navigate in small bodies of water, such as man-made and natural ponds, rivers, canals, low draft areas such as marshes and swamps, boat basins, shipping lanes, pre-construction and monitoring of marine infrastructure, and nearshore applications. The novel vehicle meets the industry and research need of developing specialized unmanned systems that are practical, inexpensive, easy to deploy and provide high spatial and temporal resolution for bathymetry and back-scatter.

Refer to caption
Figure 1: A photo of Bathy-drone system collecting data in a retention pond in Citra, FL.

The Bathy-drone can collect bathymetry, sonar imagery, and bottom hardness while traveling at speeds of 0-24 km/h (0-15 mph) [9]. The tethered configuration and travel speed allow for use in waters with swift currents. An area of more than 10 acres was surveyed using the Bathy-drone on one drone battery charge and in less than 25 min[9]. The vessel is of fiberglass composite construction, with a soft-edge skiff-like planing hull shape. This shape proved robust in both straight/level tracking and corner turns once a trim plate and fins were added. Additional considerations in the design included providing adequate volume in the hull to house the sonar console and ports to provide easy access to the microSD cards, batteries, and sonar console keypad. The sonar unit is a low-cost commercial off-the-shelf recreational fish-finder by Lowrance, model Elite ti7, with an active scan transducer. This transducer provides side-scan and down-scan capability at multiple frequencies. The current Bathy-drone vessel and electronics payload weighs approximately 14 lbs. Heavy lift drones such as the DJI Matrice 600 used for this project can easily lift the vessel to the water surface. When testing for accuracy compared to the traditional state-of-the-art survey methods, the Bathy-drone produces contour plots where mean residual is −2.64 cm, the median is 0.95 cm, and the standard deviation is 16.98 cm[9]. When comparing repeated missions of the Bathy-drone for precision, the resulting residuals are summarized as a mean of 21.6 cm, a median of 18.7 cm, and standard deviation of 16.7 cm[9]. The resolution of bottom geometry is dependent on many factors such as shape, texture and density, but qualitatively objects as small as 0.5m (2-3ft) have been observed.

III Sparse Point Cloud Generation and Automatic Object Detection using Bathy-drone

The down-scan depth and position data collected from the Bathy-drone’s sonar scan can be used to generate underwater topographic maps called isobaths which are then converted to a point-cloud structure that can be helpful to understand the bottom geometry of the body of water. Point clouds can be used as a representation of bathymetry because they can be manipulated and stored with relatively low computational cost. However, only using the down-scan sonar depth only returns depth values at the path of the sonar. Therefore using the side-scan sonar images to generate additional points is useful to better represent the bathymetry. Additionally, during operation, detecting objects on the seabed using the side-scan sonar image also provides value to better understand the underwater environment. The same techniques to find the location of new points to the point cloud can be used to then locate the objects detected in the bathymetric map.

III-A Sparse Point Cloud Generation from Echo-sounder

The algorithm presented in this paper generates a 3D point cloud in a computationally fast manner using only the side-scan sonar image with known position provided by GPS (the path of the boat, Fig. 2), and down-scan sonar depth measurement at each data point. To accomplish this the side-scan image is assumed to have a property of linear mapping between distances to pixels on the port and starboard side. The output of the algorithm is an unordered point cloud data structure of x,y,zx,y,z points in the East North Up (ENU) coordinate system.

Refer to caption
Figure 2: Path of the boat in ENU with the sampled depth from down-scan sonar

For the Bathy-drone system, transformed GPS coordinates of latitude and longitude are used to find the planar (x,y)(x,y) position of the vehicle in time using the Spherical Pseudo-Mercator projection. The depth measurement provided by the down-scan gives the third coordinate to construct a sparse 3D PC. To increase the density of the PC, especially outside of the path traveled by the vehicle, the first return in side-scan sonar image is also used to generate additional points. The first return is the first reading of the water floor from each sonar beam. The first reading can be visually seen in a side-scan sonar image as the first brightly colored pixel after the ‘dead zone’ in the middle of the image.

The first return of sonar reflection will be calculated using the sonar image (Fig. 3). The raw image provided by the Lowrance fish-finder is a two-dimensional array of size 700×K700\times K, where KK is a pixel number that grows as the sonar samples more points in a single image. The returned sonar intensity values are within the range of I[0,2.14]×109I\in[0,2.14]\times 10^{9} (Fig. 4(a)). However, it is noted that the sonar image can have negative returned intensity near the middle of the deadzone (Fig. 3) where the return is a pure reflection of the sonar itself. Therefore, the middle line values are set to zero in order to have only positive intensity. The intensity is scaled to be in between 0 and 1 in order to be able to generalize to other sonar sensors and have intuitive understanding of the intensity values as shown in Fig. 4(b).

Refer to caption
Figure 3: Pixel distance in the side-scan sonar image that is converted to r1r_{1} with the PPDPPD ratio
Refer to caption
(a) Raw intensity values of a side-scan sonar image column
Refer to caption
(b) Scaled intensity values, and set the intensity values inside the black dots to zero
Figure 4: Raw and normalized intensity values from the side-scan sonar

After the data is processed, the algorithm starts by using thresholding in the returned intensity to detect the first return. The algorithm finds the number of pixels from the middle of the image, the path of the robot, to the edge of the lighter-colored selection and maps those pixels to a distance in meters, r1r_{1}, in Fig. 5). Using the assumed property of linear scaling between the pixels in the image and the distances in the image, a pixels per distance, denoted by PPDPPD, is calculated by dividing pp, the number of pixels for the first return, by dd, the depth returned by down-scan sonar on a point in the map where the depth of the floor is known and the neighboring depth values are as close to flat as possible (1). We determined this point by looking at the flattest portion of the pond in the depth-map interpolated using the down-scan sonar data. With the PPDPPD value, and the α2\alpha_{2} angle found in the specification sheet provided by the sonar manufacturer, the depth at the first return on the starboard and the port side is calculated.

PPD=(pixels)(distance)=pdPPD=\frac{\text{(pixels)}}{\text{(distance)}}=\frac{p}{d} (1)
Refer to caption
Figure 5: Geometry of the side-scan sonar beam used to get the first return

The calculated depth is associated with a sampled point in the down-scan sonar, however, the point is in the reference frame of the boat, so it had to be transformed to the ENU coordinate system. To accomplish this, the heading 2 of the boat was calculated to be the direction of the ii sampled point to the i+1i+1 sampled point,

𝐡𝐢=𝐩𝐢+𝟏𝐩𝐢\mathbf{h_{i}}=\mathbf{p_{i+1}}-\mathbf{p_{i}} (2)
𝐭𝐢=𝐡𝐢×(𝐳^)where𝐳^=[0,0,1]\mathbf{t_{i}}=\mathbf{h_{i}}\crossproduct(\mathbf{-\hat{z}})\quad\textnormal{where}\quad\mathbf{\hat{z}}=[0,0,1] (3)
𝐭𝐢^=𝐭𝐢|𝐭𝐢|\mathbf{\hat{t_{i}}}=\frac{\mathbf{t_{i}}}{|\mathbf{t_{i}}|} (4)

Then the perpendicular vector 3, the direction where the side-scan sonar images are taken (Fig. 6), and the position of the boat at the sample point are used to project the distance in the starboard and the port side of the point to ENU using the distance yiy_{i} (6), seen in Fig. 5, and found using trigonometry.

zi=\displaystyle z_{i}= ri,1cos(α2)\displaystyle r_{i,1}\cos(\alpha_{2}) (5)
yi=\displaystyle y_{i}= ri,1sin(α2)\displaystyle r_{i,1}\sin(\alpha_{2}) (6)
𝐪(starboard,𝐢)=\displaystyle\mathbf{q_{(\text{starboard},\,i)}}= 𝐩𝐢+𝐭𝐢^yi\displaystyle\mathbf{p_{i}}+\mathbf{\hat{t_{i}}}\,y_{i} (7)
𝐪(port,𝐢)=\displaystyle\mathbf{q_{(\text{port},\,i)}}= 𝐩𝐢𝐭𝐢^yi\displaystyle\mathbf{p_{i}}-\mathbf{\hat{t_{i}}}\,y_{i} (8)
Refer to caption
Figure 6: Heading and perpendicular vector geometries based on the boat position measurements

The result is a 2D vector 𝐪(𝐣,𝐢)\mathbf{q_{(j,\,i)}} (7) and (8) which is the position in the plane parallel to the surface of the water of the additional generated points from starboard and port side. To get the (x,y,z)(x,y,z) ordered list of the PC we append the ziz_{i} depth (5).

The final step of the process is to apply an outlier removal algorithm. The algorithm removes points that have few neighbors in a given sphere around them [14]. Two parameters can be changed to vary the minimum number of points inside the sphere and the radius of the sphere.

III-B Automatic Object Detection Algorithm

To perform the object detection task, the proposed work uses You-only-look-once(YOLO) structure. YOLOv4 is the fourth generation of the algorithm that uses neural networks to provide real-time object detection, which has the capability of detecting single or multiple objects in a single picture and predicting a bounding box around the objects. YOLOv4 consists of CSPDarknet53 as the backbone of the algorithm, spatial pyramid pooling (SPP) and Path Aggregation Network (PANet) as the neck, and YOLOv3 as the head. The SPP block over the CSPDarknet53 can increase the receptive field drastically and separate out the most significant context feature while having no negative impact on the operation speed. The PANet is the path-aggregation neck which is used as a method of parameter aggregation from different backbone levels for different detector levels [13]. One of the reasons the YOLO architecture was chosen is that there has been previous work with previous versions (YOLOv2) to perform object detection using side-scan and forward-looking sonar images [15, 16].

Refer to caption
Figure 7: Example of side-Scan sonar image obtained from real underwater environment used for training

In the network training process, side-scan sonar images obtained from Bathy-drone platform (Fig. 7) and synthetic sonar images generated by a sonar image simulation are used. The synthetic data are used in the training process in order to obtain a batch of training images, which are not available from physical experiments due to time and cost constraints. The gaming framework, Unreal Engine (UE) is used as a simulation tool to generate synthetic sonar imagery data. The UE simulation models the way the sound waves get propagated underwater and then reflected by an object by approximating it as a light beam being generated from an overhead camera. An example of the generated synthetic images from the UE simulation is shown in Fig. 8(a). To increase the similarity between synthetic and real sonar images, a post-processing step was added. Specifically, the images are converted into grey scale, and then, speckle noise is added to the images. Fig. 8(b) shows an example of the post-processed image.

Refer to caption
(a) Example Image of an object obtained from Unreal Engine
Refer to caption
(b) Image obtained from simulation after pre-processing
Figure 8: Raw and pre-processed image generated in the UE simulator

Images obtained from UE after the pre-processing and real environment will form a data set that was divided into training, validation, and test set. There are 332 images in the training set, 41 images in the test set, and 33 images in the validation set. The performance of the YOLO network was evaluated with mean average precision (mAP) which is a common metric that evaluates the object detection model [17]. To calculate the mAP metric, average precision(AP) must be found for each class, and then, the average over the number of classes is taken.

mAP=1Ni=1NAPmAP=\frac{1}{N}\sum_{i=1}^{N}AP (9)

The Intersection over Union (IoUIoU) is another metric that is used to evaluate the network [18, 17]. IoU represents the overlap of the predicted bounding box to the ground truth box.

IoU=(Area of Overlap)(Area of Union)IoU=\frac{\text{(Area of Overlap)}}{\text{(Area of Union)}} (10)

A higher IoU metric means the predicted bounding box is more similar to the ground truth bounding box. A confusion matrix is also used to determine the detection performance of the network.

IV Experimental Results

IV-A Sparse Point Cloud Generated from Echo-sounder

The first return calculated by the algorithm of a side scan sonar image strip can be seen in Fig. 9. The threshold value is set as 0.3. This threshold value needs to be between 0 and 1 because the side scan sonar image intensity values are scaled such that all values are ranged from 0 to 1. The first return for the entire side-scan sonar image of a run is shown in Fig. 10.

Refer to caption
Figure 9: Side-scan sonar image strip with the first return colored red (starboard) and blue (port)
Refer to caption
Figure 10: Side-scan sonar image of a complete run with the first return colored red (starboard) and blue (port)

The point cloud generated by only using the down-scan sonar can be seen in Fig. 11. The sparse PC generated using the algorithm explained above can be seen in Fig. 12, where the red and blue points are the ones calculated by the first return of the starboard and port sides respectively. The sparse point clouds can be seen to be less dense than the down-scan sonar cloud, a reason for this is that there are many points sampled at the same position and therefore the heading vector (and perpendicular vector) were found to be zero and no sparse point was calculated. Another reason for the sparsity is because approximately 10% of the points are removed by the outlier removal algorithm [14].

Refer to caption
Figure 11: Point cloud data that are generated only using depth data from the path in Fig. 2
Refer to caption
Figure 12: Sparse PC generated using the depth data and side-scan sonar images. The black points represent the points generated from the depth data, and the red and blue points represent the points generated from the starboard and port side of side-scan sonar images, respectively.

IV-B Automatic Object Detection Algorithm

Fig. 13 presents object detection results on a rock underwater that was generated from UE. Table II presents the evaluation of the model with metrics of mAP and IoU and Table I shows the confusion matrix values, in which True Positive (TP) means how many objects that the model predicts its positive class correctly. False Positive (FP) means how many objects the model predicts positive class incorrectly, and False Negative (FN) means how many objects the model predicts the negative class incorrectly.

Refer to caption
Figure 13: Objects detection on rock simulation image
Refer to caption
Figure 14: Object detection on in real Image
Positive Negative
Positive 34 46
Negative 0 0
TABLE I: Confusion matrix
Metrics Value
IoU 76.26%
mAP 93.17%
TABLE II: Metrics to evaluate the model

Fig. 14 presents how the YOLOv4 algorithm performs object detection on the image obtained from UE. The number presented next to the box represents the object detection confidence level from the algorithm. The confidence level is high when the algorithm is tested using synthetic images as shown in Fig. 14. However, the object detection confidence score drops on the images obtained from the real environment. The drop could be mitigated by collecting more actual sonar images of objects with the sides-scan sonar.

V Conclusions

This paper presents a computationally efficient point cloud generation algorithm for bathymetric mapping. The proposed algorithm is developed for mapping pond-scale water bodies using a low-cost platform named Bathy-drone. This Bathy-drone platform consists of a drone and an unpowered boat that is tethered to the drone. The boat includes a low-cost fish finder that includes down-scan and side-scan sonar sensors. The proposed algorithm generates a sparse 3D point cloud of the bathymetry using the down-scan and side-scan sonar measurement. The proposed algorithm first generates a sparse point cloud based on the depth data collected from the down-scan and the first returned signal of side-scan sonar sensors. Then, the algorithm performs automatic object detection from a side-scan sonar sensor in order to capture details in the pond. The proposed algorithm is tested based on the data obtained using Bathy-drone at a retention pond in Citra, FL. Future work will be dedicated to generate point cloud information of the objects detected using the side-scan sonar image. Shading techniques and a priori information of objects such as geometry from YOLO classification labels will be used to generate point clouds of the objects and include them in the bathymetric map. Based on the low computational complexity of the algorithm, another future work includes the real-time implementation of the algorithm onboard Bathy-drone such that the guidance system can automatically detect areas that need a closer look to complete the bathymetric map, and generate trajectories that the Bathy-drone will autonomously track.

Acknowledgment

The authors would like to thank the support from Aurigo Software Technologies, Inc.

References

  • [1] M. Lyons, S. Phinn, and C. Roelfsema, “Integrating quickbird multi-spectral satellite and field data: Mapping bathymetry, seagrass cover, seagrass species and change in moreton bay, australia in 2004 and 2007,” Remote Sensing, vol. 3, no. 1, pp. 42–64, 2011. [Online]. Available: https://www.mdpi.com/2072-4292/3/1/42
  • [2] B. Costa, T. Battista, and S. Pittman, “Comparative evaluation of airborne lidar and ship-based multibeam sonar bathymetry and intensity for mapping coral reef ecosystems,” Remote Sensing of Environment - REMOTE SENS ENVIRON, vol. 113, pp. 1082–1100, 05 2009.
  • [3] L. M. Wedding, A. M. Friedlander, M. McGranaghan, R. S. Yost, and M. E. Monaco, “Using bathymetric lidar to define nearshore benthic habitat complexity: Implications for management of reef fish assemblages in hawaii,” Remote Sensing of Environment, vol. 112, no. 11, pp. 4159–4165, 2008, applications of Remote Sensing to Monitoring Freshwater and Estuarine Systems. [Online]. Available: https://www.sciencedirect.com/science/article/pii/S0034425708002137
  • [4] H. Rice, S. Kelmenson, and L. Mendelsohn, “Geophysical navigation technologies and applications,” in PLANS 2004. Position Location and Navigation Symposium (IEEE Cat. No.04CH37556), 2004, pp. 618–624.
  • [5] L. LUCIDO, B. PESQUET-POPESCU, J. OPDERBECKE, V. RIGAUD, R. DERICHE, Z. ZHANG, P. COSTA, and P. LARZABAL, “Segmentation of bathymetric profiles and terrain matching for underwater vehicle navigation,” International Journal of Systems Science, vol. 29, no. 10, pp. 1157–1176, 1998. [Online]. Available: https://doi.org/10.1080/00207729808929605
  • [6] J. Cooper and F. Navas, “Natural bathymetric change as a control on century-scale shoreline behavior,” Geology, vol. 32, no. 6, pp. 513–516, 06 2004. [Online]. Available: https://doi.org/10.1130/G20377.1
  • [7] A.-C. Wölfl, H. Snaith, S. Amirebrahimi, C. W. Devey, B. Dorschel, V. Ferrini, V. A. I. Huvenne, M. Jakobsson, J. Jencks, G. Johnston, G. Lamarche, L. Mayer, D. Millar, T. H. Pedersen, K. Picard, A. Reitz, T. Schmitt, M. Visbeck, P. Weatherall, and R. Wigley, “Seafloor mapping – the challenge of a truly global ocean bathymetry,” Frontiers in Marine Science, vol. 6, 2019. [Online]. Available: https://www.frontiersin.org/articles/10.3389/fmars.2019.00283
  • [8] “Noaa bathymetric data viewer,” https://www.ncei.noaa.gov/maps/bathymetry/, accessed: 2022-08-30.
  • [9] A. L. Diaz, A. E. Ortega, H. Tingle, A. Pulido, O. Cordero, M. Nelson, N. E. Cocoves, J. Shin, R. R. Carthy, B. E. Wilkinson, and P. G. Ifju, “The Bathy-Drone: An Autonomous Unmanned Drone-Tethered Sonar System,” Drones, vol. 6, no. 8, p. 220, Aug. 2022.
  • [10] G. Zheng, H. Zhang, Y. Li, and J. Zhao, “A universal automatic bottom tracking method of side scan sonar data based on semantic segmentation,” Remote Sensing, vol. 13, no. 10, 2021. [Online]. Available: https://www.mdpi.com/2072-4292/13/10/1945
  • [11] Y. Yu, J. Zhao, Q. Gong, C. Huang, G. Zheng, and J. Ma, “Real-time underwater maritime object detection in side-scan sonar images based on transformer-yolov5,” Remote Sensing, vol. 13, no. 18, 2021. [Online]. Available: https://www.mdpi.com/2072-4292/13/18/3555
  • [12] A. J. Kaeser and T. L. Litts, “A novel technique for mapping habitat in navigable streams using low-cost side scan sonar,” Fisheries, vol. 35, no. 4, pp. 163–174, 2010. [Online]. Available: https://afspubs.onlinelibrary.wiley.com/doi/abs/10.1577/1548-8446-35.4.163
  • [13] A. Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, “Yolov4: Optimal speed and accuracy of object detection,” 2020.
  • [14] Q.-Y. Zhou, J. Park, and V. Koltun, “Open3D: A modern library for 3D data processing,” arXiv:1801.09847, 2018.
  • [15] D. Einsidler, M. Dhanak, and P.-P. Beaujean, “A Deep Learning Approach to Target Recognition in Side-Scan Sonar Imagery,” in OCEANS 2018 MTS/IEEE Charleston, Oct. 2018, pp. 1–4.
  • [16] D. Karimanzira, H. Renkewitz, D. Shea, and J. Albiez, “Object Detection in Sonar Images,” Electronics, vol. 9, p. 1180, Jul. 2020.
  • [17] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 779–788.
  • [18] J. Yu, Y. Jiang, Z. Wang, Z. Cao, and T. Huang, “UnitBox,” in Proceedings of the 24th ACM international conference on Multimedia.   ACM, oct 2016. [Online]. Available: https://doi.org/10.11452F2964284.2967274