Next Article in Journal
A Novel Central Camera Calibration Method Recording Point-to-Point Distortion for Vision-Based Human Activity Recognition
Next Article in Special Issue
Development of Fixed-Wing UAV 3D Coverage Paths for Urban Air Quality Profiling
Previous Article in Journal
A Perspective on Passive Human Sensing with Bluetooth
Previous Article in Special Issue
Heterogeneous Autonomous Robotic System in Viticulture and Mariculture: Vehicles Development and Systems Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Visual Servoing Scheme for Autonomous Aquaculture Net Pens Inspection Using ROV

by
Waseem Akram
1,*,
Alessandro Casavola
1,
Nadir Kapetanović
2,* and
Nikola Miškovic
2
1
Department of Informatics, Modeling, Electronics, and Systems (DIMES), University of Calabria, 87036 Rende, Italy
2
Laboratory for Underwater Systems and Technologies (LABUST), Faculty of Electrical Engineering and Computing, University of Zagreb, 10000 Zagreb, Croatia
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(9), 3525; https://doi.org/10.3390/s22093525
Submission received: 7 April 2022 / Revised: 27 April 2022 / Accepted: 29 April 2022 / Published: 5 May 2022
(This article belongs to the Collection Smart Robotics for Automation)

Abstract

:
Aquaculture net pens inspection and monitoring are important to ensure net stability and fish health in the fish farms. Remotely operated vehicles (ROVs) offer a low-cost and sophisticated solution for the regular inspection of the underwater fish net pens due to their ability of visual sensing and autonomy in a challenging and dynamic aquaculture environment. In this paper, we report the integration of an ROV with a visual servoing scheme for regular inspection and tracking of the net pens. We propose a vision-based positioning scheme that consists of an object detector, a pose generator, and a closed-loop controller. The system employs a modular approach that first utilizes two easily identifiable parallel ropes attached to the net for image processing through traditional computer vision methods. Second, the reference positions of the ROV relative to the net plane are extracted on the basis of a vision triangulation method. Third, a closed-loop control law is employed to instruct the vehicle to traverse from top to bottom along the net plane to inspect its status. The proposed vision-based scheme has been implemented and tested both through simulations and field experiments. The extensive experimental results have allowed the assessment of the performance of the scheme that resulted satisfactorily and can supplement the traditional aquaculture net pens inspection and tracking systems.

1. Introduction

Today, fish farming plays a key role in food production, and the number of fish farms is increasing rapidly [1]. Typically, fish farming is carried out in open sea net cages that are natural marine environments. These fish cages are prone to various environmental changes that include biofouling, the growth of organisms such as algae, mussels, hydroids and many more. Furthermore, the water movement also causes net deformation and increased stress on the mooring. These environmental changes may cause harm to the net status and fish health. For example, if netting damage happens, and if it not is discovered in time, the fish escape from the net, decreasing growth performance and food efficiency. Thus, to obtain sustainable fish farming, inspection and maintenance must be performed on a regular and efficient basis [2].
Traditionally, fish net pens inspection and maintenance are carried out by expert divers. However, this method poses high risks to human life and health because of strong oceans waves and currents in the marine environment. A recent trend in the literature is the use of remotely operator vehicles (ROVs) or autonomous underwater vehicles (AUVs) for underwater fish net pens inspection tasks. These vehicles offer low-size and low-cost effective solutions for the aforementioned tasks and can automate the operations using advanced information and communication technology, intelligence control and navigation systems. The use of sonar, compass and depth sensors allows real-time localization without the need for a geographic positioning system (GPS). Furthermore, the use of camera sensors and current computer vision methodologies provides real-time streaming of the environment and interpretation of the captured scenes [3].
In recent years, many researchers have shown increased interest in the development of methods and techniques for autonomous fish net pens inspection using ROVs/AUVs controlled by computer vision approaches. Some of these works are discussed in a later section of this article (see related work Section 2). From our review of the current state of the art, we have seen many research studies that have made use of computer vision technology for addressing the net inspection problem. Currently, these studies cover net damage detection methods including hole detection, biofouling detection, deformation detection, etc. These detection tasks have been performed through traditional computer vision techniques, and also some work has proposed the use of deep-learning methods. However, this area of research is still under development, and there have been considerable research efforts specifically devoted to the integration of control and detection techniques. The current solutions only focus on the detection part, while there are less or very few attempts toward the automatic control and navigation of the vehicle in the net inspection task.

Main Contribution

This paper demonstrates how a low-cost and camera-equipped ROV, integrated with a topside server on the surface, can be used for auto-inspection and relative positioning in the aquaculture environment. The main objective of this work is to develop a vision-based positioning and control scheme to cut the use of Ultra-Short Baseline (USBL) and to increase vehicle navigation autonomously. In this regard, we first reviewed in-depth the related work dealing with the autonomous vision-based net inspection problem. From our review, we noticed that most of the work only focused on damage detection and its relative position extraction. None of the related work deals with auto-positioning as well as navigation in a uniform and integrated structure.
In this paper, we present an underwater robotic application for automating the net inspection task in underwater fish farms. The strategy consists of an integrated and hybrid modular-based scheme, developed by using existing and well-known frameworks and tools, i.e., ROS, OpenCV, and Python, which is assessed in both virtual and real environments. To enable cost-effective fish net pens inspection, we propose the use of two-parallel ropes attached to the net plane. More specifically, we use traditional computer vision techniques, i.e., SURF (Speeded Up Robust Features) image features and Canny edge detector, for the detection of reference points in the camera images captured in run-time and point triangulation and PnP techniques for vehicle positions relative to the net. Both monocular and stereo imaging techniques were utilized to assess the robustness and correctness of the scheme. In addition, the ROV is directed by using a closed-loop control law to traverse the net from top to bottom along the net plane. As a consequence, we test our methods in simulation as well as in a real environment that illustrates the natural condition of the net installed on a farm. The corresponding results are presented and discussed.

2. Related Works

In this section, we review the currently available solutions for the fish net pens tracking and inspection problem. We start by looking in-depth at the available literature, their contributions and what they missed.
In [4], an autonomous net damage detection based on a curve features method was proposed. In this scheme, a bilateral filter, an interclass variance method and a gradient histogram were used in the prepossessing part. Next, the peak cure of the mesh was calculated and the position of the curve was determined for the tracking objective. The detection was performed by determining the characteristic of the net mesh in the image. The results were assessed both with simulations and real field experiments. However, the work only describes the image processing steps for the whole detection of the net, while the vehicle control and guidance aspects have not been addressed.
In [5], the authors proposed a real-time net structure monitoring system. The idea of integrating positioning sensors with a numerical analysis model is examined. The acoustic sensors were installed on the net. Then, because of different ocean currents and waves, the net position differences were calculated at different time steps. The scheme was used to determine the current velocity profiles in the aquaculture environment that cause the net deformation. However, the scheme needs to integrate communication sensors to provide the positions data to the end users.
In [6], the authors proposed a method for the pose, orientation and depth estimation of the vehicle relative to the net. The fast Fourier transform method was used to estimate the depth values from the camera images with known camera parameters. The scheme was tested both in virtual and real environments. However, also in this work, the vehicle control problem is not addressed.
Fish cage dysfunctionalities trigger system loss both from an economic and operational perspective. The bad infrastructure of the net allows fish to escape. To reduce the death rate of the fish, a periodic inspection is required. In this regard, authors of [7] discussed the design of a small-sized autonomous vehicle-based inspection system for underwater fish cages. The scheme offers net hole detection features while the vehicle navigates autonomously during the inspection. The depth estimation is carried out using the OpenCV triangulation method based on the target detection in the camera image. Based on the depth information, the vehicle is instructed to move forward/backward. The scheme was tested successfully in a real environment. However, to achieve more autonomy in the system, top-down movement control is required.
ROV/AUV-based aquaculture inspection poses localization issues as the GPS system does not work underwater. Alternatively, the surface vehicle area is easy to deploy and maintain with fewer limitations on communication and localization. In [8], authors discussed the design and implementation of an omnidirectional surface vehicle (OSV) for fish cage inspection tasks. A depth-adjustable camera was installed with the vehicle that captures the net structure at different depths. Furthermore, the net damage detection problem was solved by using a pretrained deep-learning-based method. However, the factors that interfere with the position estimation were not incorporated. In [9], authors presented an extension of a former work by incorporating the artificial-intelligence-based mission planning technique. A hierarchical task network was exploited to determine the rules for vehicle movement. However, the scheme is not validated in a realistic environment.
The traditional methods for biofouling removal are costly and have a great impact on fish net stability and fish health. The waste products are left in the water creating a bad environment for the fish. In this regard, ROV/AUV-based biofouling detection and removal provide a more sophisticated solution. In addition, static sensors are also used to regularly monitor environmental conditions. Authors of [10] reported a detailed theoretical analysis of robotic solutions for biofouling prevention and inspection in fish farms. Various technical and operational requirements are proposed and discussed. The study proposed an automatic robotic system for biofouling detection and cleaning that consists of environmental condition monitoring, net and biofouling inspection, growth prevention and fish monitoring inside the cages. As a result, that work proposed specifications and requirements for the development of such a system that offers detailed guidelines for the deployment of the robotic system for the aquaculture inspection task.
In [11], the authors proposed a novel method for net inspection problems. This work suggested the use of a Doppler velocity log (DVL) to approximate the vehicle’s relative position in front of the local region of the net. The position coordinates are then used in line-of-sight guidance control laws for the heading movement at a constant depth and angle with respect to the net plan. However, the scheme required noise handling in DVL to achieve better tracking results. Furthermore, due to an unfriendly environment, a more robust control law is required to deal with the model uncertainty.
In fish farming, water quality matters for the health of cultured fish. Thus, water quality assessment is also an issue of great interest in the fish farming environment. In [12], the authors presented a fish cage inspection system that integrates monitoring water quality along with the net status. Different sensors were installed on the net to monitor potential hydrogen (pH), oxidation reduction potential (ORP), dissolved oxygen (DO) and temperature. For net damage detection, a Hough Transform method was used to construct the net mesh, and based on the incomplete net pattern, the damaged part was detected in the camera image. Although the work was tested in an experimental environment, the vehicle was controlled manually. Similarly, the authors in [13] deployed hardware and software solutions including SeaModem for communication, HydroLab for water quality monitoring, and energy harvesting system through propellers in underwater fish farms via acoustic IoT networks.
Another regular inspection of the fish cage net is carried out in [14]. In this work, the distance control scheme is presented for net status inspection through video streaming in a real-time environment. This scheme requires a physical object attached to the net considered as the target location. Then, computer vision methods, e.g., canny edge detector, are used to detect the target in the image under fixed distance and angle. The target information is then used to instruct the vehicle to move forward/backward toward the net plan. Although the presented scheme is simple and easy to deploy, it requires that predetermined target objects be attached to the net surface. Additionally, the controller is not robust to environmental disturbances and noise.
Traditional positioning methods involving the use of a long-baseline and ultra-short baseline methods, require predeployed and localized infrastructure, increasing the cost and operational complexity. On the other hand, the laser and optical systems are easy to deploy and are efficient solutions in a dynamic environment. In this regard, the authors of [15] proposed a laser–camera triangulation-based autonomous inspection of the fish cage net. In this scheme, the idea is to project two parallel laser lines on the net plan. By using image processing techniques, the lines were extracted from the images and their positions were estimated by the triangulation method. This approach showed better results when compared to the DVL method. However, this work only suggested the position estimation for the net tracking problem and does not consider the underlying control problem. The laser triangulation method needs to be used in a closed-loop under a tracking controller suitably designed for tracking purposes.
A trend in recent years is to introduce artificial intelligence and Internet of Things technology in aquaculture systems to obtain real-time information and optimal aquaculture performance. In [16], an attempt has been made toward the application of an IoT-based smart fish net cage system. In this work, the authors developed a smart cage system that integrates artificial intelligence, IoT, a cloud system, big data analysis, sensors and communication technology. The system communicates field information to the cloud where the big data analysis is performed. The system generates real-time information related to fish health, survival rate and food residuals. However, this work only considered data collection and processing. Vehicle autonomous control and guidance problems are not considered.
Fish cages are floating structures, and it is difficult to get a planned image of the net through camera imaging. More robust image processing is required to deal with the different net structures, shapes and sizes. The different blurred scenes should also be considered. In this regard, the authors in [17] studied net hole detection of different shapes and sizes under different underwater conditions. In this work, a combination of Hough Transform and statistical analysis methods were used to perform a local and global search for detection problems. The work was only tested on the offline images sequence. However, the work needs to be tested on a real vehicle in real-time systems to verify its relevance.
Recent studies in [18,19,20] discussed the development and results of the HECTOR- heterogeneous autonomous robotic system in viticulture and mariculture project. The purpose of the project is to use an unmanned aerial vehicle (UAV), unmanned surface vehicle (USV) and ROV in an integrated and coordinated manner to carry out different missions such as vineyard surveillance, spraying and bud-rubbing in the viticulture domain and fish net monitoring in the mariculture domain. The research carried out as a part of the HECTOR project in [21] developed an autonomous control scheme that allows vehicles to navigate autonomously while performing the detection of net status. In this work, an ROV was allowed to move autonomously and stream video to the topside computer, perform image processing to detect two parallel ropes in an image considered as target positions and then generate the velocity commands to the vehicle for implementing a distance and top/down control. Additionally, a pretrained deep neural network was used to perform real-time biofouling detection on the net. However, the proposed scheme is not robust in the sunlight and produces blurred images in a real-time environment.
In the literature, pose estimation is mainly performed with feature-matching techniques. However, such approaches are prone to generate inconsistent results in the estimation because of the similarity in different regions of the net plan. To overcome this problem, the authors in [22] proposed a novel pose estimation method by considering junction detection. In this work, the knots of the net and their topology in the camera image were used in the pose estimation relative to the camera position. This approach cuts the computation burden of feature extraction from the image on the system performance. However, vehicle control and localization are not discussed. Moreover, the pose estimation is not robust in distorted images.
As we have seen, ROVs are mostly used for the autonomous fish cage inspection problem. However, ROVs feature low maneuverability and low efficiency in limited working space and a long and dynamic environment. To improve the performance of ROVs in inspection tasks, the authors of [23] developed a novel inspection scheme called Sea Farm inspector. The system integrates a ROV with a surface vehicle for the fish net inspection and tracking problem. The surface vehicle is responsible for controlling and communicating with the ROV during the operations. Furthermore, the design and control scheme is described in the work. However, this work is at the initial phase and the real implementation is ongoing. Further extensions have been undertaken in the work (17). In the latter study, the authors provide system coherence by properly integrating a surface vehicle, winch and ROV. However, the camera integration for net inspection is still being developed.
Next, a summary of the reviewed related work is shown in Table 1.

3. Design, Algorithm and Implementation

In this section, we present the design and implementation details of the proposed vision-based control scheme for the net inspection problem.

3.1. Overview

The proposed scheme follows a modular approach consisting of two modules. The first module is responsible for distance estimation using traditional computer vision techniques. The second module is responsible to guide and control the vehicle movement along the net plane to inspect its status. In this work, two different designs are investigated, namely Method 1 and Method 2, based on the camera use. There are a certain number of other applications that can profit from the proposed visual control strategies, see, e.g., [24,25,26].
In the following, a detailed description of the system is provided.

3.2. Distance Estimation

In this section, we describe distance detection by considering a reference point in the image frame of the net. We employed two different methods that are described hereafter:

3.2.1. Method 1: Stereo Image Design

In this section, we discuss the design of the proposed Method 1. The proposed idea is taken from [27] and further elaborated in this work. In the latter, a vision-based positioning system is presented for docking operation of the USVs in the harbor environment. We extend the solution of that previous work by allowing the online detection of the target positions in images, and the generated path is used to solve a fish net tracking problem for an ROV model.
In Figure 1, a schematic of Method 1 is shown. The forward-looking (left and right) cameras installed on the vehicle are used to collect the image sequences. The “cv-bridge” package is used to convert the obtained images from ROS to OpenCV formats. Next, the obtained images are forwarded to the object detector that extracts the net and draws a bounding box around the region of interest (ROI) using a canny edge detection algorithm. Next, from the right image, the image SURF features are extracted from the bounding box and searched in the corresponding bounding box of the left image along the horizontal lines. The matched points are used to compute the disparity map based on the triangulation method. Finally, the disparity map is used to obtain the relative positions of the vehicle with the objective to traverse the net from top to bottom while keeping a safe distance from it.
In this method, a stereo imaging design is employed. First, two raw images from the available cameras are collected as shown in Figure 2.
The obtained images may contain some noise and distortion. Then, we recover the rectified images by using the “unDistortRectify” OpenCV method. This requires the availability of the camera matrix and distortion coefficients obtained from the calibrated camera installed along with the simulator.
A = f x 0 c x 0 f y c y 0 0 1
where f x and f y are the focal lengths, and c x and c y are principal points. Moreover,
K = [ k 1 , k 2 , k 3 , p 1 , p 2 ]
are the distortion parameters, where k 1 , k 2 , k 3 denote the radial distortion parameters and p 1 , p 2 the tangential ones [15].
Next, a small region of interest (ROI) is selected in the image. Because the whole net has the same pattern, it is easy to select a small portion of the net to reduce the computational burden during the features extraction. The selected region of interest of size “300 by 200” pixel is shown in Figure 3.
Next, we can recover the edges in the ROI image by calling up the “Canny” edge detection algorithm. The Canny algorithm is a popular mathematical operator that finds prominent edges in images using a multi-stage approach consisting of noise reduction, finding intensities, non-maximum suppression and thresholding. This step is required to check if there are enough pixels related to the net area inside the selected ROI. Otherwise, the search and extraction of the features of interest may not be effective. The final result containing strong edges in the image is shown in Figure 4.
The next step is to design the stereo vision system. Here, the epipolar geometry concept is employed. The basic idea of the system is to search the SURF features in the bounding box containing the detected edges in the right image. SURF features are a scale and rotation-invariant interest point detector and descriptor [28]. Next, the same feature extraction steps are followed in the bounding box by considering horizontal scale lines corresponding to the left image. The next phase is to perform feature matching and find the best match in the corresponding left and right images. For points matching purposes, the K-Nearest Neighbour routine by OpenCV is used, and a filter is applied to detect only the best matches. Those are the ones that have a constant distance less than 0.6, and they are labeled as a best-matched point. Finally, the best match points are obtained and drawn as shown in Figure 5.
Algorithm 1 returns the pixel positions of best-matched points in the left and right images that are further sorted according to the distance value, and the points pair with minimum distance is selected which is used to calculate the pixel difference of these two points to compute the disparity. Then, the distance value is obtained with the help of the following formulas:
f = ( w / 2 ) / ( tan ( f o v / 2 ) ) d = | c x l c x r | d i s t a n c e = ( f b ) / ( d ) x = ( ( c x l c x ) b ) / ( d ) y = ( ( c y l c y ) b ) / ( d )
where f denotes the focal length, w denotes the width of the image frame, f o v the camera angle, d the disparity, c x l and c x r the pixel coordinates of the points in the left and right images, c x and c y the center midpoints of the image, and b the baseline, that is the distance between the two cameras.
Algorithm 1 Stereo vision-based positioning algorithm
Initialization:
1:
set: camera-left, camera-right
Online Phase
1:
for t > 0 do
2:
    get-images: Read images from both cameras
3:
    do-rectify: Remove distortion
4:
    get-roi: Select region of interest (ROI)
5:
    get-edges: Apply canny edge detector
6:
    draw-contours: Calculate a rectangular region containing the detected edges
7:
    find-features: Extract features present in contours
8:
    match-features: Match features pairs in the second image
9:
    filter-matched-features: Apply filter to get the best match feature pairs
10:
  return the pixel positions of the matched best feature pair
11:
end for

3.2.2. Method 2: Monocular Image Design

In this section, we discuss the design of the proposed Method 2 that is achieved by elaborating and extending ideas presented in [21]. Specifically, we have generalized the approach for the Blueye Pro ROV in a field environment and performed the assessment of the performance of the scheme. The basic idea of the scheme here is to identify two parallel ropes attached on the net in the image frame and then determine the position of the rope in the image to perform the distance estimation of the net with respect to the vehicle.
In Figure 6, a schematic of Method 2 is depicted. Both methods share the same functionality except for the usage of the cameras. In Method 2, a monocular camera is used. Here, the idea is having two parallel ropes along the net surface. Then, by means of edge detection and Hough transform algorithms, the ropes in the image are extracted, and pixel distance is calculated. Next, by using the computed pixel distance and knowing the real distance between the ropes, the positions are obtained which are the necessary input to the vehicle control and navigation algorithms.
The vehicle on-board camera is used to capture the cage net as shown in Figure 7. The input image is of size “1920 × 1080”—a high resolution image. To make the detection process easier and robust, it is necessary to undertake some preprocessing steps. Thus, first we modify the input image by applying the “CLAHE” OpenCV method. The CLAHE (contrast limited adaptive histogram equalization) recalculates the values of each pixel in the image and redistributes the brightness level, thereby increasing contrast in the image. This results in better visibility of the image objects and makes the identification easier. Next, the image is converted to gray and the “Bilateral filter” by OpenCV is used. The filter makes use of one or more Gaussian filters and blurs the neighborhood pixels of similar intensities while preserving the edges. The image dimension is reduced by a 25 % of the original one with the intention to eliminate the unnecessary details in the image. The resulting image is then used for the distance estimation process.
The next essential step is the identification of the two parallel ropes in the image. The ropes are recovered by applying the Canny edge detection algorithm. The algorithm follows a three step-procedure which include noise reduction, the calculation of the intensity gradient, the suppression of false edges and hysteresis thresholding. The resultinb image is shown in Figure 8.
The ropes in the image can essentially be considered as parallel straight lines. As we are only interested in these lines, we can freely discard the minor edges and only extract the large edges in the image. Therefore, the detection of the lines is achieved by applying the Hough transform method. This method requires an input image containing all edge information obtained from the previous steps and uses gradient information for the detection of the lines. The gradient is a measurement of the changing intensities of pixel coordinates inside an image and mathematically can be written as:
f ( x , y ) = G x G y = f / x f / y
where G x is the gradient of the x-axis, G y is the gradient of the y-axis, f / x shows change in intensity of the x-axis, and f / y shows change in intensity of the y-axis. Furthermore, the size and direction of the gradient are calculated by:
| f | = G x 2 + G x 2 θ = arctan G y / G x
Given the size ( | f | ) and direction ( θ ) of the gradient, the direction of an edge is determined by a perpendicular line at any given point in the image. Next, to draw the lines in the image, the polar coordinate system is used
r = x cos ( θ ) + y sin ( θ )
The pair ( r , θ ) shows the intersection point of a line that passes through two points ( x ) and ( y ) . Here, r denotes the distance from the origin to the nearest point on the line, and θ denotes the angle between the x-axis and the line which connects the origin with that nearest point. Thus, each line in the image is constructed, and the resulting image is shown in Figure 9.
The overall procedure of Method 2 is summarized in the following algorithm.
Algorithm 2 returns pixel positions of the detected two parallel lines in the image based on pixel differences and is calculated by:
d = | P L P R | d i s t a n c e = ( f o b j e c t / d ) s c a l e
where the term d denotes the difference between average pixel values detected in left and right ropes in the images that are denoted by PL and PR, respectively, and the o b j e c t denotes the real distance between the two ropes that need to be known in advance, and the term s c a l e takes into account the ROV camera tilt angle.
Algorithm 2 Monocular vision-based positioning algorithm
Initialization:
1:
set: camera
Online Phase
1:
for t > 0 do
2:
    get-images: Read image from camera
3:
    pre-process: Improve image quality
4:
    get-edges: Apply canny edge detector
5:
    draw-lines: Apply Hough-lines transform algorithm
6:
    separate-lines: Get the two parallel lines
7:
    return the pixel positions of the obtained parallel lines
8:
end for

3.3. Control Law

The ROV is described in 4 DOFs: surge, sway, heave and yaw. To solve the control design problem, first we assumed that:
  • The roll and pitch motion is passively stabilized by gravity and can therefore be neglected.
  • The vehicle is neutrally buoyant, and the motion in heave can therefore be neglected.
In particular, here we focus on the control of surge, sway, heave and yaw speed of the vehicle to perform the net pens inspection task. The control law is designed which directs the ROV heading toward the net pen and makes the ROV traverse the net pen with a desired distance and speed. This way, the camera view is directed toward the net such that the ROI stays in camera view while the ROV is traversing.
The control module is used to instruct the vehicle to perform a predetermined course outside the net to generate live streaming to the topside user and to inspect its status [21]. This part complements the methodology of the auto-inspection of nets using ROVs. The control commands use the position and distance information obtained from the object detection module via computer vision methods on the acquired images of the vehicle cameras. To this end, a simple but efficient control law is synthesized that generates the velocity set point for vehicle movement based on the reference position and distance data. The overall navigation problem can be stated as follows:
Given the depth information, generate the velocity commands to move the vehicle from top to bottom under a certain predefined distance while keeping the net plane in a parallel position with respect to the vehicle position.
In view of the above statement, the algorithm works as follows. First, the distance estimation module is called that preprocesses and detects target points in the images to identify the reference point. It checks if the target is visible or not. If the target is not visible, a rotation command is sent to the vehicle. Once the target is detected, it checks if the distance between the vehicle and the net is in the range of the predefined distances. If too far or too close, the forward/backward commands are sent to the vehicle, respectfully. Once the vehicle is at the desired distance, a top to bottom movement command is sent to the vehicle until the bottom area is detecte, and the navigation is stopped. The overall control procedures are explained with the help of Algorithm 3.
The goal of this study is the design of a control method allowing the ROV to traverse the net pens from top to bottom. Once the ROV reaches the bottom of the net pens, the distance estimation module is not receiving the input and sends a stop command to the vehicle. Once the one drive is completed, ROV is manually lifted to the docking/surface position. Incorporating the autonomous docking capability, in addition to the proper control of the heading and velocity, is out of the paper scope and will be addressed in a future study.
Algorithm 3 Control and Navigation algorithm
Initialization:
1:
set: target positions
2:
choose: ref-distance
3:
store: net-distance, wanted-distance, x-force, z-force
Online Phase
1:
for t > 0 do
2:
    compute: the net-distance by solving (3) or (7)
3:
    if net-distance > ref-distance then
4:
        move-fwd
5:
    else if net-distance < ref-distance then
6:
        move-bwd
7:
    else if net-distance==ref-distance then
8:
        move-dwn
9:
    else
10:
        wait
11:
    end if
12:
end for

4. Results

The proposed schemes used, both in simulations and experiments, a ROV and net pens with the same characteristics. First, the structure of the net pen was developed in the blender tool that supposedly roughly covered the size of the net pens used in the experiments. In addition, the methods only need to know in advance the distance between the reference points attached to the net pens. The ROV was used for image acquisition and tracking purposes, which is evident both from the undertaken simulations and experiments.
Following the design of the proposed vision-based control scheme, we move over to the results. The results are divided into two parts, i.e., Simulation and Experiments, where each focuses on the different design choices proposed in this work. Here, we discuss how the experiments were conducted and what we achieved. Finally, based on the obtained results, we make some conclusions.

4.1. Simulation Results

4.1.1. Description

To test the proposed Method 1 and Method 2 schemes in the simulation setting, we adopted the “unmanned underwater vehicle simulator” (UUV) simulator [29]. It is a set of packages that includes plugins and ROS applications that allow carrying out simulations of underwater vehicles in Gazebo and Rviz tools.
The simulation environment consisted of an underwater scene and the ROV vehicle that performed the tracking tasks. To simulate the net-tracking scenario, we designed the net structure using the blender tool. The simulation was performed on ROS Melodic distribution installed on Ubuntu 18.04.4 LTS.
In the simulator, the vehicle model named “rexrov-default” was used that consists of a mechanical based with two cameras and other sensing devices, e.g., inertial measurement unit (IMU) and LIDAR. Furthermore, the Doppler velocity log (DVL) was also available to measure the vehicle velocity during the simulation. The model also consisted of the implementation of Fossen’s motion equations for the vehicle [30] and the thruster management package.
By following the stereo and monocular imaging design, the two forward-looking cameras were used during images acquisition with the objective of testing the vision-based auto-inspection of the net structure. The simulation setup is shown in Figure 10 where the vehicle model is cloned in the underwater scene and facing toward the net plane.

4.1.2. Results

In this section, the simulation results are described. The main objective of the simulation was to test the performance of both Method 1 and Method 2 described in the earlier sections for distance estimation and tracking purposes. First, as an example, the working of the scheme is shown in Figure 11 where different ROS nodes are called in the terminal windows and communicate with the vehicle.
During the simulation, we observed the estimation of the distance between the vehicle camera and the net plane via Method 1 and Method 2. This is shown in Figure 12. From our results, we found that the monocular image-based method produced smooth results compared with the stereo-image-based method. The signal produced by Method 1 had multiple abrupt changes during the simulation. This was due to the fact that if the obtained images were blurred or not flat, the scheme extract features incorrectly caused the ambiguous estimation. From our experiments, we learned that Method 1 is highly influenced by the choice of hardware, feature extraction and computation cost. In contrast, Method 2 showed reasonable performance for the distance estimation. We can conclude that Method 2, which makes use of the monocular imaging design technique, is feasible for tracking underwater fish net pens at sea. The second test that we performed was finalized to observe the vehicle states during the simulation. This is shown in Figure 13, where the vehicle positions (x, y, z and yaw) achieved by applying Method 1 and Method 2 can be seen. In terms of performance, we noticed that Method 2 showed better performance compared to Method 1. This can be seen more clearly in the state x and angle y a w . With Method 1, the path covered was not linear despite the absence of external noise imposed in the control path. Furthermore, the state z confirmed the top/down traversing during the simulation. From our results, we can conclude that Method 2 is more suitable to be used for tracking. The third test was organized to analyze the velocity profiles during the simulation. This is shown in Figure 14 and Figure 15. Here, the results show the velocity set points for x and z. The x set point was used for the longitudinal movement while the z set poin twas used for the depth movement of the vehicle. The results confirm that whenever the vehicle stays at the desired distance from the net, the controller generates the z velocity set point correctly. From our analysis, we can conclude that the proposed schemes could be used to address the tracking problem.

4.2. Experimental Results

4.2.1. Description

To test the proposed Method 2 scheme in a real environment setting, a small set of experiments were conducted during the workshop “BTS (Breaking the Surface) 2022” organized by the Laboratory for Underwater Systems and Technologies (LABUST) at the Faculty of Electrical Engineering and Computing, University of Zagreb in October 2021 at Biograde de Muro, Croatia. The experiments were performed in the pool setup as shown in Figure 16.
Blueye Pro ROV as shown in Figure 17 was acquired from the LABUST during the experiments. The Blueye Pro ROV is produced by Blueye Robotics and has dimensions of 48.5 × 25.7 × 35.4 cm length, width and height, respectively. The ROV weighs 9 kg, it can go down to 300 m depth, and it has a 400 m cable for tethering. It is also equipped with a battery allowing 2 h of autonomy. The ROV can be used in saltwater, brackish, or freshwater. The drone moves with help of the available four thrusters of 350 w.
The ROV has a full HD forward-looking camera with a [−30 deg, 30 deg] tilt angle and 25–30 fps imaging capability. Additionally, it is equipped with a powerful light that ensures well-lit imagery under low-light scenarios. Other sensory devices including IMU, accelerometer, compass and temperature are also installed on it. The ROV allows control and communication with a topside user on the surface through both WiFi and Ethernet cable in a wireless environment [18]. Furthermore, the ROV is integrated with an open-source BlueyeSDK-ROS2 interface (see [31]). This allowed for achieving a connection between the Python-based ROS nodes running on a topside computer and the ROV.

4.2.2. Results

In this section, the experimental results are described. The main objective of the experiments was to test the performance of the proposed vision-based positioning and control scheme for the underwater fish net pens inspection problem using the Blueye Pro ROV. Here, we performed experiments by following the proposed Method 2 that makes use of the monocular imaging-based design technique.
Generally, the adopted ROV had several challenges during the experiments. First, there is no interface provided by the manufacturer where one can get access to the vehicle’s internal parameters which are necessary for the feedback control. Second, there is a lack of vehicle position, velocity and depth estimation. The only possible way to interact with the vehicle is to use the Blueye Python SDK that comes with the vehicle. SDK allowed us to subscribe to the “Thruster force” topic and publish a constant velocity set point to the surge, sway, heave and yaw motion of the vehicle as shown in Figure 18. Another main problem faced during the experiments was the uncalibrated vehicle camera that generated noisy images resulting in degradation of the algorithm performance.
Despite the above-mentioned challenges, the vehicle was used for the distance estimation of the vehicle relative to the fish net pens in a field trial as shown in Figure 16, and some results were collected. In Figure 19 the running interface is shown. In the figure, one can clearly see that the algorithm is working by identifying the ropes as two parallel lines in the input image. Next, these lines are used to estimate the distance to the fish net pens. The estimated distance is shown in Figure 20. The reference distance used during the experiments was 200 cm, and based on the current distance, the vehicle was instructed to move forward/backward. While the vehicle got in range of the wanted distance, the top to down motion was called. Here, we performed two different experiments and examined the estimated distance. The results show similar performance from both experiments. However, false estimation was also observed during the experiments. By clearly examining the results and after tuning the algorithm parameters, we concluded that the estimation process is influenced by the input data. From the experiments, we learned that the algorithm was showing performance degradation with sunlight and poor weather. Moreover, the distance estimation data showed that the proposed scheme was capableof being used for the distance estimation and tracking of the fish cage net. For thoroughness, the thruster force profiles during Experiments 1 and Experiments 2 are shown in Figure 21 and Figure 22, respectively. Here, the results confirmed that the control part successfully generated the velocity set point x and z whenever necessary on the basis of the reference distance.

4.3. Comment

This work aimed at the design and development of an autonomous fish net pens tracking system. The obtained results are promising and indicate the capability of the schemes for efficiently detecting, localizing, and tracking the fish net pens in the real environment. However, the robustness of the proposed methods has to be routinely tested with a sequence of sea trials over time.

5. Conclusions

In this paper, a vision-based positioning and control scheme is described that can be used for auto-inspection of the fish net pens in underwater fish farms. We employed both stereo and monocular image design approaches for the input data acquisition and the traditional computer vision methods for the target position detection. The vision algorithm was integrated with a control module that allows the vehicle to perform the traversing along the net plane. The system was tested both in a simulation and in a real environment. In terms of performance, we found that the monocular image-based method is more suitable than the stereo-image-based method. From the obtained results, we learned that the stereo-image-based method is highly influenced by the choice of hardware design, features extraction and computation cost. In contrast, the monocular image-based method is found to be easily adopted in real applications because of fewer requirements and lower computation costs. The scheme also avoids tracking the image features and does not suffer from repeated scenes in the input data. In the future, we are interested to overcome the existing limitations by performing more experiments, providing the state’s data, and evaluating the results with the true positions and state data. We also plan to modify the control part and integrate it with the feedback control law to make the scheme more automatic.

Author Contributions

Conceptualization, N.K. and N.M.; Data curation, A.C.; formal analysis, W.A. and N.K.; funding acquisition, N.M.; investigation, N.M.; methodology, W.A. and N.K.; project administration, N.M.; resources, N.K. and N.M.; software, W.A. and N.K.; supervision, A.C. and N.M.; validation, W.A., A.C. and N.K.; visualization, W.A.; writing—original draft, W.A. and A.C.; writing—review and editing, W.A., A.C., N.K. and N.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the European Regional Development Fund, Operational Programme Competitiveness and Cohesion 2014–2020, through project Heterogeneous Autonomous Robotic System in Viticulture and Mariculture (HEKTOR)—grant number KK.01.1.1.04.0036; and the European Regional Development Fund through the Interreg Italy–Croatia InnovaMare project (Partnership ID 10248782).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Matej Fabijanić from LABUST for providing contributions to algorithm development and sea trials. We would also like to thank Đula Nađ, Fausto Ferreira, Igor Kvasić, Nikica Kokir, Martin Oreč, Vladimir Slošić, and Kristijan Krčmar for providing useful discussions and assistance during sea trials. Our thanks also go to Marco Lupia from DIMES for providing help on ROS.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jovanović, V.; Svendsen, E.; Risojević, V.; Babić, Z. Splash detection in fish Plants surveillance videos using deep learning. In Proceedings of the 2018 14th Symposium on Neural Networks and Applications (NEUREL), Belgrade, Serbia, 20–21 November 2018; pp. 1–5. [Google Scholar]
  2. Liao, W.; Zhang, S.; Wu, Y.; An, D.; Wei, Y. Research on intelligent damage detection of far-sea cage based on machine vision and deep learning. Aquac. Eng. 2022, 96, 102219. [Google Scholar] [CrossRef]
  3. Ubina, N.A.; Cheng, S.C. A Review of Unmanned System Technologies with Its Application to Aquaculture Farm Monitoring and Management. Drones 2022, 6, 12. [Google Scholar] [CrossRef]
  4. Zhao, Y.P.; Niu, L.J.; Du, H.; Bi, C.W. An adaptive method of damage detection for fishing nets based on image processing technology. Aquac. Eng. 2020, 90, 102071. [Google Scholar] [CrossRef]
  5. Su, B.; Kelasidi, E.; Frank, K.; Haugen, J.; Føre, M.; Pedersen, M.O. An integrated approach for monitoring structural deformation of aquaculture net cages. Ocean. Eng. 2021, 219, 108424. [Google Scholar] [CrossRef]
  6. Schellewald, C.; Stahl, A.; Kelasidi, E. Vision-based pose estimation for autonomous operations in aquacultural fish farms. IFAC-PapersOnLine 2021, 54, 438–443. [Google Scholar] [CrossRef]
  7. Chalkiadakis, V.; Papandroulakis, N.; Livanos, G.; Moirogiorgou, K.; Giakos, G.; Zervakis, M. Designing a small-sized autonomous underwater vehicle architecture for regular periodic fish-cage net inspection. In Proceedings of the 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 18–20 October 2017; pp. 1–6. [Google Scholar]
  8. Tao, Q.; Huang, K.; Qin, C.; Guo, B.; Lam, R.; Zhang, F. Omnidirectional surface vehicle for fish cage inspection. In Proceedings of the OCEANS 2018 MTS/IEEE Charleston, Charleston, SC, USA, 22–25 October 2018; pp. 1–6. [Google Scholar]
  9. Lin, T.X.; Tao, Q.; Zhang, F. Planning for Fish Net Inspection with an Autonomous OSV. In Proceedings of the 2020 International Conference on System Science and Engineering (ICSSE), Kagawa, Japan, 31 August–3 September 2020; pp. 1–5. [Google Scholar]
  10. Ohrem, S.J.; Kelasidi, E.; Bloecher, N. Analysis of a novel autonomous underwater robot for biofouling prevention and inspection in fish farms. In Proceedings of the 2020 28th Mediterranean Conference on Control and Automation (MED), Saint-Raphaël, France, 15–18 September 2020; pp. 1002–1008. [Google Scholar]
  11. Amundsen, H.B.; Caharija, W.; Pettersen, K.Y. Autonomous ROV inspections of aquaculture net pens using DVL. IEEE J. Ocean. Eng. 2021, 47, 1–19. [Google Scholar] [CrossRef]
  12. Betancourt, J.; Coral, W.; Colorado, J. An integrated ROV solution for underwater net-cage inspection in fish farms using computer vision. SN Appl. Sci. 2020, 2, 1–15. [Google Scholar] [CrossRef]
  13. Cario, G.; Casavola, A.; Gjanci, P.; Lupia, M.; Petrioli, C.; Spaccini, D. Long lasting underwater wireless sensors network for water quality monitoring in fish farms. In Proceedings of the OCEANS 2017-Aberdeen, Aberdeen, UK, 19–22 June 2017; pp. 1–6. [Google Scholar]
  14. Livanos, G.; Zervakis, M.; Chalkiadakis, V.; Moirogiorgou, K.; Giakos, G.; Papandroulakis, N. Intelligent navigation and control of a prototype Autonomous underwater vehicle for automated inspection of aquaculture net pen cages. In Proceedings of the 2018 IEEE International Conference on Imaging Systems and Techniques (IST), Krakow, Poland, 16–18 October 2018; pp. 1–6. [Google Scholar]
  15. Bjerkeng, M.; Kirkhus, T.; Caharija, W.; T Thielemann, J.; B Amundsen, H.; Johan Ohrem, S.; Ingar Grøtli, E. ROV Navigation in a Fish Cage with Laser-Camera Triangulation. J. Mar. Sci. Eng. 2021, 9, 79. [Google Scholar] [CrossRef]
  16. Chang, C.C.; Wang, J.H.; Wu, J.L.; Hsieh, Y.Z.; Wu, T.D.; Cheng, S.C.; Chang, C.C.; Juang, J.G.; Liou, C.H.; Hsu, T.H.; et al. Applying Artificial Intelligence (AI) Techniques to Implement a Practical Smart Cage Aquaculture Management System. J. Med. Biol. Eng. 2021, 41, 652–658. [Google Scholar] [CrossRef]
  17. Paspalakis, S.; Moirogiorgou, K.; Papandroulakis, N.; Giakos, G.; Zervakis, M. Automated fish cage net inspection using image processing techniques. IET Image Process. 2020, 14, 2028–2034. [Google Scholar] [CrossRef]
  18. Kapetanović, N.; Nad, D.; Mišković, N. Towards a Heterogeneous Robotic System for Autonomous Inspection in Mariculture. In Proceedings of the OCEANS 2021 Conference and Exposition, San Diego—Porto (Hybrid), San Diego, CA, USA, 20–23 September 2021; pp. 1–6. [Google Scholar]
  19. Rezo, M.; Čagalj, K.M.; Kovačić, Z. Collecting information for biomass estimation in mariculture with a heterogeneous robotic system. In Proceedings of the 44th International ICT Convention MIPRO, Opatija, Croatia, 27 September–1 October 2021; pp. 1295–1300. [Google Scholar]
  20. Goričanec, J.; Kapetanović, N.; Vatavuk, I.; Hrabar, I.; Kurtela, A.; Anić, M.; Vasilijević, G.; Bolotin, J.; Kožul, V.; Stuhne, D.; et al. Heterogeneous autonomous robotic system in viticulture and mariculture-project overview. In Proceedings of the 16th International Conference on Telecommunications-ConTEL, Zagreb, Croatia, 30 June–2 July 2021; pp. 1–8. [Google Scholar]
  21. Borković, G.; Fabijanić, M.; Magdalenić, M.; Malobabić, A.; Vuković, J.; Zieliński, I.; Kapetanović, N.; Kvasić, I.; Babić, A.; Mišković, N. Underwater ROV Software for Fish Cage Inspection. In Proceedings of the 2021 44th International Convention on Information, Communication and Electronic Technology (MIPRO), Opatija, Croatia, 27 September–1 October 2021; pp. 1747–1752. [Google Scholar]
  22. Duda, A.; Schwendner, J.; Stahl, A.; Rundtop, P. Visual pose estimation for autonomous inspection of fish pens. In Proceedings of the OCEANS 2015-Genova, Genova, Italy, 18–21 May 2015; pp. 1–6. [Google Scholar]
  23. Osen, O.L.; Leinan, P.M.; Blom, M.; Bakken, C.; Heggen, M.; Zhang, H. A novel sea farm inspection platform for norwegian aquaculture application. In Proceedings of the OCEANS 2018 MTS/IEEE Charleston, Charleston, SC, USA, 22–25 October 2018; pp. 1–8. [Google Scholar]
  24. DeCarlo, R.A.; Zak, S.H.; Matthews, G.P. Variable structure control of nonlinear multivariable systems: A tutorial. Proc. IEEE 1988, 76, 212–232. [Google Scholar] [CrossRef] [Green Version]
  25. Conte, G.; Scaradozzi, D.; Mannocchi, D.; Raspa, P.; Panebianco, L.; Screpanti, L. Development and experimental tests of a ROS multi-agent structure for autonomous surface vehicles. J. Intell. Robot. Syst. 2018, 92, 705–718. [Google Scholar] [CrossRef]
  26. Djapic, V.; Nad, D. Using collaborative autonomous vehicles in mine countermeasures. In Proceedings of the OCEANS’10 IEEE SYDNEY, Sydney, NSW, Australia, 24–27 May 2010; pp. 1–7. [Google Scholar]
  27. Volden, Ø.; Stahl, A.; Fossen, T.I. Vision-based positioning system for auto-docking of unmanned surface vehicles (USVs). Int. J. Intell. Robot. Appl. 2021, 6, 86–103. [Google Scholar]
  28. Ferreira, F.; Veruggio, G.; Caccia, M.; Bruzzone, G. Real-time optical SLAM-based mosaicking for unmanned underwater vehicles. Intell. Serv. Robot. 2012, 5, 55–71. [Google Scholar] [CrossRef]
  29. Manhães, M.M.M.; Scherer, S.A.; Voss, M.; Douat, L.R.; Rauschenbach, T. UUV Simulator: A Gazebo-based package for underwater intervention and multi-robot simulation. In Proceedings of the OCEANS 2016 MTS/IEEE Monterey, Monterey, CA, USA, 19–23 September 2016. [Google Scholar] [CrossRef]
  30. Fossen, T.I. Handbook of Marine Craft Hydrodynamics and Motion Control; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  31. Kapetanović, N.; Vuković, J. Blueye SDK-ROS2 Interface. 2021. Available online: https://github.com/labust/blueye-ros2-pkg.git (accessed on 22 July 2021).
Figure 1. Method 1. A stereo image design scheme for positioning and auto-inspection of the underwater fish cage net adopted from [27].
Figure 1. Method 1. A stereo image design scheme for positioning and auto-inspection of the underwater fish cage net adopted from [27].
Sensors 22 03525 g001
Figure 2. Input images from the vehicle cameras: (a) left camera’s image; (b) right camera’s image.
Figure 2. Input images from the vehicle cameras: (a) left camera’s image; (b) right camera’s image.
Sensors 22 03525 g002
Figure 3. This figure illustrates how ROI is extracted. This region is used for edge detection and features extraction.
Figure 3. This figure illustrates how ROI is extracted. This region is used for edge detection and features extraction.
Sensors 22 03525 g003
Figure 4. Output images after applying Canny edge detector: (a) detected edges in ROI; (b) resulted image.
Figure 4. Output images after applying Canny edge detector: (a) detected edges in ROI; (b) resulted image.
Sensors 22 03525 g004
Figure 5. Matched points in both right and left image.
Figure 5. Matched points in both right and left image.
Sensors 22 03525 g005
Figure 6. Method 2. Monocular image design scheme for positioning and auto-inspection of the underwater fish cage net adopted from [21].
Figure 6. Method 2. Monocular image design scheme for positioning and auto-inspection of the underwater fish cage net adopted from [21].
Sensors 22 03525 g006
Figure 7. An example of the input image from the blueye vehicle camera obtained during the experiments.
Figure 7. An example of the input image from the blueye vehicle camera obtained during the experiments.
Sensors 22 03525 g007
Figure 8. Results after preprocessing the original image and applying the Canny edge detector.
Figure 8. Results after preprocessing the original image and applying the Canny edge detector.
Sensors 22 03525 g008
Figure 9. Output image with two identified parallel lines after applying OpenCv methods.
Figure 9. Output image with two identified parallel lines after applying OpenCv methods.
Sensors 22 03525 g009
Figure 10. Simulation setup. Model initialized in Gazebo simulator.
Figure 10. Simulation setup. Model initialized in Gazebo simulator.
Sensors 22 03525 g010
Figure 11. Simulation screen captured during run-time.
Figure 11. Simulation screen captured during run-time.
Sensors 22 03525 g011
Figure 12. Fish cage net distance estimation during simulation.
Figure 12. Fish cage net distance estimation during simulation.
Sensors 22 03525 g012
Figure 13. Vehicle positions during simulation.
Figure 13. Vehicle positions during simulation.
Sensors 22 03525 g013
Figure 14. Velocity profile by Method 1.
Figure 14. Velocity profile by Method 1.
Sensors 22 03525 g014
Figure 15. Velocity profile by Method 2.
Figure 15. Velocity profile by Method 2.
Sensors 22 03525 g015
Figure 16. Field trial view before experiments.
Figure 16. Field trial view before experiments.
Sensors 22 03525 g016
Figure 17. The Blueye Pro ROV.
Figure 17. The Blueye Pro ROV.
Sensors 22 03525 g017
Figure 18. Screenshot of rqt-graph of all the ROS nodes during run-time.
Figure 18. Screenshot of rqt-graph of all the ROS nodes during run-time.
Sensors 22 03525 g018
Figure 19. Experiments screen captured during run-time.
Figure 19. Experiments screen captured during run-time.
Sensors 22 03525 g019
Figure 20. Fish cage net distance estimation.
Figure 20. Fish cage net distance estimation.
Sensors 22 03525 g020
Figure 21. Velocity profile during Experiment 1.
Figure 21. Velocity profile during Experiment 1.
Sensors 22 03525 g021
Figure 22. Velocity profile during Experiment 1.
Figure 22. Velocity profile during Experiment 1.
Sensors 22 03525 g022
Table 1. Review of fish net tracking and inspection techniques.
Table 1. Review of fish net tracking and inspection techniques.
Ref.TechniqueTaskRemarks
[4]Bilateral filterDamage detectionDoes not incorporate the vehicle control
[5]Kalman filterStructure detectionPositions data is not communicated
[6]Fourier TransformPose estimationDoes not incorporate the vehicle control
[7]Canny edge detectorHole detectionDoes not perform top-down tracking
[8]Deep learningDamage detectionExperience pose estimation error
[11]DVLNet inspectionNot robust to noise
[12]Hough transformDamage detection and water quality monitoringThe vehicle is controlled manually
[13]IoT networkWater quality monitoringDoes not consider net tracking
[14]Canny edge detectorNet status inspectionRequired predetermined target location on the net plane
[15]Canny edge detectorNet status inspectionDoes not consider vehicle control
[17]Hough transformHole detectionWorked on offline images without control system
[21]Canny edge detectorNet status inspectionNot robust to sunlight and blurred images
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Akram, W.; Casavola, A.; Kapetanović, N.; Miškovic, N. A Visual Servoing Scheme for Autonomous Aquaculture Net Pens Inspection Using ROV. Sensors 2022, 22, 3525. https://doi.org/10.3390/s22093525

AMA Style

Akram W, Casavola A, Kapetanović N, Miškovic N. A Visual Servoing Scheme for Autonomous Aquaculture Net Pens Inspection Using ROV. Sensors. 2022; 22(9):3525. https://doi.org/10.3390/s22093525

Chicago/Turabian Style

Akram, Waseem, Alessandro Casavola, Nadir Kapetanović, and Nikola Miškovic. 2022. "A Visual Servoing Scheme for Autonomous Aquaculture Net Pens Inspection Using ROV" Sensors 22, no. 9: 3525. https://doi.org/10.3390/s22093525

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop