Elsevier

Computers & Geosciences

Volume 95, October 2016, Pages 105-117
Computers & Geosciences

Research paper
Multi-view horizon-driven sea plane estimation for stereo wave imaging on moving vessels

https://doi.org/10.1016/j.cageo.2016.07.012Get rights and content

Highlights

  • We propose an optical horizon estimation technique to recover the attitude of a moving stereo rig with respect to the sea plane.

  • Projective distortion is corrected to account for the non-uniform distribution of the reconstructed point cloud.

  • The alignment of each sea surface is improved without using external positioning sensors.

Abstract

In the last few years we faced an increased popularity of stereo imaging as an effective tool to investigate wind sea waves at short and medium scales. Given the advances of computer vision techniques, the recovery of a scattered point-cloud from a sea surface area is nowadays a well consolidated technique producing excellent results both in terms of wave data resolution and accuracy. Nevertheless, almost all the subsequent analyses tasks, from the recovery of directional wave spectra to the estimation of significant wave height, are bound to two limiting conditions. First, wave data are required to be aligned to the mean sea plane. Second, a uniform distribution of 3D point samples is assumed. Since the stereo-camera rig is placed tilted with respect to the sea surface, perspective distortion do not allow these conditions to be met. Errors due to this problem are even more challenging if the optical instrumentation is mounted on a moving vessel, so that the mean sea plane cannot be simply obtained by averaging data from multiple subsequent frames. We address the first problem with two main contributions. First, we propose a novel horizon estimation technique to recover the attitude of a moving stereo rig with respect to the sea plane. Second, an effective weighting scheme is described to account for the non-uniform sampling of the scattered data in the estimation of the sea-plane distance. The interplay of the two allows us to provide a precise point cloud alignment without any external positioning sensor or rig viewpoint pre-calibration. The advantages of the proposed technique are evaluated throughout an experimental section spanning both synthetic and real-world scenarios.

Introduction

In the recent years there has been a growing interest in remote and proximal observation of sea surface waves. New classes of instruments (e.g. radars, laser scanners, optical) have been placed aboard satellites, airplanes, or ships at sea, facing the need for new processing tools for specific tasks. In this context, observations of sea surface fields by means of optical-based stereo systems, after the pioneering work of Schumacher (1939), Holthuijsen (1979), and Banner et al. (1989), are nowadays becoming accessible to a large number of scientists given the advances in the processing of digital images (Kosnik and Dulov, 2011, Gallego et al., 2011, Benetazzo et al., 2012, Benetazzo et al., 2015). The efforts made to develop both hardware and software for stereo processing has allowed stereo wave imaging to become a well consolidated and accurate tool for wave observations, that indeed is providing new insights into the field of sea surface waves (Banner et al., 2014, Leckler et al., 2015, Yurovskaya et al., 2013). So far, snapshots of the sea surface have been mostly taken from fixed platforms at sea, as this condition greatly eases the cameras deployment and the image processing. Indeed, the time-constant rigid-motion that places all the acquired surfaces into a common geo-referenced frame can be estimated a priori exploiting the statistical nature of the waves (Benetazzo et al., 2016). However, this design for stereo systems is starting to become too restrictive for oceanographers, who need to collect wave data in as many different conditions and locations as possible, eventually where the use of a fixed platform is unfeasible. Thus, first applications of stereo systems mounted on oceanographic vessels is taking place, facing the nontrivial task of accounting for the camera motion during the stereo processing (Brandt et al., 2010).

In this study we present an analysis of spatio-temporal 3D wave fields that take advantage of a multi-view horizon estimate approach for compensating ship motion. The paper is organized as follows. We start with an overview of the related work in the field of horizon estimation, highlighting the novelties of our proposed approach with respect to the literature. Then, in Section 2 we describe the stereo system used to reconstruct the sea surface. In Section 3 we present our multi-view horizon estimation approach to recover the sea surface plane normal. Then, in Section 4 we describe how to use the recovered normal to estimate a homography relating to the camera and sea plane. Such homography is used to define a spatial density estimator to weight the 3D point samples and account for perspective distortion. This allows a robust recovery of plane distance and, consequently, a precise alignment of the scattered point data. An extensive experimental part (Section 5) demonstrates the strengths of our method comparing with both synthetic and real-world data acquired in a controlled environment (i.e. a fixed oceanographic platform). Furthermore, a use-case is presented to show the behaviour of the approach “in the wild” while investigating the wave statistics on a moving ship during a cruise. Finally, in Section 6 we draw some concluding considerations.

Being able to recover the attitude and position of a moving vehicle subject to 6 degrees of freedom has been a topic of pivotal importance for both aerial and marine operations. With the former, the recent availability of inexpensive unmanned aerial vehicles to be used in dangerous or remote situations raised the need to improve the automatic control and stabilization beyond the accuracy obtainable with purely inertial sensors. In the latter, the precise estimation of ship motion is an aid for navigation safety (Liu et al., 2008), vessel detection (Fefilatyev et al., 2007) or environmental sciences (Williams and Howard, 2011).

In this plethora of different applications, vision based approaches have gained an increased popularity due to their potential ability to hinge the estimation on well-localizable visual features occurring on the landscape. Among all, the horizon line is usually the weapon of choice as is expected being always visible if no other assumptions can be made to the application scenario. Even if seeing the horizon is not a sufficient condition to recover the full 6 degrees of freedom of a vessel (the estimation is recovered up to any rotation around an arbitrary axis orthogonal to the ground plane), position can be usually obtained with enough accuracy through GPS sensors and is less critical with respect to attitude recovery. In fact, by constraining to a physically plausible kinematics of a ship, statistical filtering methods can improve the accuracy of common marine GPS to tens of centimeters. Given the importance of the horizon as a visual feature, sophisticated techniques have been proposed to provide a precise estimation even when the acquired images are affected by cluttering objects or optical artifacts (sun-glares, fog, low-contrast, etc.).

Following the taxonomy introduced by Shabayek et al. (2012), horizon line estimation methods are essentially divided into two categories. In the former, a Sky/Ground segmentation is performed by means of statistical models over the feature space being either color (Gallagher et al., 2004), textures (Cherian et al., 2009; Schwendeman and Thomson, 2015), polarization (Shabayek et al., 2012) or a combination of the previous (Todorovic et al., 2003). In the latter, the horizon line is directly estimated by extracting significant image edges (usually after a sequence of multiple filtering stages) and a statistically robust aggregation of contour segments composing the horizon or localizing vanishing points. For instance, Wang et al. (2009) propose an horizon extraction method from ocean observations by weighting the color channels to enhance sky and sea signals while reducing the effect of noise and highlights. Thereafter, contours are extracted with Canny edge detection and the predominant line clustered with a classical Hough transform. Similarly, Williams and Howard (2011) describe a rather sophisticated sequence of adaptive histogram thresholding, region growing and Gaussian mixture model to preprocess the images before binarization and contour extraction.

When horizon estimation is used to control an UAV, temporal coherency can be exploited to aid the extraction considering multiple video frames (Bao et al., 2003). Moreover, optical flow can be extracted between subsequent images as an aid to cluster horizon edges from spurious clutter arising from natural or human-made features on the landscape (Dusha et al., 2007). When an inertial measurement unit (IMU) is available, its data can be used as an initial prior to limit the image analysis on a reduced area of the image (Hugues et al., 2014). Furthermore, Angelino et al. (2013) suggest the usage of an Unscented Kalman Filter (UKF) to describe the dynamic system non-linear equations to effectively fuse attitude information coming from GPS, IMU and cameras.

If stereo vision is available, the complete pose can be obtained as a natural consequence of multi-view triangulation. For instance, Wang et al. (2005) use stereo vision to recover the full position and attitude of a UAV flying near the ground whereas Eynard et al. (2010) use an hybrid stereo system composed of a central and a fish-eye camera to measure the altitude without any external sensor. However, stereo reconstruction can only be used for pose estimation if we assume a non-moving set of landmarks being present in a scene. This is obviously not the case when observing the continuously changing sea surface.

In this paper we propose a novel horizon estimation technique to recover the stereo rig attitude with respect to the sea surface. Once the relative orientation of the rig and the sea plane is assessed (up to a rotation around the plane normal, that cannot be inferred by the horizon alone), its distance is estimated in a robust way that compensates the non-uniform distribution of all the reconstructed point samples.

With respect to the aforementioned existing approaches, our method introduces several novelties. To our knowledge, it presents the first intrinsically multi-view approach being able to fuse the horizon line estimates coming from different images. In our case, just the two images of the stereo rig are used but the method itself is flexible enough to comprise many different (pre-calibrated) image sources. Second, we use a prior given by the 3D reconstruction as a first guess to refine the estimated horizon line. In this sense, our method is more like an horizon refinement approach rather than a full estimator being able to locate the horizon anywhere in the image. Nevertheless, this prior allows us to pose the problem as an energy optimization spanning the entire image domain and avoiding voting schemes such as the commonly used Hough Transform. This both improves the accuracy of the method and reduces the computational effort. Third, our method is not based on a preliminary edge extraction step. Instead, it directly operates on image pixels in a way similar to active contours (Kass et al., 1988). This has the double benefit of getting rid of multiple edge extraction parameters and allowing a better line localization since no information is loss due to binarization.

The horizon line allows to recover the parameters defining the normal of the sea plane. This is of crucial importance for the alignment of the reconstructed data, that can be directly rotated so that the sea plane normal coincides with the z-axis. Also, we can exploit the relative angles to correct for perspective distortion in an optimal way to account for the non-uniform sampling of the reconstructed points with respect to the sea plane. This non-uniformity is typical of all the stereo methods that operate by matching corresponding pixels directly on the image space. These methods are the only feasible on a moving scenario (like the one investigated in this work) for which the pose of the cameras from the observed scene cannot be reliably provided a priori. On the contrary, when the pose can be properly calibrated, methods that directly operate on the object space like the variational approach in Gallego et al. (2008) can estimate a smooth spatially uniform triangulated surface at a cost of being computationally intensive for long sequences.

Section snippets

The WASS pipeline

Before proceeding with the description of the sea plane alignment technique, we give a brief description of the 3D reconstruction machinery and algorithms used in this work. Our Wave Acquisition Stereo System (WASS) is essentially composed of two digital cameras firmly placed side-by-side looking toward the sea surface. Each camera is equipped with a 5.0 mm low radial distortion lens and a CCD with a pixel size of 3.45 μm producing images with a resolution of 2048×2456 pixels. An electronic

Sea plane normal estimation

The first step to align the reconstructed stereo data is to estimate the normal n=(nxnynz)T of the sea plane with respect to the camera reference frame. We start by considering the case of a single camera oriented toward the surface so that the image plane is not parallel to (the trivial case in which they are parallel would not produce significant perspective distortion). In this context, the second camera of the rig is just used to perform 3D triangulation so that the reconstructed point

Plane distance recovery

So far we showed that plane normal n can be estimated without the reconstructed point cloud through the horizon line extracted from an image frame. From the normal n, a transformation Hr mapping points from the first camera image plane to an ideal camera whose optical axis is orthogonal to the sea plane can be recovered up to scale. This transformation is particularly interesting for two reasons: it allows the rectification of the acquired image to remove the perspective distortion (Fig. 4)

Experimental section

In this section we test the proposed approach under different conditions. We start with some synthetic cases to quantitatively evaluate the sea plane recovery accuracy in a simulated scenario. In Section 5.3 we test the horizon estimation in a controlled real-world environment during the acquisition of 3D data from a fixed platform. Then, in Section 5.4 we present a real-case study of the data acquired during a cruise over the Garda Lake, Italy.

Conclusions

In this paper we proposed a novel technique that exploits the observed horizon to provide a better alignment of the reconstructed 3D sea surface with respect to the mean sea plane. We avoid the combinatorial search of the horizon line over an image assuming that a 3D reconstruction of the surface have already been produced by the stereo vision machinery and refine the initial coarse estimate by means of a multi-view energy optimization involving all image pixels. Thus, this approach recovers

References (42)

  • A. Benetazzo

    Measurements of short water waves using stereo matched image sequences

    Coast. Eng.

    (2006)
  • A. Benetazzo et al.

    Stereo wave imaging from moving vesselspractical use and applications

    Coast. Eng.

    (2016)
  • Albarelli, A., Rodolà, E., Torsello, A., 2010. Robust camera calibration using inaccurate targets. In: Proceedings of...
  • Angelino, Baraniello, V., Cicala, L., 2013. High altitude UAV navigation using IMU, GPS and camera. In: 2013 16th...
  • M.L. Banner et al.

    Linking reduced breaking crest speeds to unsteady nonlinear water wave group behavior

    Phys. Rev. Lett.

    (2014)
  • M.L. Banner et al.

    Wavenumber spectra of short gravity waves

    J. Fluid Mech.

    (1989)
  • Bao, G., Zhou, Z., Xiong, S., Lin, X., Ye, X., 2003. Towards micro air vehicle flight autonomy research on the method...
  • A. Benetazzo et al.

    Observation of extreme sea waves in a spacetime ensemble

    J. Phys. Oceanogr.

    (2015)
  • Benetazzo, A., Bergamasco, F., Barbariol, F., Torsello, A., Carniel, S., Sclavo, M., 2014. Towards an operational...
  • Benetazzo, A., Fedele, F., Gallego, G., Shih, P.-C., Yezzi, A., 2012. Offshore stereo measurements of gravity waves...
  • A. Brandt et al.

    Three-dimensional imaging of the high sea-state wave field encompassing ship slamming events

    J. Atmos. Ocean. Technol.

    (2010)
  • Brodtkorb, P., Johannesson, P., Lindgren, G., Rychlik, I., Rydén, J., Sjö, E., 2000. WAFO—a Matlab toolbox for the...
  • J. Canny

    A computational approach to edge detection

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1986)
  • Cherian, A., Morellas, V., Papanikolopoulos, N., 2009. Accurate 3d ground plane estimation from a single image. In:...
  • R.O. Duda et al.

    Use of the Hough transformation to detect lines and curves in pictures

    Commun. ACM

    (1972)
  • Dusha, D., Boles, W.W., Walker, R.A., 2007. Attitude estimation for a fixed-wing aircraft using horizon detection and...
  • Eynard, D., Vasseur, P., Demonceaux, C., Fremont, V., 2010. UAV altitude estimation by mixed stereoscopic vision. In:...
  • Fefilatyev, S., Goldgof, D.B., Langebrake, L., 2007. Toward detection of marine vehicles on horizon from buoy camera,...
  • Gallagher, A., Luo, J., Hao, W., 2004. Improved blue sky detection using polynomial model fit. In: 2004 International...
  • Gallego, G., Benetazzo, A., Yezzi, A., Fedele, F., 2008. Wave statistics and spectra via a variational wave acquisition...
  • G. Gallego et al.

    A variational stereo method for the three-dimensional reconstruction of ocean waves

    IEEE Trans. Geosci. Remote Sens.

    (2011)
  • Cited by (0)

    View full text