Next Article in Journal
Estimation of Eye Closure Degree Using EEG Sensors and Its Application in Driver Drowsiness Detection
Next Article in Special Issue
A Sensorized Nuss Bar for Patient-Specific Treatment of Pectus Excavatum
Previous Article in Journal
Active Disaster Response System for a Smart Building
Previous Article in Special Issue
Performance of SOI Bragg Grating Ring Resonator for Nonlinear Sensing Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Calibration of Action Cameras for Photogrammetric Purposes

by
Caterina Balletti
1,†,
Francesco Guerra
1,†,
Vassilios Tsioukas
2,*,† and
Paolo Vernier
1,†
1
IUAV di Venezia-S. Croce 191, 30135 Venezia, Italy
2
Aristotle University of Thessaloniki, Univ. Box. 432, Thessaloniki 54124, Greece
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2014, 14(9), 17471-17490; https://doi.org/10.3390/s140917471
Submission received: 28 July 2014 / Revised: 28 August 2014 / Accepted: 5 September 2014 / Published: 18 September 2014
(This article belongs to the Special Issue State-of-the-Art Sensors Technology in Italy 2014)

Abstract

: The use of action cameras for photogrammetry purposes is not widespread due to the fact that until recently the images provided by the sensors, using either still or video capture mode, were not big enough to perform and provide the appropriate analysis with the necessary photogrammetric accuracy. However, several manufacturers have recently produced and released new lightweight devices which are: (a) easy to handle, (b) capable of performing under extreme conditions and more importantly (c) able to provide both still images and video sequences of high resolution. In order to be able to use the sensor of action cameras we must apply a careful and reliable self-calibration prior to the use of any photogrammetric procedure, a relatively difficult scenario because of the short focal length of the camera and its wide angle lens that is used to obtain the maximum possible resolution of images. Special software, using functions of the OpenCV library, has been created to perform both the calibration and the production of undistorted scenes for each one of the still and video image capturing mode of a novel action camera, the GoPro Hero 3 camera that can provide still images up to 12 Mp and video up 8 Mp resolution.

1. Introduction

During the last few years, passive sensors providing visual image information [1], as well as the development of software solutions for the extraction of point clouds from a set of un-oriented images have received increasing attention, not only from the surveying community, but also from non-experts (such as archeologists). The reason is due to some clear advantages provided from these systems: the improved resolution and sensitivity of photo sensors and the availability of low cost off-the-shelf digital cameras which can be used in several photogrammetric applications, such as aerial mapping (e.g., aimed to structural and archaeological surveying) or emergency documentation.

Normally in photogrammetric applications, amateur cameras, semi-professional and professional cameras (dSLR) are not appropriate for measurement purposes. Making it necessary to calibrate them by estimating their intrinsic parameters in order to be able to extract accurate 3D metric information from their images [2]. The estimation of these geometric characteristics, i.e., the focal length (f) of the lens, the coordinates of the centre of projection of the image (xp, yp), the radial lens distortion coefficients (k1, k2, k3) [2], is performed through the camera calibration process.

A full review of camera calibration methods and models developed in the last few years can be found in [35] describing the commonly adopted methods such as Tsai [6] or Zhang [7]. These are all based on the pinhole camera model including terms for modeling radial distortion. Nowadays there are several software applications (i.e., Photomodeler, Agisoft Lens, iWitness, MicMac, 3DF Zephir, etc.), mainly produced by the computer vision scientific community, that can automatically perform camera self-calibration. They also offer the possibility to work with several cameras and sensors to obtain dense point clouds or 3D models suitable for different fields of applicFation.

In the wide panorama of low-priced consumer grade devices [8] (including nowadays even smartphones or other similar mobile devices) [9,10], action cameras are more widespread and have been used more and more during extreme activities such as free fall, parapent flying, underwater swimming and diving. Their photogrammetric use though has not been easy since they could not supply high resolution still images and video and additionally their geometry are far away from the theoretical model of central projection due to their wide angle or fish eye lenses.

Several approaches have recently been testing the use of the GoPro Hero 3 camera (GoPro, Inc., San Mateo, CA, USA) [1113], particularly for the recording and/or photogrammetric use in underwater conditions even using a stereoscopic configuration to obtain better results or to produce models from immersive videos. However, this was not possible since the baseline of the cameras was very short [1416].

The camera's characteristics are ideal not only for underwater conditions, but also in UAV configurations. The camera is very light and additionally it can be controlled through a WiFi connection from a cell phone or tablet [17].

2. Experimental Section

2.1. Camera Specifications

The weight (Table 1) and dimensions (Table 2) of the camera are ideal for photogrammetric applications, especially when it is necessary to place it on an UAV or on top of a diver's mask.

The GoPro Hero 3 camera has a rather big sensor which is able to provide four different still image resolutions (Table 3) and nine different video resolutions (Table 4).

The modes of the camera used to capture still images or video sequences vary depending on the needs of the application and the user demands. The frame rate of the coarser video resolution (WVGA 848 × 480) is about 240 FPS, while the finer resolution (4K Cin 4096 × 2160) can provide a minimum of 12 FPS (Table 4).

Additionally, the image dimensions ratio also varies and can be 4:3, 16:9 or 17:9. Therefore, the sensor part used to acquire the still images or the video sequences is not always the same. For this reason separate self-calibration procedures have to be applied in order to calculate the camera calibration parameters. Using our software, we have calibrated the GoPro camera in several of its resolutions, however for photogrammetric purposes the most useful is the largest one, hence for the accuracy assessment of our calibration procedure (see Section 3) we have used only 12 Mpixels GoPro images.

2.2. Calibration Model and Undistorted Images

A modified mathematical of the Brown Calibration model [18] has been used to determine all the appropriate parameters that determine the lens deformation appearing on an image [19]. The specific model uses odd and even order polynomial coefficients to model the radial and the tangential distortion of the lenses (Equations (14)):

d x radial = x ( 1 + k 1 r 2 + k 3 r 4 + k 5 r 5 )
d y radial = y ( 1 + k 1 r 2 + k 3 r 4 + k 5 r 5 )
d x tan = x + [ 2 p 1 x y + p 2 ( r 2 + 2 x 2 ) ]
d y tan = y + [ p 1 ( r 2 + 2 y 2 ) + 2 p 2 x y ]

The developed software, containing all the procedures needed to calibrate not just the GoPro camera, but any imaging sensor that is able to create digital photographic images, is compiled in Microsoft Visual Studio 2010, using the functions of the graphics library OpenCV (version 2.4.3). Within the OpenCV library there are also functions that implement the modified Brown Calibration Model. With the use of certain image analysis and adjustment functions of OpenCV it was feasible to create a software application running in Windows OS that is able to:

  • Identify a big number of ground control points (targets) on images printed on a special calibration sheet (chessboard or circle patterns);

  • Calculate through adjustment all the camera calibration parameters; except for the radial and tangential calibration parameters the camera constant (f) value and the principal point's coordinates (xo, yo) are determined;

  • And (optionally) create the undistorted (idealized) images (Figure 1).

The estimated accuracy of the calibration is given by the OpenCV routine that simultaneously calculates the extrinsic and intrinsic parameters of all the images used in the calibration process. The value is given in pixels and is referred to as Average Reprojection Error. It concerns the difference of target points location defined in the original images from their estimated location using the calculated exterior, interior and calibration parameters.

Examining the displacement of a point on the original (Figure 2a) and the idealized image (Figure 2b) we can identify the radial and tangential deformation. The radial deformation (cyan line) is collinear to the radius of the correct location of the point from the image center and the tangential (blue) line is vertical to the radius. Its tangential vector is relatively small to the total deformation error (Figure 3).

2.3. Recognized Problems and Finding Solutions

Although there exists an OpenCV sample software application (running under the Windows command line environment) able to deliver camera calibration, it cannot be used to calibrate high resolution digital cameras (more than 20 Mpixels). The specific software was originally created for the calibration of small frame cameras and especially for web cameras in Computer Vision Applications.

Certain problems appeared when we tried to identify control points of the printed calibration sheet on high resolution images. Especially, when we tried to calibrate a full frame camera (Nikon D800 (NIKON CORPORATION, Tokyo, Japan)) of a 36.3 Mpixels resolution the software was unable to identify the points of a chessboard calibration field. The effect of defocus in the part of the image that was away from the center of the image (or the point in focus) introduced the biggest problem (Figure 4).

Therefore we decided to use a regular grid of circles as calibration field. In the case of the circular grid, the barycentric algorithm that we used to identify the exact location of the circles' centers, was not affected by the defocusing effect. Another issue appearing in some cases is the bad lighting of the calibration sheet during the image capture. Not all the people that want to calibrate their cameras have the experience and awareness to deliver appropriate calibration image shots, especially students that want to use their amateur cameras for photogrammetric recording purposes. Low luminance and a gloss effect (Figure 5) that might appear when the light is hitting the calibration field page from a certain angle might confuse the software and block the algorithm from identifying the control points on the page.

In general all calibration software functions as a “black box” application and the user has no option to adjust the settings of the algorithms to identify the control points.

When using wide angle lenses and in cases where the camera location has a big inclination with the calibration page, the size of the control points' circles varies a lot. In some cases the circle radius of the most remote circles could be three times smaller than those close to the lens, especially in the case of wide angle lenses (Figure 6).

This last issue severely hinders the success of the calibration adjustment due to possible wrong numbering of the control points on the calibration field. The calibration adjustment function assigns internally to every control point a ground space coordinate according to the distance parallel to the x and y direction of the columns and rows of the calibration sheet. Therefore for the case of a 5 × 7 (rows × columns) calibration grid and a distance of 37 mm between each row and column, the first point on the top left corner of the calibration field gets ground space coordinates (0, 0, 0), and all the control points on the same (first) row have the same y (= 0 mm) and increasing x (from 37.0 to 222.0 mm) and the control points in the succeeding columns get x coordinates from 0 mm to 222 mm and y coordinates from 37 mm to 148 mm. For all the points the height (z coordinate) is the same and is 0.0. Therefore the calibration sheet has to be planar in order to achieve the highest accuracy of the calibration parameter calculation. To identify properly a control point at its correct location on the image the software applies a sorting method. For the case of a regular grid of circles of 5 rows and 7 columns the first seven points, identified on the calibration image, are supposed to be visibly closer to the top edge (first image row of the image). However, if the camera's primary axis (implemented by the line connecting the center of lens and the center of the sensor) is rotated very much the software erroneously sorts the control points and the software fails to calibrate the camera (Figure 7).

2.4. Software Options

For all the above-mentioned problems and to increase the functionality of a calibration application, we decided to implement a tool with a graphical user interface giving to the user the opportunity to have better control on the calibration algorithms, and therefore to:

  • Adjust the threshold algorithm and isolate the circular control points' grid from any noise that might appear on the edges of the calibration images;

  • Define the minimum radius of the control point circles since the wide angle lens and the camera position might affect rapidly the variance of the circles' diameter;

  • Examine the correct location of the control points on the calibration sheet by providing the correct numbering of all the control points on the image;

  • Undistort the still images and create their idealized respective images; the idealized images may be used in single photo photogrammetric procedures to provide rectified photos, i.e., of building facades [20];

  • Isolate frames of a video sequence capturing the circles calibration grid and save them as independent images to perform the calibration of the video sensor; since the calibration of the video sensor is performed the undistortion of any video sequence is feasible.

The created software is characterized by its simplicity and the possibility to interact with other 3D recording software. The user has just to import a set of images using the menus of the Graphical User Interface (GUI) of the software and define the sensor size of the camera to calibrate. The images should be captured in such a way that the calibration page will cover almost the total width and height of the sensor from different locations above of the planar surface of the calibration field (Figure 8). We avoided using a single image calibration setup [21,22], since the result could be inaccurate due to potential asymmetry of the camera construction.

Although the information of the sensor's dimension optionally delivers the calibration parameters in pixels (f, xo, yo) it is important for a photogrammetrist to identify the intrinsic parameters of the cameras in millimeters. The rest of the parameters (k1, k2, k3, p1, p2) are dimensionless numbers.

The sensor dimensions of any commercial camera is easy to be found using the Digital Photography Review [23]. Reviews of almost all of the new and old commercial cameras, including their basic features, are available on the site. Unfortunately though, there is no appropriate information for the GoPro camera or for any cell phone camera. For the GoPro Hero 3 camera the sensor size information can be derived from the official site of the manufacturing company [24], since it gives the dimensions of a single pixel on the camera image in microns (1.55 μm). Therefore for the highest still image resolution the sensor size is estimated to 6.20 × 4.65 mm. Although, the dimensions of the sensor could be calculated from its diagonal (1/2.3") we decided to use the value calculated by multiplying the image width and height with the pixel size given at the official site of the manufacturing company, because the calculated focal length in the case of the pixel size defined sensor dimension (6.20 × 4.65) was 2.6 mm, a value much closer to the nominal (3 mm) given by the manufacturer of the camera. In the case of the sensor size defined by its diagonal in inches, the focal length was estimated at 2.59 mm. In any case the created undistorted images are not affected at all by the sensor size in mm.

Since the user has added the images of the calibration sheet, has defined the sensor size dimensions and also the distance of the control points grid in mm, the calibration (identification of control points on every image and adjustment to calculate the calibration parameters) can follow. The software browses every image in a graphical window on which the user has to define:

  • The threshold value to separate the control points circles from the surrounding area

  • The minimum radius of circles to be identified as control points and

  • The area including points (Figure 9)

Although the target identification process is semi-automatic, the total time spent by an average user (namely the students of our post graduate course in IUAV) to adjust the correct parameters and define the area including the targets did not exceed a couple of minutes. Since this operation is performed only once for every camera (or at least once every six months for professional photogrammetric labor) we decided to follow an absolutely correct calibration process, allowing the user to watch and verify the correct location and numbering (Figure 10) of targets on the calibration field, rather than letting the software perform it without any human interaction and operate as a “black-box” device.

After the user has used the images and the control points have been identified, the software prints in the main window application the results (Figure 11), which can be saved for future use in a project (text) file. In our case the focal length of the GoPro camera was about 2.60 mm and the estimated accuracy (Average Reprojection Error) of the calibration was 3.91 pixels. The calibration project file might be used in the future to undistort images taken with the same camera's modes like the one used to capture the calibration images.

The image undistortion procedures can follow right after the calibration procedure or can be performed in another occasion since the calibration project file has been stored in the computer's disk file system. The user has just to load the images (at least two) and for every single image, the software performs the undistortioning process (Figure 12a,b).

Finally, although not so much photogrammetry related, an important feature of our software is the possibility to calibrate the sensor when capturing video sequences and perform further the undistortioning of a whole video file.

Since the part of the sensor that is used to capture every frame of any video mode is smaller than the full resolution of 12 Mpixels we must calculate the exact dimensions of the sensor to perform the calibration. For this reason another simple software applications was created, which is able to calculate the part of the sensor used to create lower resolution (<12 Mpixels) images, using an affine transformation. We can load and browse in two separate graphics windows a full resolution 12 Mpixels image and a smaller one, taken from the same camera location, and if we can identify at least three common points, we can estimate the dimensions of the sensor part used to capture the low resolution (still or video) image (Figure 13), given that the sensor size of the full resolution image is 6.20 × 4.65 mm.

The video file depicting the calibration sheet can be loaded in the software. The user can isolate any image from the video and save it as a single calibration image. Following the similar procedure of still images' calibration we can perform the calibration of the video frames. After the completion of the video frame calibration, the user can load an original video file and recreate another one, idealized, with respect to the original, using the undistorted frames (Figure 14).

Finally, the software has the ability to import xml files of camera calibration parameters produced by Agisoft Lens or Agisoft Photoscan. Agisoft Lens is a free software that can calculate the same calibration parameters (f, xp, yp, k1, k2, k3, p1, p2) using a calibration screen, while Agisoft Photoscan is a commercial low cost application that can calculate these parameters using just the conjugate points on overlapping scenes of the recorded objects. Unfortunately, the software does not provide the possibility to create the idealized images, but using the undistortion option of our software we can create the undistorted images produced by any calibrated camera.

3. Accuracy Assessment

The use of a widespread and low-cost digital camera combined with easy to use calibration software, such us the free one implemented by the authors (also considering other commercial solutions), offers even non-expert users the possibility to obtain metric data: just considering the immediate results obtainable in teaching or in other different researching areas such as archaeology.

We have tested the use of the GoPro Hero 3 camera on top of a Parrot AR.Drone 2.0 (GPS edition) and have taken a series of images to test the functionality of the camera to provide images of 3D architectural details on a facade not easily reachable from the ground. Both the camera and the drone was controlled from separate mobile devices. The images taken from drone were used to create a complete 3D model of the logia in the university campus giving most accurate results.

For the 3D photogrammetric processing we have used the software Agisoft Photoscan Professional (version 1.04) and the accuracy in alignment of the GoPro distorted images was about 0.035 m (Figure 15). Using the undistorted ones (Figure 16), the RMS is halved (0.015 m).

Moreover, as shown by some tests we performed using different modes of the camera and capturing schemes and considering other published experiences [21], the geometric scheme that optimizes results, in terms of accuracy, is that of zenithal strip coverage of the objects to be recorded (e.g. parallel to the average plane of the surveyed surface for the architectonical facades), with high overlap (Table 5). Oblique photos demonstrated greater RMS error in the alignment as can also be seen in Table 6 and Figure 16.

This integrated system (GoPro + Parrot AR.Drone) is going to be applied and an in depth test in a archeological surveying campaign that will be carried out in the ancient city of Sepinum, in the south of Italy, to verify its reliability, comparing the results obtained by using a “typical” photogrammetric equipment (such as calibrated camera mounted on a drone).

Within the archaeological documentation or supporting restoration, undistorted images are useful when we want to rectify a planar surface (i.e., building's facade) to further digitize its architectural details. After several tests we have made, we came to the conclusion that the accuracy of the rectification is increased significantly when undistorted images are used.

In the case of the GoPro camera, the originally created images are not appropriate to be used for the rectification of planar facades. RMS error (Table 7), due to the distortion, exceeds 1cm and additionally, points that lie very close to the edges of the image are totally unusable (Figure 17a). The rectification of the undistorted image gave very good results (Table 8) and also the curvature effect (of straight lines in ground space) was totally eliminated (Figure 17b).

The same rectification procedure was used also for original and undistorted images of a dSLR NIKON D100 camera (NIKON CORPORATION, Tokyo, Japan) equipped with a very good quality lens (NIKOR 20 mm/2.8 D AF). The RMS errors (Tables 9 and 10) and rectification residuals (Figure 18a,Tables 9 and 10) and rectification residuals (b) were very small, however they were worse than the undistorted GoPro camera case.

4. Conclusions

The possibility to calibrate a low cost camera and especially the GoPro Hero 3 camera which is additionally an action camera is providing lots of benefits:

  • The cost of camera is very low and does not exceed 500€;

  • The cost of accessories necessary to use it in extreme conditions is zero since it is included in the package;

  • The cost of additional equipment to carry the camera for other activities is very small. For UAV applications a low cost drone might be more than enough to carry it and perform remote shootings using just a cell phone applications.

The developed application itself is easy to use by novice users, gives a very good visual corrected result (of the undistorted still images and videos) even to non-photogrammetrists, but provides professional photogrammetrists all the appropriate information to apply the resulting calibration parameters and the derived products for further exploitation in photogrammetric software.

Author Contributions

C.B., F.G., P.V. and V.T. conceived and designed the experiments; V.T., P.V. and C.B. performed the experiments; C.B. and V.T. analyzed the data; F.G. contributed equipment, materials and analysis tools; C.B. and V.T. wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Remondino, F.; El-Hakim, S. Image-based 3D modelling: A review. Photogramm. Rec. 2006, 21, 269–291. [Google Scholar]
  2. Habib, A.; Pullivelli, A.; Mitishita, E.; Ghanma, M.; Kim, E.M. Stability analysis of low-cost digital cameras for aerial mapping using different georeferencing techniques. Photogramm. Rec. 2006, 21, 29–43. [Google Scholar]
  3. Remondino, F.; Fraser, C. Digital Camera Calibration Methods: Considerations and Comparisons. Proceedings of the International Archives of Photogrammetry, Remote Sensing and the Spatial Sciences, Dresden, Germany, 25–27 September 2006; pp. 266–272.
  4. Fraser, C. Automatic Camera Calibration in Close-Range Photogrammetry. Proceedings of the ASPRS 2012 Annual Conference, Sacramento, CA, USA, 19–23 March 2012; pp. 19–27.
  5. Toschi, I.; Rivola, R.; Bertacchini, E.; Castagnetti, C.; Dubbini, M.; Capra, A. Validation test of open source procedures for digital camera calibration and 3D image based modeling. ISPRS Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2013, 1, 647–652. [Google Scholar]
  6. Tsai, R.Y. A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses. IEEE Int. J. Robot. Autom. 1987, 3, 323–344. [Google Scholar]
  7. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar]
  8. Nakano, K.; Chikatsu, H. Camera-Variant Calibration and Sensor Modeling for Practical Photogrammetry in Archeological Sites. Remote Sens. 2011, 3, 554–569. [Google Scholar]
  9. Sirmacek, B.; Lindenbergh, R. Accuracy assessment of building point clouds automatically generated from iphone images. ISPRS Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2014, 1, 547–552. [Google Scholar]
  10. Masiero, A.; Guarnieri, A.; Vettore, A.; Pirotti, F. An ISVD-Based Euclidian Structure from Motion for Smartphones. Proceedings of ISPRS Technical Commission V Symposium, Riva del Garda, Italy, 23–25 June 2014; pp. 401–406.
  11. PhotoModeler Quick Start Guide. Available online: http://www.photomodeler.com (accessed on 10 September 2014).
  12. AgiSoft StereoScan. Available online: http://www.agisoft.ru/ (accessed on 10 September 2014).
  13. MicMac, Software for Automatic Matching in the Geographical Context. Available online: http://www.micmac.ign.fr/index.php?id=6 (accessed on 10 September 2014).
  14. Schmidt, V.E.; Rzhanov, Y. Measurement of micro-bathymetry with a GOPRO underwater stereo camera pair. Proceedings of Oceans, Hampton Road, VA, USA, 14–19 October 2012; pp. 1–6.
  15. Thoeni, K.; Giacomini, A.; Murtagh, R.; Kniest, E. A comparison of multi-view 3D reconstruction of a rock wall using several cameras and a laser scanner. Proceedings of ISPRS Technical Commission V Symposium, Riva del Garda, Italy, 23–25 June 2014; pp. 573–580.
  16. Wallis Forum 2013. Available online: http://www.walis.wa.gov.au/forum/presentations/day-two/slides/3d-modelling-of-underwater-objects-using-photogrammetry-petra-helmholz (accessed on 15 June 2014).
  17. Lehmann Aviation. Available online: http://www.lehmannaviation.com/la/la100.php (accessed on 15 June 2014).
  18. Brown, D.C. Close-range camera calibration. Photogramm. Eng. 1971, 37, 855–866. [Google Scholar]
  19. Fryer, J.G. Camera Calibration. In Close Range Photogrammetry and Machine Vision; Atkinson, K.B., Ed.; Whittler Publishing: Caithness, Scotland, UK, 1996; pp. 156–179. [Google Scholar]
  20. Tsioukas, V. Simple Tools for Architectural Photogrammetry. Proceedings of the XXI International CIPA Symposium, Athens, Greece, 1–6 October 2007; pp. 535–539.
  21. PhotoModeler. Available online: http://info.photomodeler.com/blog/using-the-gopro-hero-3-for-`3d-photogrammetry-modeling-and-measuring/ (accessed on 15 August 2014).
  22. HTW-Mechlab. Available online: http://www.htw-mechlab.de/index.php/intrinsic-camera-parameter-fuer-gopro-hd-hero2/ (accessed on 15 August 2014).
  23. Digital Photography Review. Available online: www.dpreview.com (accessed on (accessed on 15 June 2014).
  24. GoPro. Available online: http://gopro.com/support/articles/hero3-faqs (accessed on 15 June 2014).
Figure 1. The (a) distorted image from the GoPro Hero 3 Camera and its (b) undistorted version. The undistorted image has been created using the special remapping function of OpenCV.
Figure 1. The (a) distorted image from the GoPro Hero 3 Camera and its (b) undistorted version. The undistorted image has been created using the special remapping function of OpenCV.
Sensors 14 17471f1 1024
Figure 2. On the (a) original and (b) idealized images the deformation error is visualized.
Figure 2. On the (a) original and (b) idealized images the deformation error is visualized.
Sensors 14 17471f2a 1024Sensors 14 17471f2b 1024
Figure 3. Detail of the distortions.
Figure 3. Detail of the distortions.
Sensors 14 17471f3 1024
Figure 4. The upper left part of the image is appearing defocused. The software was unable to identify the intersection of the squares and perform the calibration.
Figure 4. The upper left part of the image is appearing defocused. The software was unable to identify the intersection of the squares and perform the calibration.
Sensors 14 17471f4 1024
Figure 5. The gloss effect on the top left corner of the calibration field is giving brighter colors to some circles of the calibration field.
Figure 5. The gloss effect on the top left corner of the calibration field is giving brighter colors to some circles of the calibration field.
Sensors 14 17471f5 1024
Figure 6. In case of wide angle lens cameras and when the inclination of the camera is big the circle's diameter of the control points might be three times smaller at the edges of the calibration image.
Figure 6. In case of wide angle lens cameras and when the inclination of the camera is big the circle's diameter of the control points might be three times smaller at the edges of the calibration image.
Sensors 14 17471f6 1024
Figure 7. High inclination of a camera's principal axis giving wrong sorting and numbering of control points (red colored points' codes). The point with No. 8 (code in white color inside the circle) is erroneously getting the code No. 4 (red colored code) just because its j coordinate on the image (distance from first image row) is smaller than the j coordinate of all the other points below the green line.
Figure 7. High inclination of a camera's principal axis giving wrong sorting and numbering of control points (red colored points' codes). The point with No. 8 (code in white color inside the circle) is erroneously getting the code No. 4 (red colored code) just because its j coordinate on the image (distance from first image row) is smaller than the j coordinate of all the other points below the green line.
Sensors 14 17471f7 1024
Figure 8. Images captured to perform the calibration.
Figure 8. Images captured to perform the calibration.
Sensors 14 17471f8 1024
Figure 9. The user defines with the use of the sliders the parameters for the software and defines on the graphical window the region including the control points.
Figure 9. The user defines with the use of the sliders the parameters for the software and defines on the graphical window the region including the control points.
Sensors 14 17471f9 1024
Figure 10. Since the control points are identified correctly the software displays the numbers of control points on the original calibration image.
Figure 10. Since the control points are identified correctly the software displays the numbers of control points on the original calibration image.
Sensors 14 17471f10 1024
Figure 11. Calibration parameters as they are calculated for the 12 Mpixels still images from the GoPro Hero 3 camera.
Figure 11. Calibration parameters as they are calculated for the 12 Mpixels still images from the GoPro Hero 3 camera.
Sensors 14 17471f11 1024
Figure 12. (a) Original and (b) undistorted images of the calibration field. A small part of the image is lost however considering the geometric correction we can accept the loss.
Figure 12. (a) Original and (b) undistorted images of the calibration field. A small part of the image is lost however considering the geometric correction we can accept the loss.
Sensors 14 17471f12 1024
Figure 13. The affine transformation gave the part of the sensor used to capture the low resolution (848 × 480) video frames. The blue rectangle defines the placement of the low resolution image in the full frame image.
Figure 13. The affine transformation gave the part of the sensor used to capture the low resolution (848 × 480) video frames. The blue rectangle defines the placement of the low resolution image in the full frame image.
Sensors 14 17471f13 1024
Figure 14. (a) Original and (b) undistorted video sequences.
Figure 14. (a) Original and (b) undistorted video sequences.
Sensors 14 17471f14 1024
Figure 15. The dense point cloud obtained using undistorted oblique images. RMS error of collected GCPs is about 0.035 m. The test shows that perpendicular undistorted images allow to obtain a higher accuracy.
Figure 15. The dense point cloud obtained using undistorted oblique images. RMS error of collected GCPs is about 0.035 m. The test shows that perpendicular undistorted images allow to obtain a higher accuracy.
Sensors 14 17471f15 1024
Figure 16. A dense point cloud obtained in Agisoft Photoscan using the GoPro undistorted images. RMS error of collected GCPs is about 0.015 m.
Figure 16. A dense point cloud obtained in Agisoft Photoscan using the GoPro undistorted images. RMS error of collected GCPs is about 0.015 m.
Sensors 14 17471f16 1024
Figure 17. The rectification of the original (a) and undistorted GoPro images (b) lead to an RMS error of the collected GCPs of 0.0103 m and 0.002 m respectively using the VeCAD-Photogrammetry application [20]. It is also visible the curvature of straight lines on the original image (a).
Figure 17. The rectification of the original (a) and undistorted GoPro images (b) lead to an RMS error of the collected GCPs of 0.0103 m and 0.002 m respectively using the VeCAD-Photogrammetry application [20]. It is also visible the curvature of straight lines on the original image (a).
Sensors 14 17471f17 1024
Figure 18. The rectification of the original (a) and undistorted NIKON D100 images; (b) lead to an RMS error of the collected GCPs of 0.0013 m and 0.0001 m using the VeCAD—Photogrammetry application [20].
Figure 18. The rectification of the original (a) and undistorted NIKON D100 images; (b) lead to an RMS error of the collected GCPs of 0.0013 m and 0.0001 m using the VeCAD—Photogrammetry application [20].
Sensors 14 17471f18 1024
Table 1. Camera weight.
Table 1. Camera weight.
WeightCamera Part
51 gramscamera
25 gramsbattery
76 gramscamera with battery
19 gramsquick-release buckle
106 gramswaterproof housing with quick-release buckle
182 gramstotal
Table 2. Camera dimensions.
Table 2. Camera dimensions.
DimensionsCamera Part
72 mm × 65 mm × 37 mmhousing size
13 mm × 16 mmLCD screen size
60 mm × 40 mm × 20 mmcamera size
Table 3. Still image resolutions.
Table 3. Still image resolutions.
Image TypeWidthHeight
12 Mp40003000
7 Mp wide30002250
7 Mp medium30002250
5 Mp medium25601920
Table 4. Video resolutions.
Table 4. Video resolutions.
Video TypeWidthHeightFPSImage Ratio (Width:Height)
1080p1920108029.97(16:9)
720p128072059.94(16:9)
1440p1920144029.97(4:3)
4K3840216014.99(16:9)
4K Cin4096216011.99(17:9)
2.7K2716152429.97(16:9)
2.7K Cin2716144023.98(17:9)
960p128096047.95(4:3)
WVGA848480240.00(16:9)
Table 5. Zenithal undistorted strip errors.
Table 5. Zenithal undistorted strip errors.
# LabelXYZX ErrorY ErrorZ Error
target 1102.153107.353100.044−0.0063120.0029430.005454
target 21105.664107.846100.5560.0120980.008989−0.015523
target 25104.728107.90698.973−0.018369−0.0040260.012258
target 27108.083108.16698.855−0.0034760.0020340.020901
target 3103.476107.535100.058−0.001126−0.000676−0.002830
target 5104.411107.669100.0650.001108−0.005677−0.006631
target 7105.357107.800100.0620.006128−0.000555−0.008944
target 9107.146108.036100.0570.009943−0.003037−0.004682

# Total Error0.0091830.0043590.011246
Table 6. Oblique undistorted strip errors.
Table 6. Oblique undistorted strip errors.
# LabelXYZX ErrorY ErrorZ Error
target 1102.153107.353100.044−0.036420−0.0368480.011161
target 21105.664107.846100.5560.0276850.014427−0.024244
target 25104.728107.90698.9730.0001800.0279690.007661
target 27108.083108.16698.855−0.030725−0.0255170.041410
target 3103.476107.535100.058−0.002215−0.000803−0.001226
target 5104.411107.669100.0650.0102430.011765−0.008236
target 7105.357107.800100.0620.0215850.019010−0.013242
target 9107.146108.036100.0570.009631−0.010095−0.013273

# Total Error0.0215220.0212180.019061
Table 7. Ground, image coordinates and their ground residuals of the original GoPro image.
Table 7. Ground, image coordinates and their ground residuals of the original GoPro image.
Point IDXYjiX ResidualY Residual
1−1.610.8291720.48641.730.00090.0641
2−0.5860.8092195.29638.970.0179−0.023
30.5810.7752730.22683.97−0.0179−0.034
41.6640.8073166.44723.970.02290.0249
51.6630.1523147.071006.23−0.00460.0136
60.5650.1652715.13959.19−0.0405−0.0355
7−0.5920.0992193.27964.710.0054−0.0356
8−1.620.1011723.2965.210.00540.0132
91.641−0.9543077.091460.360.02790.0242
100.512−0.9242656.131433.9−0.02450.0064
11−0.579−0.9432190.11428.90.0092−0.0046
12−1.638−0.9161731.911405.23−0.002−0.0137

RMS Error0.010945
Table 8. Ground, image coordinates and their ground residuals of the undistorted GoPro image.
Table 8. Ground, image coordinates and their ground residuals of the undistorted GoPro image.
Point IDXYjiX residualY residual
1−1.610.8291703.97588.51−0.00130.0078
2−0.5860.8092209.41586.760.00150.0014
30.5810.7752815.96597.280.00560.0048
41.6640.8073429.66561.77−0.0062−0.0062
5−1.620.1011716.3952.810.0022−0.0052
6−0.5920.0992198.11953.3−0.0042−0.0075
70.5650.1652768.16923.410.0019−0.0014
81.6630.1523342.25931.630.0010.0013
9−1.638−0.9161730.761405.370.00590
10−0.579−0.9432190.981428.61−0.00780.0006
110.512−0.9242682.771432.91−0.00640.0002
121.641−0.9543209.041461.310.00790.0042

RMS Error0.002092
Table 9. Ground, image coordinates and their ground residuals of the original NIKON D100 image.
Table 9. Ground, image coordinates and their ground residuals of the original NIKON D100 image.
Point IDXYjiX residualY residual
1−1.610.8291599.22569.09−0.00030.0057
2−0.5860.8091876.33556.10.00310.0001
30.5810.7752270.87544.23−0.0023−0.0023
41.6640.8072735.2500.860.0044−0.0012
5−1.620.1011596.36819.95−0.00060.0007
6−0.5920.0991872.55826.29−0.0002−0.0016
70.5650.1652259.95806.93−0.0056−0.002
81.6630.1522725.48821.06−0.0017−0.0005
91.641−0.9542692.671347.20.00210.0046
100.512−0.9242228.751262.34−0.00290.0004
11−0.579−0.9431870.541216.440.00510.0003
12−1.638−0.9161591.621163.88−0.0012−0.0041

RMS Error0.001299
Table 10. Ground, image coordinates and their ground residuals of the undistorted NIKON D100 image.
Table 10. Ground, image coordinates and their ground residuals of the undistorted NIKON D100 image.
Point IDXYjiX residualY residual
1−1.610.8291599.29567.49−0.00130.0024
2−0.5860.8091878.83553.340.00110.0011
30.5810.7752281.46538.24−0.00180.0003
41.6640.8072767.29487.130.0016−0.0022
5−1.620.1011596.37819.79−0.00150.0004
6−0.5920.0991873.35825.650.0024−0.0005
70.5650.1652268.08804.35−0.0013−0.0015
81.6630.1522752.99816.50.0005−0.0008
91.641−0.9542719.061354.25−0.00150.0027
100.512−0.9242236.031265.73−0.00180.0013
11−0.579−0.9431871.51217.050.0057−0.0002
12−1.638−0.9161591.71163.88−0.0023−0.003

RMS Error0.000981

Share and Cite

MDPI and ACS Style

Balletti, C.; Guerra, F.; Tsioukas, V.; Vernier, P. Calibration of Action Cameras for Photogrammetric Purposes. Sensors 2014, 14, 17471-17490. https://doi.org/10.3390/s140917471

AMA Style

Balletti C, Guerra F, Tsioukas V, Vernier P. Calibration of Action Cameras for Photogrammetric Purposes. Sensors. 2014; 14(9):17471-17490. https://doi.org/10.3390/s140917471

Chicago/Turabian Style

Balletti, Caterina, Francesco Guerra, Vassilios Tsioukas, and Paolo Vernier. 2014. "Calibration of Action Cameras for Photogrammetric Purposes" Sensors 14, no. 9: 17471-17490. https://doi.org/10.3390/s140917471

Article Metrics

Back to TopTop