Next Article in Journal
PCB-Based Magnetometer as a Platform for Quantification of Lateral-Flow Assays
Previous Article in Journal
Optical Harmonic Vernier Effect: A New Tool for High Performance Interferometric Fiber Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Global Positioning from a Single Image of a Rectangle in Conical Perspective

by
Manuel Estrems Amestoy
1 and
Óscar de Francisco Ortiz
2,*
1
Mechanics, Materials and Manufacturing Engineering department, Technical University of Cartagena, 30202 Cartagena, Spain
2
Department of Engineering and Applied Technologies, University Center of Defense, San Javier Air Force Base, MDE-UPCT, 30720 Santiago de la Ribera, Spain
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(24), 5432; https://doi.org/10.3390/s19245432
Submission received: 22 November 2019 / Revised: 5 December 2019 / Accepted: 6 December 2019 / Published: 10 December 2019
(This article belongs to the Section State-of-the-Art Sensors Technologies)

Abstract

:
This article presents a method to obtain the overall positioning of the focus of a camera from an image that includes a rectangle in a fixed reference with known position and dimension. This technique uses basic principles of descriptive geometry introduced in engineering courses. The document will first show how to obtain the dihedral projections of a rectangle after three turns and one translation. Secondly, we will proceed to obtain the image of the rectangle rotated in a conical perspective, taking the elevation plane as the drawing plane and a specific point in space as the view point, and represented in the dihedral system. Thirdly, we proceed with the inverse perspective transformation; we will expose a method to obtain the coordinates in the space of a rectangle obtained from an image. Finally, we check the method experimentally by taking an image of the rectangle with a camera in which the coordinates in the drawing plane (center of the image) are the only available position information. Then, the positioning and orientation of the camera in 3D will be obtained.

1. Introduction

Pose determination is to estimate the position and orientation of one calibrated camera using a set of correspondences between 3D control points and 2D image points [1]. Determination of surface orientation has important applications such robotics, object recognition, 3D measurement or tracking of moving objects. Magee [2] was the first to present a procedure for determining the unique position of a robot in a three dimensional space. That method has been continuously improved in different areas as large non-cooperative satellites [3] or Unmanned Aerial Vehicle (UAV) Control [4,5]. Different methods for monocular pose estimation have been studied in the past [6,7,8,9]. More recently, marker-based positioning systems as ArUco, Chilitags, ApriTags, or ArToolKit, among others, have been introduced to estimate quantitative changes in distances and orientations in many technological applications, such as autonomous robots [10,11,12], unmanned vehicles [13,14,15,16], or virtual assistants [17,18,19,20].
The calibration and orientation of a camera from its images has been obtained through different approaches in the past with good precision through techniques such as using a single image with four coplanar control lines [21], three coplanar circles [22], using parallelogrammatic grid points [23], or even using only three points in the world coordinate system when a multiple camera system is used [24]. Becker [25] introduced a new technique using an iterative method which solves the parameters that minimize vanishing point dispersion to solve for radial and decentering lens distortion directly from the results of vanishing point estimation, precluding the need for special calibration templates. Single image based reconstruction has been deeply studied by many authors such as Delage [26], Wilczkowiak et al. [27], Sturm and Maybank [28] or Micusik et al. [29], assuming perpendicularity and parallelism to recover the lack of information. Other authors such as Penna [30] showed that there is sufficient information in the two-dimensional perspective projection of an arbitrary quadrilateral of known shape and size in three-space to determine the exact three-dimensional coordinates of its vertices, generalizing known results for rectangles. Duan [1] used the projection of a trapezium for pose estimation and plane measurement in a very simple way. An iterative algorithm was used by Hong & Yang [31] to establish the relationship between parameters and the world coordinates of a given 3D calibration point. Nevertheless, additional studies via rectangular structures as in Haralick [8] or Wefelscheid [32] use similar concepts with a different approach. In contrast, our research used the information provided by the dihedral projections of a rectangle to determine the image of the rectangle rotated in a conical perspective.
Computer vision has been used in areas, such as unmanned vehicles, to estimate relative 3D position and altitude using algorithms based on four feature points, such as square and parallel relations, to avoid complicated calculations [33]. An algorithm for pose estimation based on volume measurement of tetrahedra composed of target points and the lens center of the vision system was proposed by Abidi [6]. 3D model reconstruction from a single image calibrating a camera and recovering the geometry and the photometry of objects was part of Guillou’s [34] research and a novel method to find the initial solutions for iterative camera pose estimation using coplanar points was provided by Zhou [35]. A general photogrammetric method for determining object position and orientation was presented by Yuan [36]. Recently Wang et al. [37] studied active relocalization of a 3D camera pose from a single reference image; a recent and challenging problem in computer vision and robotics. Pose estimation of smooth metal parts is an important task in intelligent manufacturing. Ulrich [38], Sakcak [39], Han [40] and He [41] proposed a solution using a monocular camera and corresponding practical algorithms.
The adjustment of tools in machining centers is usually the slowest and most critical operation in the positioning of the machined parts. The provision of a tool that includes machine displacements and images with edge detection can be adjusted at micrometric scales without the need for lasers or probes. Other possible application could be the metrology by vision, since in the case of characteristics to be measured in the same plane of a rectangle of known dimensions, the dihedral perspective of the aforementioned characteristic can be obtained and non-contact metrological checks can be performed immediately (in real time) compensating many of the existing errors. This is an essential aspect to achieve the efficiency and flexibility required by controls in production systems in Industry 4.0.
In this work a new method to obtain the camera coordinates of a rectangle from its image is proposed. This method is based in the principles of descriptive geometry as developed by Monge [42], which is studied in basic engineering courses. In order to explain the method a remembrance of the construction of a rectangle in conical perspective is described, and an inverse path is proposed. Finally, an experiment has been designed to check the precision of the method.

2. Dihedral Projection of a Rectangle. Rotations and Translations

In this case the problem input data is the dihedral projection of a rectangle in which the length of one side L is known. Therefore, it is represented by its coordinates x * and z * . This rectangle is rotated by three angles ϕ , ξ , and θ . The transformation matrices are applied to obtain a global rotation matrix and the translation is made to the point X 0 , the coordinates of the vertices are then obtained and presented in a table of dihedral information. The Top View of the dihedral would be represented by the x y plane, and the elevation of the dihedral is the x z plane. The projections of the rectangle on both planes will be its dihedral representation [42,43].

3. Conical Projection

With the point of view with coordinates ( V x , V y , 0 ) and represented in the same dihedral system as the rectangle, where the Front View coincide with the image plane and V y is the focal distance, the vertices coordinates ( x * , z * ) in the rectangle in the conical perspective are obtained. The method used consists in creating, from the Top View, a line that passes through ( V x , V y , 0 ) and the Top Projection of the point P ( P x , P y , 0 ) obtaining the intersection with the image plane which will be the coordinate x p * . This coordinate x p * is calculated by drawing the line that passes through ( V x , V y , 0 ) and the Front View of the point P ( P x , P y , 0 ) and obtaining the intersection with the vertical line that starts at x * . Consequently, the conical projection of the point in the image plane with coordinates ( x * , z * ) is calculated. When this operation is performed Figure 1 with the four points in the rectangle, the rectangle in conical perspective is obtained.

4. Obtaining the Possible Front View and Top View of Dihedral Projection of the Rectangle

Using the coordinates in the conical perspective and knowing the projection of the point of view in the Front and Top planes, the coordinates of the edges of the rectangle are calculated (Figure 2).
To do this, we use part of the geometric method described by Wefelscheid et al. [32] obtaining auxiliary points that help us calculate the dihedral projection from the conical projection. These auxiliary points are:
  • Point M * . Intersection between the lines joining P 1 * P 3 * and P 2 * P 4 * .
  • Vanishing point V 1 * as intersection between the lines P 1 * P 2 * and P 3 * P 4 * .
  • Vanishing point V 2 * as intersection between the lines P 1 * P 4 * and P 2 * P 3 * .
  • Midpoint of edges P 12 * , P 23 * , P 34 * , P 14 * as intersection of the lines that are drawn from the vanishing point to M * with the respective edges.
The auxiliary points are represented in Figure 3.
These operations can be done graphically by drawing on paper, so it is computationally reduced to intersections between lines that are defined each by two points as in Table 1 as represented in Figure 3.
After obtaining these points, a proposal of the Front View of the rectangle is based on two graphic properties:
  • The points of the Front View projection are in the lines that start from the center point V whose coordinates are ( V x , 0 , V z ) and go though the point of the image P 1 * , P 2 * , P 3 * , P 4 * , M * , P 12 * , P 23 * , P 34 * , P 14 * .
  • In the dihedral projection the center points are in the geometric center of the segment of the side, dividing this side in two. For example, P 12 is in the center point of the segment that joins P 1 and P 2 .
  • Opposite sides are parallel in the dihedral projection.
By taking advantage of these two properties and a trigonometric interrelation, a first proposal of a rectangle in Front View can be obtained by the following procedure:
  • In the triangle P 1 * P 2 * V * which is divided by the segment V * P 12 * , a line that starts at P 12 and its intersection with the lines V * P 1 * and V * P 2 * is equidistant, in a way that a possible point P 12 in Front View can be obtained as shown in Figure 4.
  • An arbitrary distance d to obtain P 12 is taken.
  • The normal vector of the line P 1 P 2 in dihedral will be found by a rotation of the vector V * P 12 an angle ω reached using the trigonometric relation (1):
    tan ω = 2 cot α + cot β .
    The deduction of this expression is detailed in Appendix A, where α is the angle between V * P 2 * and V * P 12 * ; and β is the angle between V * P 1 * and V * P 12 * represented in the Figure 4 and expressed by the Equations (2) and (3).
    α = arccos V * P 2 * · V * P 12 * | V * P 2 * | | V * P 12 * |
    β = arccos V * P 1 * · V * P 12 * | V * P 1 * | | V * P 12 * | .
  • Points P 1 and P 2 are obtained from the intersection of the line defined by the point P 12 and the vector V * P 12 rotated an angle ω . Once the orientations are calculated, starting from a point in the line V * P 1 * and drawing a line that intersects the line V * P 2 * gets the hypothetical side P 1 P 2 already in the Front View of the dihedral projection.
  • With the presumed points P 1 and P 2 of the Front View in the dihedral projection, it is possible to calculate, with the central point V in the Top view (which is at a distance equal to the focal distance from the drawing plane), the projection in Top view of points P 1 and P 2 as shown in Figure 5.
    To get the Top View of the rectangle from the hypothetical Front projection of the side P 1 P 2 it is possible to obtain its Top projection from the Top projection of V which is at a focal distance from the drawing plane. As an example, the y coordinate of the point P 1 will be obtained by the intersection of the line that joins V in the Top View with the point x P 1 * and the coordinate x P 1 as shown in Figure 6 and in Equation (4). In the same way we proceed to obtain the y coordinate of the point P 2 .
    y p 1 = V y + V y V x x P 1 *
  • With the Top and Front projection of the points P 1 y P 2 , the length in pixel units of the segment is calculated, and the real distance d that means the length P 1 P 2 matched with the length of the edge of the rectangle is obtained.
    With the complete coordinates of the points P 1 y P 2 the distance between both points is calculated.
    Being a proportional geometric problem, the solution is found in a single step from the application of Thales’ theorem.
  • A similar procedure is done with the triangle P 2 * P 3 * V * divided by V * P 23 . Consequently, the two orientations of the edges which will have the projection of the rectangle in the dihedral system are calculated.
    Getting the rest of the points is direct as we have the orientations in Top View, finding the points P 3 y P 4 using a correlative method.
Once the three-dimensional coordinates of the rectangle are found, it is possible to perform any operation related to positioning and orientation of the camera or distance calculation and angle modification. The full method is represented in Figure 7.

5. Comments on the Described Method and Comparison with Previous Ones

The new method has many advantages over the methods of Haralik [8] and Wefelscheid [32] which are the most used:
  • It has the advantage of working with points and lines as it works in descriptive geometry science, making the calculations much more intuitive, based on simple sequences.
  • The method makes little use of trigonometric functions. The only trigonometric relation used is the tangent angle function between two vectors which induces very few floating point errors. In addition to this, a rotation is applied on the vectors to redraw the rectangle edges in dihedral.
  • It is a direct method without iterations or matrix inversions.
  • As it is sequential, we can perform checks and easily determine where an error may have occurred. Once the calculations have been verified, the equations in mega formulas that save the calculation times can be exposed. The algebraic operations to obtain the points barely exceed one hundred which equates to less than thousandths of a second of computer time.
With the results, several verification can be performed since it provides data which can already be calculated such as:
  • The length of the second edge of the rectangle, since it has not been used for the calculation of the inverse perspective.
  • The spatial lines that join the point V with the vanishing points V 1 and V 2 in the drawing plane, are parallel to the sides of the rectangle so they are perpendicular to each other. Consequently, the scalar product must be zero, which means that the starting data (the focal length) can actually be determined from the vanishing points [34].

6. Positioning of the Camera in Coordinate System Defined in the Rectangle

Object tracking systems in space through images, navigation systems or calculation of distances and angles in images can be easily made from the coordinates of known rectangles that serve as a reference (Figure 8). Therefore, they can be used for tracker systems in positioning of parts in specific coordinate systems.
In global coordinates (5):
i = P 1 P 2 | P 1 P 2 | j = P 4 P 1 | P 4 P 1 | k = P 1 P 2 P 4 P 1 | P 1 P 2 P 4 P 1 | = i j .
The position in global coordinates would be given by (6):
X v = x v x M y v y M z v z M ,
and the check in local coordinates would be obtained according to (7):
X v = M V · i M V · j M V · k .
In global coordinates the orientation of the camera follows the j v vector (8):
0 1 0 j v .
In rectangle coordinates (9) is obtained, which is the projection of j on each of the three axes that coincides with the component in y of the three vectors i , j , and k expressed in global coordinates.
j v x j v y j v z = i · j v j · j v k · j v .

7. Experimental Tests

For the test of the method all the images are taken with a CASIO Exilim EX-ZR200 digital camera with a resolution of 4608 × 3456 (16MPixels) and a sensor dimension of 6.16 × 4.62 mm (1/2.3”). After the calibration the focal distance is 4.6 mm and the central point is not in the middle of the image but at coordinates (2186, 1991). The camera is calibrated by standard method of Computer Vision by images taken of a chessboard. To analyze the position a Coordinate Measuring Machine (CMM) model Pioneer DEA 03.10.06 with measuring strokes 600 × 1000 × 600 mm has been used as seen in Figure 9. The Maximum Permissible Error of the DEA in the measurements is 2.8 + 4.0 L/1000 μ m. The software for the measurements was PC-DMIS.
The procedure followed was as follows:
  • Place a DIN A4 size paper on the granite table.
  • Position the camera on a tripod.
  • Take photo of DIN A4 paper remotely so as not to influence the captured image.
  • Take six points of the camera housing according to the 3-2-1 method [44].
  • Calculate the position of the camera focus with respect to the center of the A4 sheet from the palpated points.
  • Contrast this position with that obtained by image analysis.
  • Repeat several times the steps 3–6, while varying the position and angles of the camera.
The parameters used for the test are summarized in Table 2. The coordinates and distances of the paper center obtained by CMM and the image analysis including the differences of the coordinates and distances between both, where subindex e refers to experimental data calculated with the CMM and subindex t refers to theoretical values calculated by the image analysis algorithm, are shown in the Table 3.
Result comments:
  • The differences in measured distances are less than 2%, e.g., less than 2 cm in 1 m distance.
  • The errors in the x coordinate are due to the parallelism of the lines with the image plane, but they hardly affect the distance calculation since the contribution of the x coordinate is small in the global calculation.
As can be seen in Table 3 and Figure 10, the difference between the distance from the camera to the center of the folio, measured by the image analysis and by the CMM, is less than 2%. This indicates that, visually, very close accuracies of the actual distances can be achieved. Nevertheless, analyzing each coordinate, a very high error in the x coordinates is observed in some points. These errors occur when the camera is facing the folio, being the x-axis parallel to the sensor plane. In consequence, the vanishing points in this direction are far apart, and the intersection between the lines is more imprecise. As in these cases, the camera is located at a small value of the x coordinate, the influence of this value on the global error is reduced. This indicates a limitation since the method works best, the closer the vanishing points remain.
Figure 11 shows the images obtained with the camera and used to verify the method presented in this article.

8. Conclusions

A new method has been proposed for rectangle reconstruction using elements of descriptive geometry, as used by Monge in 1847 [42], and of extensive knowledge by engineering users since it is taught in the early stages of such studies. The method presented is mainly based on the intersection between lines, as their calculations are fast and stable in computing and, therefore, minimize errors and optimize computation. The proposed process uses very few trigonometric functions of small angles that are the main source of errors in other methods, so very few floating-point errors are introduced. Additionally, the trigonometric functions are mainly used for the rotation of vectors to align the edges in dihedral projections, which also reduces the errors.
In addition, a procedure was carried out to experimentally test the calculations. The proposed technique was tested in a CMM by locating the camera through the palpation using the 3-2-1 method and the position given by the CMM was compared with the calculation from the image taken by the camera. The proposed method provides maximum errors of 2% in the measured distances. The big errors detected in individual coordinates are due to the parallelism of two sides with the image plane since, in this case, the vanishing point is distance in space and its determination by the intersection of two almost parallel lines has more variability.

Author Contributions

Conceptualization, M.E.A. and Ó.d.F.O.; methodology, M.E.A. and Ó.d.F.O.; software, Ó.d.F.O. and M.E.A.; validation, Ó.d.F.O. and M.E.A.; formal analysis, M.E.A.; writing–original draft preparation, Ó.d.F.O.; writing–review and editing, M.E.A. and Ó.d.F.O.; supervision, M.E.A.; visualization, Ó.d.F.O. and M.E.A.

Funding

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
CMMCoordinate Measurement Machine
DINDeutsches Institut für Normung (German Institute for Standardization)

Appendix A. Obtaining the Expression of tanω

In order to identify the sign of α and β the vector product is calculated and the corresponding sign is found (A1) and (A2):
sin α = V * P 2 * V * P 12 * | V * P 2 * | | V * P 12 * |
sin β = V * P 1 * V * P 12 * | V * P 1 * | | V * P 12 * | .
The angle ω can be obtained from α and β applying the sine law (A3):
L 2 sin α = | V * P 12 * | sin ( π α ω ) = | V * P 12 * | sin α cos ω + cos α sin ω
L 2 sin β = | V * P 12 * | sin ( ω β ) = | V * P 12 * | sin ω cos β + cos ω sin β
L cos ω 2 sin α = | V * P 12 * | sin α + cos α tan ω
L cos ω 2 sin β = | V * P 12 * | cos β tan ω sin β
L cos ω 2 = | V * P 12 * | 1 + cot α tan ω = | V * P 12 * | cot β tan ω 1
1 + cot α tan ω = cot β tan ω 1
2 + cot α tan ω = cot β tan ω .
Consequently, the angle of rotation ω can be calculated using the Equation (A10):
tan ω = 2 cot α + cot β .

References

  1. Duan, F.; Fuchao, W.; Hu, Z. Pose determination and plane measurement using a trapezium. Pattern Recognit. Lett. 2008, 29, 223–231. [Google Scholar] [CrossRef]
  2. Magee, M.; Aggarwal, J. Determining the position of a robot using a single calibration object. In Proceedings of the IEEE International Conference on Robotics and Automation, Atlanta, GA, USA, 13–15 March 1984; Volume 1, pp. 140–149. [Google Scholar] [CrossRef]
  3. Gao, X.H.; Liang, B.; Pan, L.; Li, Z.H.; Zhang, Y.C. A Monocular Structured Light Vision Method for Pose Determination of Large Non-cooperative Satellites. Int. J. Control. Autom. Syst. 2016, 14, 1535–1549. [Google Scholar] [CrossRef]
  4. Martinez, C.; Mondragon, I.F.; Olivares-Mendez, M.A.; Campoy, P. On-board and Ground Visual Pose Estimation Techniques for UAV Control. J. Intell. Robot. Syst. 2011, 61, 301–320. [Google Scholar] [CrossRef] [Green Version]
  5. Zhang, L.; Zhai, Z.; He, L.; Wen, P.; Niu, W. Infrared-Inertial Navigation for Commercial Aircraft Precision Landing in Low Visibility and GPS-Denied Environments. Sensors 2019, 19, 408. [Google Scholar] [CrossRef] [Green Version]
  6. Abidi, M.A.; Chandra, T. A new efficient and direct solution for pose estimation using quadrangular targets—Algorithm and evaluation. IEEE Trans. Pattern Anal. Mach. Intell. 1995, 17, 534–538. [Google Scholar] [CrossRef] [Green Version]
  7. Haralick, R.M. Using perspective transformations in scene analysis. Comput. Graph. Image Process. 1980, 13, 191–221. [Google Scholar] [CrossRef]
  8. Haralick, R.M. Determining camera parameters from the perspective projection of a rectangle. Pattern Recognit. 1989, 22, 225–230. [Google Scholar] [CrossRef]
  9. Quan, L.; Lan, Z. Linear N-point camera pose determination. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 774–780. [Google Scholar] [CrossRef]
  10. Sim, R.; Little, J. Autonomous vision-based robotic exploration and mapping using hybrid maps and particle filters. Image Vis. Comput. 2009, 27, 167–177. [Google Scholar] [CrossRef]
  11. Valencia-Garcia, R.; Martinez-Béjar, R.; Gasparetto, A. An intelligent framework for simulating robot-assisted surgical operations. Expert Syst. Appl. 2005, 28, 425–433. [Google Scholar] [CrossRef]
  12. Pichler, A.; Akkaladevi, S.; Ikeda, M.; Hofmann, M.; Plasch, M.; Wögerer, C.; Fritz, G. Towards Shared Autonomy for Robotic Tasks in Manufacturing. Procedia Manuf. 2017, 11, 72–82. [Google Scholar] [CrossRef]
  13. González, D.; Pérez, J.; Milanés, V. Parametric-based path generation for automated vehicles at roundabouts. Expert Syst. Appl. 2017, 71, 332–341. [Google Scholar] [CrossRef] [Green Version]
  14. Sanchez-Lopez, J.; Pestana, J.; De La Puente, P.; Campoy, P. A reliable open-source system architecture for the fast designing and prototyping of autonomous multi-UAV systems: Simulation and experimentation. J. Intell. Robot. Syst. 2015, 84, 779–797. [Google Scholar] [CrossRef] [Green Version]
  15. Romero-Ramirez, F.J.; Muñoz-Salinas, R.; Medina-Carnicer, R. Speeded up detection of squared fiducial markers. Image Vis. Comput. 2018, 76, 38–47. [Google Scholar] [CrossRef]
  16. Germanese, D.; Leone, G.R.; Moroni, D.; Pascali, M.A.; Tampucci, M. Long-Term Monitoring of Crack Patterns in Historic Structures Using UAVs and Planar Markers: A Preliminary Study. J. Imaging 2018, 4, 99. [Google Scholar] [CrossRef] [Green Version]
  17. Pflugi, S.; Vasireddy, R.; Lerch, T.; Ecker, T.; Tannast, M.; Boemke, N.; Siebenrock, K.; Zheng, G. Augmented marker tracking for peri-acetabular osteotomy surgery. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Seogwipo, South Korea, 11–15 July 2017; pp. 937–941. [Google Scholar] [CrossRef]
  18. Lima, J.P.; Roberto, R.; Simões, F.; Almeida, M.; Figueiredo, L.; Teixeira, J.M.; Teichrieb, V. Markerless tracking system for augmented reality in the automotive industry. Expert Syst. Appl. 2017, 82, 100–114. [Google Scholar] [CrossRef]
  19. Chen, P.; Peng, Z.; Li, D.; Yang, L. An improved augmented reality system based on AndAR. J. Vis. Commun. Image Represent. 2016, 37, 63–69. [Google Scholar] [CrossRef]
  20. Khattak, S.; Cowan, B.; Chepurna, I.; Hogue, A. A real-time reconstructed 3D environment augmented with virtual objects rendered with correct occlusion. In Proceedings of the 2014 IEEE Games Media Entertainment Toronto, ON, Canada, 22–24 October 2014; pp. 1–8. [Google Scholar]
  21. Shang, Y.; Yu, Q.; Zhang, X. Analytical method for camera calibration from a single image with four coplanar control lines. Appl. Opt. 2004, 43, 5364–5369. [Google Scholar] [CrossRef]
  22. Cai, Y.; Huang, Y. A Robust Linear Camera Calibration Based on Coplanar Circles. In Proceedings of 2013 Chinese Intelligent Automation Conference: Intelligent Information Processing. Chinese Assoc Automat, Intelligent Automat Comm; Yangzhou University: Yangzhou, China, 2013; Volume 256, pp. 521–529. [Google Scholar] [CrossRef]
  23. Takahashi, A.; Ishii, I.; Makino, H.; Nakashizuka, M. A camera calibration method using parallelogrammatic grid points. IEICE Trans. Inf. Syst. 1996, E79D, 1579–1587. [Google Scholar]
  24. Mozerov, M.; Amato, A.; Al Haj, M.; Gonzalez, J. A Simple Method of Multiple Camera Calibration for the Joint Top View Projection. In Computer Recognition Systems 2; Springer: Berlin, Germany, 2007; Volume 45, pp. 164–170. [Google Scholar]
  25. Becker, S.; Bove, V. Semiautomatic 3-D model extraction from uncalibrated 2-D camera views. Visual data exploration and analysis II. In Proceedings of the Society of Photo-Optical Instrumentation Engineers (SPIE), Orlando, FL, USA, 19–21 April 1995; Volume 2410, pp. 447–461. [Google Scholar] [CrossRef] [Green Version]
  26. Delage, E.; Lee, H.; Ng, A.Y. Automatic single-image 3D reconstructions of indoor Manhattan world scenes. Robot. Res. 2007, 28, 305–321. [Google Scholar]
  27. Wilczkowiak, M.; Boyer, E.; Sturm, P. Camera calibration and 3D reconstruction from single images using parallelepipeds. In Proceedings of the 8th IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; IEEE Computer Society: Washington, DC, USA, 2001; Volume 1, pp. 142–148. [Google Scholar]
  28. Sturm, P.; Maybank, S. A Method for Interactive 3D Reconstruction of Piecewise Planar Objects from Single Images. In Proceedings of the 10th British Machine Vision Conference (BMVC ’99), Nottingham, UK, 13–16 September 1999; pp. 265–274. [Google Scholar]
  29. Micusik, B.; Wildenauer, H.; Kosecka, J. Detection and matching of rectilinear structures. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; IEEE Computer Society: Washington, DC, USA, 2008; pp. 1–7. [Google Scholar] [CrossRef]
  30. Penna, M. Determining camera parameters from the perspective projection of a quadrilateral. Pattern Recognit. 1991, 24, 533–541. [Google Scholar] [CrossRef]
  31. Hong, Z.Q.; Yang, J.Y. An algorithm for camera calibration using a three-dimensional reference point. Pattern Recognit. 1993, 26, 1655–1660. [Google Scholar] [CrossRef]
  32. Wefelscheid, C.; Wekel, T.; Hellwich, O. Monocular Rectangle Reconstruction Based on Direct Linear Transformation. In Proceedings of the VISAPP 2011: Proceedings of the International Conference on Computer Vision Theory and Applications, Vilamoura, Portugal, 5–7 March 2011; Institute for Systems and Technologies of Information, Control and Communication: Setubal, Portutal, 2011; pp. 271–276. [Google Scholar]
  33. Shunliang, P.; Xiaojian, W.; Weiqun, S.; Zishan, S. A faster relative 3D position and attitude algorithm based on special four-point feature. In Signal Analysis, Masurement Theory, Photo-Electronic Technology, and Artificial Intellingence, Pts 1 and 2; Beijing Univ Aeronaut & Astronaut: Beijing, China, 2006; Volume 6357, pp. 1–2. [Google Scholar] [CrossRef]
  34. Guillou, E.; Meneveaux, D.; Maisel, E.; Bouatouch, K. Using vanishing points for camera calibration and coarse 3D reconstruction from a single image. Vis. Comput. 2000, 16, 396–410. [Google Scholar] [CrossRef]
  35. Zhou, K.; Wang, X.J.; Wang, Z.; Wei, H.; Yin, L. Complete Initial Solutions for Iterative Pose Estimation From Planar Objects. IEEE Access 2018, 6, 22257–22266. [Google Scholar] [CrossRef]
  36. YUAN, J. A General Photogrammetric Method for Determining Object Position and Orientation. IEEE Trans. Robot. Autom. 1989, 5, 129–142. [Google Scholar] [CrossRef]
  37. Wang, P.; Xu, G.; Cheng, Y.; Yu, Q. Camera pose estimation from lines: A fast, robust and general method. Mach. Vis. Appl. 2019, 30, 603–614. [Google Scholar] [CrossRef]
  38. Ulrich, M.; Wiedemann, C.; Steger, C. CAD-Based Recognition of 3D Objects in Monocular Images. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation ICRA, Kobe, Japan, 12–17 May 2009; pp. 2090–2097. [Google Scholar]
  39. Sakcak, B.; Bascetta, L.; Ferretti, G. Model based Detection and 3D Localization of Planar Objects for Industrial Setups. In Proceedings of the 13th International Conference on Informatics in Control, Automation and Robotics (ICINCO), Lisbon, Portugal, 29–31 July 2016; Volume 2, pp. 360–367. [Google Scholar] [CrossRef] [Green Version]
  40. Han, P.; Zhao, G. CAD-based 3D objects recognition in monocular images for mobile augmented reality. Comput. Graph. UK 2015, 50, 36–46. [Google Scholar] [CrossRef]
  41. He, Z.; Jiang, Z.; Zhao, X.; Zhang, S.; Wu, C. Sparse Template-Based 6-D Pose Estimation of Metal Parts Using a Monocular Camera. IEEE Trans. Ind. Electron. 2019, 390–401. [Google Scholar] [CrossRef]
  42. Monge, G. Géométrie Descriptive; J. Klostermann Fils: Paris, France, 1847; p. 184. [Google Scholar]
  43. Adler, A.A. The Theory of Engineering Drawing; D. Van Nostrand Company: New York, NY, USA, 1912; p. 362. [Google Scholar]
  44. Estrems, M.; Sánchez, H.; Faura, F. Influence of Fixtures on Dimensional Accuracy in Machining Processes. Int. J. Adv. Manuf. Technol. 2003, 21, 384–390. [Google Scholar] [CrossRef]
Figure 1. Plotting a point (P) in a conical perspective from its dihedral projection. Subindex f refers to the Front View and subindex t refers to the Top View. (a) Point dihedral construction view; (b) Dihedral projections.
Figure 1. Plotting a point (P) in a conical perspective from its dihedral projection. Subindex f refers to the Front View and subindex t refers to the Top View. (a) Point dihedral construction view; (b) Dihedral projections.
Sensors 19 05432 g001
Figure 2. Actual conical perspective of the object froma photograph: (a) Initial image; (b) Referencemodel.
Figure 2. Actual conical perspective of the object froma photograph: (a) Initial image; (b) Referencemodel.
Sensors 19 05432 g002
Figure 3. Obtaining auxiliary points from the four vertices of the rectangle.
Figure 3. Obtaining auxiliary points from the four vertices of the rectangle.
Sensors 19 05432 g003
Figure 4. Line orientation that subdivides the segment of the side in two equal segments.
Figure 4. Line orientation that subdivides the segment of the side in two equal segments.
Sensors 19 05432 g004
Figure 5. Obtaining the Top View from the Front View.
Figure 5. Obtaining the Top View from the Front View.
Sensors 19 05432 g005
Figure 6. Obtaining the Top projection of a point of the conical with a known Front projection.
Figure 6. Obtaining the Top projection of a point of the conical with a known Front projection.
Sensors 19 05432 g006
Figure 7. Summary for obtaining the dihedral from the conical projection (blue: dihedral; black: conical).
Figure 7. Summary for obtaining the dihedral from the conical projection (blue: dihedral; black: conical).
Sensors 19 05432 g007
Figure 8. Obtaining camera positioning with respect to rectangle coordinate system.
Figure 8. Obtaining camera positioning with respect to rectangle coordinate system.
Sensors 19 05432 g008
Figure 9. (a) Set-up used for the measurements of the position of the camera in rectangle coordinates through a CMM; (b) Points of the camera touched by the probe for the implementation of the 3-2-1 method [44].
Figure 9. (a) Set-up used for the measurements of the position of the camera in rectangle coordinates through a CMM; (b) Points of the camera touched by the probe for the implementation of the 3-2-1 method [44].
Sensors 19 05432 g009
Figure 10. Graphical representation of the theoretical and experimental values of the center points of the rectangles. Distance errors have also been included for each point.
Figure 10. Graphical representation of the theoretical and experimental values of the center points of the rectangles. Distance errors have also been included for each point.
Sensors 19 05432 g010
Figure 11. Images analyzed during the experimental test.
Figure 11. Images analyzed during the experimental test.
Sensors 19 05432 g011
Table 1. Obtaining auxiliary points as intersections of lines that go through two points.
Table 1. Obtaining auxiliary points as intersections of lines that go through two points.
Support PointLine 1Line 2
M * P 1 * P 3 * P 2 * P 4 *
V 1 * P 1 * P 2 * P 3 * P 4 *
V 2 * P 1 * P 4 * P 2 * P 3 *
P 12 * P 1 * P 2 * M * V 2 *
P 23 * P 2 * P 3 * M * V 1 *
P 34 * P 3 * P 4 * M * V 2 *
P 14 * P 1 * P 4 * M * V 1 *
Table 2. Parameters used for the test.
Table 2. Parameters used for the test.
Camera ParametersA4 Sheet ParametersFocus Position in Camera Coordinates
Sensor size1/2.3”6.16 × 4.62 mmwidth210mmx30mm
Width4608pixelsheight297mmy30mm
Height3456pixels z3.5mm
Focus3433pix (4.6 mm)
V x 2186pixels
V y 1991pixels
1 pixel0.0013368mm
Table 3. Coordinates, distances, and errors obtained by CMM and image analysis where subindex e refers to experimental data calculated with the CMM and subindex t refers to theoretical values calculated by the image analysis algorithm. Coordinates x, y, z and distances d are expressed in mm. Differences of the coordinates and distances (errors) are presented in percentages.
Table 3. Coordinates, distances, and errors obtained by CMM and image analysis where subindex e refers to experimental data calculated with the CMM and subindex t refers to theoretical values calculated by the image analysis algorithm. Coordinates x, y, z and distances d are expressed in mm. Differences of the coordinates and distances (errors) are presented in percentages.
CMMImage
Image x e y e z e d e x t y t z t d t ϵ x (%) ϵ y (%) ϵ z (%) ϵ d (%)
1201.22222.21367.44474.21210.09200.44365.64466.914.41−9.79−0.49−1.54
2−195.04218.71367.43469.98−180.47215.48374.60468.32−7.47−1.481.95−0.35
388.39469.44378.18609.27111.54473.32380.00617.1526.190.830.481.29
4157.90225.84374.79465.19148.48212.45381.40461.13−5.97−5.931.76−0.87
5−113.73672.02390.02785.27−120.61687.26391.43800.066.062.270.361.88
6128.21190.00366.23432.05102.81185.02370.68426.85−19.82−2.621.21−1.20
7184.55592.80384.91730.50197.23600.11389.99742.386.881.231.321.63
8−80.07460.32374.07598.53−60.11462.76378.33600.75−24.930.531.140.37
9−113.76128.64352.65392.24−89.32120.79354.29384.83−21.48−6.100.47−1.89

Share and Cite

MDPI and ACS Style

Estrems Amestoy, M.; de Francisco Ortiz, Ó. Global Positioning from a Single Image of a Rectangle in Conical Perspective. Sensors 2019, 19, 5432. https://doi.org/10.3390/s19245432

AMA Style

Estrems Amestoy M, de Francisco Ortiz Ó. Global Positioning from a Single Image of a Rectangle in Conical Perspective. Sensors. 2019; 19(24):5432. https://doi.org/10.3390/s19245432

Chicago/Turabian Style

Estrems Amestoy, Manuel, and Óscar de Francisco Ortiz. 2019. "Global Positioning from a Single Image of a Rectangle in Conical Perspective" Sensors 19, no. 24: 5432. https://doi.org/10.3390/s19245432

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop