The attitude parameter is an important state parameter for a long axisymmetric target. The plane intersection (PI) method is a commonly used method for attitude estimation. However, this method only uses planes’ information in the object space under a multiple camera system (more than two cameras simultaneously observing a target). We propose two methods to address the aforementioned issue. One method involves minimizing the square of the object-space angle residual (OAR) and the other method involves minimizing the square of the image-space angle residual (IAR). The linear optimization method is used for the above minimizing problems. The simulation results demonstrate that the IAR method has higher accuracy than the PI and OAR methods under multiple and dual camera systems because it incorporates information of a pair of corresponding image points. Furthermore, our experiments have shown that the linear method is generally faster, and it has an equivalent accuracy compared to the iterative method. |
1.IntroductionThe axis attitude is an important parameter for the motion state analysis, performance measurement, and aerodynamic parameter identification of rockets or missiles. Optical measurement is the main measurement method of a shooting range due to its noncontact, passive, high frequencies and high accuracy characteristics. Generally, a shooting range uses theodolites to record the target launching and running process, followed by image processing to obtain the target yaw angle and the pitch angle. The pose measurement of the target mainly has the following types of methods: The first method is to use plane intersection (PI) to obtain the target’s central axis attitude,1 This method assumes that the target is a symmetrical object, and two cameras are used to simultaneously record the target. The central axis of the target can then be extracted from the image. Based on the principle of light propagation along a straight line, the target axis must be located on a plane consisting of the optical center and the target central axis in the image. The intersection line of the planes of the two cameras can be calculated to obtain the target axis representation. This method gives more accurate measurements than the method based on endpoints, and the form is intuitive. This method also shows that the longer the target axis in an image, the higher the pose measurement accuracy. Many researchers have extended this method to other applications.2–4 The accuracy of target axis extraction in an image plays an important role in improving the accuracy of the PI method.5 This kind of method can obtain only the target’s central axis attitude, not the target’s position or the roll angle. The second method is to use the information of the target’s structure and then match its projection on the image with the contour extracted from the image.6–16 Zhao et al.7 proposed a pose estimation algorithm based on the inner angle and the triangle constraint. Liu and Hu8 gave a pose estimation method based on a monocular image for a rotating solid shape spacecraft that utilized the inherent characteristics of a spacecraft. However, this method was not applicable for other types of targets. Peng et al.9,10 proposed some methods for non-cooperative space targets that were based on only stereo cameras or fusion with laser radar. Becke and Schlegl11 proposed a method for estimating a target axis based on the contours under monocular or multi-image conditions. Hanek et al.15 used the linear extreme contours of points, lines, and cylinders to estimate the position and orientation of a camera in the world coordinate system. The third method is to use some cooperation signs (points or lines) on the target. First, we achieve the cooperation signs’ coordinate in the target’s coordinate system, and then, we solve the pose from the target to the camera’s coordinate system.17–26 Horn et al.20 utilized the point cloud of the side surface, which achieved by a laser profile sensor to measure the pose of cylindrical components, but this method cannot work at high frequencies. Some researchers estimated camera’s pose by the correspondence of lines or epipolar constraint.21–25 Zhou et al.26 proposed a method to track and estimate the pose of known rigid objects efficiently in a complex environment, which based on a 3D particle filter with M-estimation optimization. The fourth method is to directly achieve the point cloud by a laser profile sensor and then use the point cloud to estimate the target’s attitude.23 For long cylindrical targets, the second method cannot achieve a high precision while the cameras are at a long distance of more than 100 m. For the third method, it is difficult to obtain the coordinates of the cooperation signs in the target coordinate system with high precision. There are two main methods for this. One method is based on model labeling, and the other method involves measurement with a total station after the mark is attached. With the first method, it is difficult to guarantee the accuracy of labeling when the target is large. In the total station measurement process, since it is necessary to move multiple positions to measure all of the feature points, it is easy to introduce conversion errors and affect the measurement accuracy. The fourth method cannot work at high frequencies. Later, in this paper, we will describe the accuracy advantage of the axis method relative to the feature points-based method in the measurement of the axis attitude for long symmetric targets. So, in this paper, we propose a method to improve the PI’s accuracy. For a multiple camera system (more than two cameras simultaneously observing a target), the planes given by the cameras from the optical center and the target axis in the image cannot intersect at the same straight line. This is caused by the inevitable errors of the camera parameter and the central axis. The solution for the PI method is to derive the direction vectors of the target using all of the plane equations. Then, the target attitude angle can be derived according to the definition of decomposition. However, the PI method lacks a geometric explanation when there are more than two cameras. In this paper, we propose two optimization methods to address this issue. The first method is based on minimizing the object-space angle residual (OAR), and the other method is based on minimizing the image-space angle residual (IAR). We also provide linear solutions and iterative solutions for the OAR and the IAR, called the OARL, IARL, OARI, and IARI, respectively. The simulation results showed the IAR obtained a higher accuracy than the PI and the OAR. Furthermore, the IARL was generally faster than the IARI. 2.Plane Intersection Method2.1.Target Attitude RepresentationFor an axisymmetric long strip target, the pose of the target is usually determined by its symmetry axis direction, which is generally represented by the direction vector or the angles. Alternatively, the yaw and pitch angles could be used to represent the direction vector, as shown in Fig. 1. For the shooting range, faces north, faces upward, and faces east. OA is the target axis, OA’ is the projection of OA on the XOZ plane, the angle between the axis and OA’() is the yaw angle, and the angle between OA and OA’() is the pitch angle. 2.2.Coordinate System DefinitionThe world coordinate system is a right-handed coordinate system, as shown in Fig. 2. The origin of the camera coordinate system is the optical center. The axis is pointing in the positive direction of the optical axis, the axis is perpendicular to the axis horizontally to the right, and the axis is perpendicular to the axis and to the axis, whose direction is determined by the right-handed definition criterion. The origin of the image coordinate system is the upper left corner. 2.3.Plane Representation by a Single CameraAs shown in Fig. 2, AB is the target axis, C1 and C2 are the optical centers of the two cameras, A1B1 is the projection of the target axis on camera C1, and A2B2 is the projection of the target axis on camera C2. Assume that the equation of the target central axis in the ’th image coordinate system is where and . The image plane can be seen as the plane in the camera coordinate system, where is the physical focal length of the camera. The axis of the camera coordinate system passes through the principal point of the image plane , so the transformation from the image coordinate system to the camera coordinate system is determined as follows: where is in the camera coordinate system, is in the image coordinate system, and is the actual physical size of the camera pixel.The central axis of the target in the image can be expressed in the camera coordinate system as follows: The plane equation (composed of the axis and the optical center of the camera) can be defined as follows: The equivalent focal length of the camera is abbreviated as follows: The plane equation can be summarized as follows: Given the parameter matrix , the plane equation is then rewritten as follows: where is a point from the camera coordinate system and is the normal vector of the plane. It is assumed that the external parameters of the ’th camera in the world coordinate system are , , where and are the rotation matrix and the translation vector from the world coordinate system to the camera coordinate system. Therefore, the point in the world coordinate system can be transformed to the camera coordinate system using Eq. (8): where is a point in the world coordinate system. The plane equation in the world coordinate system can then be derived as follows:So far, the expression of the plane (composed of optical center of the camera and the central axis of the target in the image) has been obtained in the world coordinate system. 2.4.Principle of Dual Camera Plane IntersectionSupposing that the plane equations of the two cameras are where and are the normal vectors of the two planes. We do not consider the atmospheric refraction in this paper, so the direction vector of the intersection line must be perpendicular to both normal vectors, it can be solved asThe yaw angle and pitch angle of the target can be determined as follows: It can be seen that the representation of the direction vector obtained by the final intersection of this method is independent of and only for the plane parameter that is related to the translation vector of the camera. This means that the final direction vector is independent of the translation vector . It can be concluded that the target position is related only to the attitude of the camera. 3.Our Method3.1.Multicamera Intersection ScenarioAs shown in Fig. 2, the three cameras C1, C2, and C3 observe the same target AB from different directions, and the projections on the image are A1B1, A2B2, and A3B3, respectively. Theoretically, the three planes C1A1B1, C2A2B2, and C3A3B3 should intersect in a straight line. However, due to the errors of the camera parameters and the axis, the direction vectors given by each set of two cameras are not completely consistent. Therefore, it is important to derive the optimal direction vector. 3.2.Optimization ConditionsThis paper proposes two optimization methods for the axis measurement problem. These two methods are based on minimizing residuals:
3.3.Object-Space Angle Residual3.3.1.Definition of the optimization problemSupposing that multiple cameras simultaneously observe a target, one can extract the target axis in the image. The plane equations of the camera’s optical center and the target axis in the image are expressed in the world coordinate system as follows: Supposing that the target axis direction is , then the angle between the central axis and the plane is For the convenience of calculation, the linear direction and the plane parameters are normalized, i.e., Then, can be expressed as follows: From the geometric understanding of the PI, one can use this angle as the residual of the obtained central axis relative to each camera plane. In a similar manner to the spatial residual for the camera calibration, minimizing this residual can provide the optimal solution: where is the number of cameras. For a small angle, ;could be approximated by , therefore, Eq. (17) could be rewritten as follows: where is within the interval of .3.3.2.Linear methodThe Levenberg–Marquardt22 can be used to solve Eq. (17) and keep low computation efficiency. The following method provides a linear solution to the minimum value by deriving Since is a function of , letting the partial derivative be 0: We can solve the equations by the singular value decomposition (SVD)27 of the coefficient matrix. Theorem 1:It is assumed that are the coefficients of the spatial plane . The necessary and sufficient condition for of the matrix in Eq. (21) is that at least two of the planes are not parallel, and all of the planes are parallel to the same straight line. The coefficient matrix of Eq. (20) can be expressed as the form of Eq. (21): Proof of the theorem’s sufficiency: Supposing two planes p1 and p2 are not parallel, and are not all equal, then the matrix has at least one second-order subdetermination, that is, not 0, i.e., the rank is equal to 2. One can then conclude that the rank of is larger than or equal to 2. Since all planes are parallel to the same straight line, all planes can be made across the same straight line by changing . Then, all planes form a planar cluster. It can be known from the properties of planar clusters that other planes can be linearly represented by p1 and p2: It can be seen that the rank of is 2. From ,28 the rank of the formula (21) is 2. The sufficiency certificate is completed. Next, the necessity of the theorem is proven. It is known from that the rank of is 2. This means that at least two planes are not parallel. Without loss of generality, p1 and p2 are chosen for the following analysis. All other planes can be represented linearly by p1 and p2, i.e., Since affects only the plane position, it does not affect the plane direction, so all planes are parallel to the straight line represented by Eq. (24): The necessity certificate is completed. According to the above Theorem 1, if there are only two cameras (not parallel), then the equation has a unique nonzero solution, and the geometric meaning of the solution is consistent with the PI method. If there are more than two cameras, unless the condition of Theorem 1 is satisfied, the system of equations does not have a strict nonzero solution, and only the least-squares solution can be obtained as the optimal solution. 3.3.3.Efficiency analysisThe linear method for minimizing the object angle residual is as follows.
It can be seen from the above analysis that the method can be solved only by calculating the coefficient matrix and solving equations with the least square method. Therefore, its time complexity is . 3.4.Image-Space Angle Residual3.4.1.Definition of the optimization problemIt is assumed that the final central axis direction is and the axis passes through a point in the world coordinate system. It is easy to see that the straight line passes through the point . In the world coordinate system, can be obtained from the intersection of the pair of corresponding image point on the central axis. The plane consisting of the straight line and the optical center of the ’th camera can be expressed as follows: where is the optical center of the ’th camera. The plane equation after simplification is where is the normal vector of the plane and the subscript means that it is from the world coordinate system. It can be seen from Eq. (25) that can be represented linearly by . Letting the relationship matrix be M, i.e.,It is then easy to prove that M is a full-rank matrix. The transformation matrix from the world coordinate system to the camera coordinate system is , which is also a full-rank matrix. The plane is expressed in the camera coordinate system as where is the normal vector of the plane and the subscript means that it is from the camera coordinate system. From the image plane equation , the direction of the intersection line is . Then, the angle between the axis of the target in the image plane and the intersection line is where and are coefficients of the linear equation in the image plane.Therefore, the optimized objective function is where is the number of cameras.3.4.2.Linear methodA method of linearly solving for the minimum value is given by derivation below. The original expression (30) is complicated, especially the partial derivative expression. The numerator and the denominator contain , which cannot be solved linearly. Therefore, the expression is deformed. If the two 2D lines , are nearly parallel, the angle is a small angle and we have . For the sake of discussion, we divide and by the larger one, respectively, so that the maximum value is 1. Then, , so is positively correlated with . is the absolute value of . Then, Eq. (31) can be used instead of Eq. (28) as the optimization solution objective function: From the previous derivation, is a linear function of . Then, the equation can be recorded as where are the coefficients of in Eq. (31) and we can think of it as representing a normal vector to a plane too. This can be thought of as a function of . Clearly, this function is a basic elementary function, so it is continuous and differentiable in the domain of the definition, and the minimum value must exist. At the point where the minimum value is obtained, the partial derivative of exists and it is 0, and the partial derivative is obtained separately: We can solve the equations by the SVD27 decomposition of the coefficient matrix. It can be understood from Theorem 1 that only two planes represented by need to be nonparallel and all planes are parallel to the same line. The rank of this matrix is 2, which indicates that the system of equations has a unique nonzero solution. Since all planes are parallel to the straight line , , and only the optical center of the two cameras and the target line is not coplanar, then the rank of the matrix of Eq. (32) is 2, and the direction solved for by the image angle residual can be used to obtain the global minimum. 3.4.3.Efficiency analysisThe linear method for minimizing the image angle residual is as follows.
It can be seen from the above analysis that the image angle residual method can be solved for only by calculating the 3D point coordinate, the projection line parameter, the solution coefficient matrix, and the least squares solution, so the time complexity is , which is the same as that of the OARL method. 4.ExperimentsBecause the ground truth was not easy to obtain in the real experiment, we used only the real data to verify the validity of the algorithm, and then we used the simulation to verify the accuracy of the algorithm. 4.1.Simulation4.1.1.Simulation environmentThe simulation platform was Windows 7, and the processor was an Intel(R) Core TM i7-6820HQ 2.7 GHz. Simulation conditions: The equivalent focal length was (2181.8, 2181.8) and the image main point was (1023.5, 1023.5). The cameras were arranged in a circle around the target, and the radius of the circle was 4.5 m. The origin of the world coordinate system was the center of the circle. The two endpoints of the target axis were and (1.5, 1.5, 1.5). The top view of simulation scenario is shown in Fig. 5. 4.1.2.Simulation description
Simulations were conducted for the following conditions, separately:
4.1.3.Simulation resultsThe different simulations and the related results are given below. Simulation 1: A Gaussian error with a mean of 0 was added to the optical center of the camera, with a 10 mm RMS value. A Gaussian error with a mean of 0 was added to the camera angle, with a 0.5 deg RMS value. A Gaussian error with a mean of 0 was added to the and directions of the two endpoints of the axis with a 1-pixel RMS value. To verify the solution of this method in the case of only two cameras, the number of cameras ranged from two to nine. The cameras were arranged on a circle with a radius of 4.5 m, the center of the circle was the origin of the world coordinate system. The angle was calculated according to the average number of cameras. The corresponding image points of the IAR were obtained through the intersection of the head points with the error added. The simulation results for the different numbers of cameras are as follows.
Table 1Relationship between angle error and number of cameras.
Note: The bold means the best results. In the subsequent simulation results, the linear and iterative results were essentially the same. For the convenience of comparison, only the OARL and IARL are discussed for the following experiment (except the running time comparison in simulation 7). Simulation 2: A Gaussian error with a mean of 0 was added to the and directions at the head and the tail of the axis with 0 to 4 pixels of RMS. We fixed the number of cameras at . The simulation results are shown in Fig. 6. Simulation 3: A Gaussian error with a mean of 0 was added to the three angles of the external parameter with 0.1 to 1 deg of RMS. The step was 0.05 deg, and the number of cameras was 5. The simulation results are shown in Fig. 7. Simulation 4: A Gaussian error with a mean of 0 was added to the three values of the camera’s optical center, with 1 to 50 mm of RMS. The step size was 5 mm and the number of cameras was 5. The simulation results are shown in Fig. 8. Simulation 5: A Gaussian error with a mean of 0 was added to the optical center of the camera, with 1 to 50 mm of RMS, and the step size was 1 mm. A Gaussian error with a mean of 0 was added to the camera angle, with 0.05 to 2.5 deg of RMS, and the number of cameras was 5. The simulation results are shown in Fig. 9. Simulation 6: A Gaussian error with a mean of 0 was added to the optical center of the camera, with 1 to 50 mm of RMS, and the step size was 1 mm. A Gaussian error with a mean of 0 was added to the camera angle, with 0.05 to 2.5 deg of RMS. A Gaussian error with a mean of 0 was added to the and directions of the two endpoints of the axis, with 0.1 to 5 pixels of RMS, and the number of cameras was 5. The simulation results are shown in Fig. 10. Simulation 7: The same error was added as in case 1. The number of cameras was increased from 3 to 100. The simulation was run 10,000 times. The results are shown in Fig. 11. 4.2.Experiment VerificationNext, we will demonstrate the PI’s advantage for the measurement of a long axisymmetric target with physical experiment 1 and demonstrate the correctness of the method proposed in this paper with physical experiment 2. 4.2.1.Physical experiment 1Physical environmentTo verify the accurate performance of the central axis intersection algorithm in the axial attitude measurement of the symmetric target, the design experiment was verified and the four commonly used methods were tested, namely The test scenario was shown in Fig. 12. The target was a cylindrical barrel. The points on the barrel were used as marker points on the target for pose calculation. The points on the wall and on the pillar were calibration control points. The optical axes of the two cameras all faced the barrel, with the center of the barrel as the intersection point. The angle was , and the optical center was about 1.5 m from the barrel. The target filled the width of the field of view as much as possible. A column of points was arranged in the direction of the camera and for 45 deg on both sides. The main purpose of this was to simultaneously improve the coverage area of the marker points and ensure the extraction accuracy of the points as much as possible. At the same time, some points were arranged in the common area of the two cameras for method 3. Some cooperative signs were also arranged on the walls around the cameras to convert the control points and marking points to the same coordinate system when the total station frame was in two different positions. The auxiliary measuring equipment consisted of a Leica TS30 and a Leica TM5100A. The Leica TS30 was used to obtain the coordinates of the control points for calibration and the coordinates of the target’s points for posture calculation. Its accuracy was 0.6 mm. The Leica TM5100A was used to measure the change in the angle of the barrel. Its accuracy was 0.5 arc sec, and the accuracy of the vertical lifting platform to which the Leica TM5100A was attached was 2 arc sec. Experimental procedureThe barrel was attached to a platform that could precisely control the direction of rotation, and a plane mirror was fixed on the barrel as the measuring reference of the Leica TM5100A. The measurement steps were as follows:
Table 2True values and algorithm results (unit: degrees).
Note: The bold means the best results. The RMS of error is , where is the value got by algorithm and is the truth value got by the total station. It is evident from Table 2 that the RMS of attitude error measured in this experimental scenario from low to high was PI, GOIAMCS, DMPI + AO, and OPNP + OI. Therefore, it can be proven that the PI method had advantages in the axis attitude measurement of a long symmetric target. 4.2.2.Physical experiment 2To verify the correctness of the algorithm, the method was also verified by physical experiments. Physical environmentThe physical environment for physical experiment 2 was shown in Fig. 13. The target was placed on the board in the center and the diagonal markers on the side were the calibration control points. The camera’s pixel size was , and the equivalent focal distance was 5200. Experimental procedureThe experiment was conducted as follows:
Table 3Experimental results (unit: degrees).
5.ConclusionThis paper proposed two methods, which were based on minimizing the residuals of the object angle and image angle. These two methods can provide geometric explanations that are lacking in the PI method. The simulation demonstrated that in the case of only two cameras, the results of the OAR method were consistent with those of the PI method, and the IAR method achieved higher accuracy than the PI and OAR method using corresponding image point information. It was not difficult to extract a pair of corresponding image points on the target axis in the actual task, so the method had important practical significance for the case of only two cameras. For a multiple camera system, the accuracies of the OAR and the PI were the same, and the accuracy of the IAR was higher than the accuracies of the OAR and the PI. Furthermore, the linear method was compared with the iterative method. We proved that the IARL was generally faster than the IARI. The IARL had great significance in the actual task measurement scenario. However, in the same manner as the PI, all of these proposed methods need the target’s center axis for measurement, so they are suitable only for the measurement of an axisymmetric target. AcknowledgmentsThe research was supported by the National Natural Science Foundation of China (Grant Nos. 11727804, 11872070, and 11802321). ReferencesQ. F. Yu, X. Y. Sun and G. J. Chen,
“A new method of measure the pitching and yaw of the axes symmetry object through the optical image,”
J. Natl. Univ. Def. Technol., 22
(2), 15
–19
(2000). Google Scholar
Y. Zhang et al.,
“Attitude measurement method research for missile launch,”
China Opt., 8
(6), 997
–1003
(2015). https://doi.org/10.3788/CO.20150806.0997 Google Scholar
K. Luo et al.,
“Measuring technology on elevation angle and yawing angle of space target based on optical measurement method,”
J. Changchun Univ. Sci. Technol.,
(2007). Google Scholar
C. Wang, Study on Methods to Improve the Attitude Measurement Precision of the Rotary Object,
(2013). Google Scholar
S. W. Ding, Research on Axis Extraction Methods for Target’s Attitude Measurement,
(2013). Google Scholar
L. Liu and Z. Zhao,
“A new approach for measurement of pitch, roll and yaw angles based on a circular feature,”
Trans. Inst. Meas. Control, 35
(3), 384
–397
(2013). https://doi.org/10.1177/0142331212451991 TICODG 0142-3312 Google Scholar
R. Zhao et al.,
“Pose estimation based on the constraints of inner angles and areas of triangles,”
Proc. SPIE, 7384 738411
(2009). https://doi.org/10.1117/12.835128 PSISDG 0277-786X Google Scholar
C. Liu and W. Hu,
“Relative pose estimation for cylinder-shaped space-crafts using single image,”
IEEE Trans. Aerosp. Electron. Syst., 50
(4), 3036
–3056
(2014). https://doi.org/10.1109/TAES.2014.120757 IEARAX 0018-9251 Google Scholar
J. Peng, W. Xu and H. Yuan,
“An efficient pose measurement method of a space non-cooperative target based on stereo vision,”
IEEE Access, 5 22344
–22362
(2017). https://doi.org/10.1109/ACCESS.2017.2759798 Google Scholar
J. Peng et al.,
“Pose measurement and motion estimation of space non-cooperative targets based on laser radar and stereo-vision fusion,”
IEEE Sens. J., 19
(8), 3008
–3019
(2018). https://doi.org/10.1109/JSEN.2018.2889469 ISJEAZ 1530-437X Google Scholar
M. Becke and T. Schlegl,
“Least squares pose estimation of cylinder axes from multiple views using contour line features,”
in IECON 2015-Conf. IEEE Ind. Electron. Soc.,
001855
–001861
(2015). https://doi.org/10.1109/IECON.2015.7392371 Google Scholar
J. B. Huang, Z. Chen and T. L. Chia, Pose Determination of a Cylinder Using Reprojection Transformation, Elsevier Science Inc.(1996). Google Scholar
C. Doignon and M. De Mathelin,
“A degenerate conic-based method for a direct fitting and 3-D pose of cylinders with a single perspective view,”
in IEEE Int. Conf. Rob. Autom.,
4220
–4225
(2007). https://doi.org/10.1109/ROBOT.2007.364128 Google Scholar
A. Y. C. Shiu and H. Zhuang,
“Pose determination of cylinders, cones, and spheres from perspective projections,”
Proc. SPIE, 1708 771
–782
(1992). https://doi.org/10.1117/12.58624 PSISDG 0277-786X Google Scholar
R. Hanek et al.,
“Yet another method for pose estimation: A probabilistic approach using points, lines, and cylinders,”
in IEEE Comput. Soc. Conf. Comput. Vision and Pattern Recognit.,
550
(1999). https://doi.org/10.1109/CVPR.1999.784961 Google Scholar
Z. Zhao and G. Wen,
“Model-based estimation of axisymmetric target’s pose and speed using line-scan camera,”
in Int. Conf. Syst. Sci. Eng. Des. and Manuf. Inf.,
31
–34
(2011). Google Scholar
Y. Zheng et al.,
“Revisiting the PnP problem: a fast, general and optimal solution,”
in ICCV 2013,
2344
–2351
(2013). Google Scholar
C.-P. Lu, G. D. Hager and E. Mjolsness,
“Fast and globally convergent pose estimation from video images,”
IEEE Trans. Pattern Anal. Mach. Intell., 22
(6), 610
–622
(2000). https://doi.org/10.1109/34.862199 ITPIDJ 0162-8828 Google Scholar
Y. X. Xu, Y. L. Jiang and C. Fang,
“Generalized orthogonal iterative algorithm for pose estimation of multiple camera systems,”
Acta Opt. Sin., 29
(1), 72
–77
(2009). https://doi.org/10.3788/AOS GUXUDC 0253-2239 Google Scholar
B. K. P. Horn, H. M. Hildren and S. Negahdaripour,
“Closed-form solution of absolute orientation using orthonormal matrices,”
Int. J. Opt. Soc. Am. A, 5 1127
–1135
(1988). https://doi.org/10.1364/JOSAA.5.001127 Google Scholar
Y. Q. Zhang et al.,
“Pose optimization based on integral of the distance between line segments,”
Sci. China Technol. Sci., 59
(1), 135
–148
(2016). https://doi.org/10.1007/s11431-015-5958-1 Google Scholar
J. Heller, M. Havlena and T. Pajdla,
“Globally optimal hand-eye calibration using branch-and-bound,”
IEEE Trans. Pattern Anal. Mach. Intell., 38
(5), 1027
–1033
(2015). https://doi.org/10.1109/TPAMI.2015.2469299 ITPIDJ 0162-8828 Google Scholar
Z. Q. Zhang, L. Zhang and G. Z. Yang,
“A computationally efficient method for hand-eye calibration,”
Int. J. Comput. Assist. Radiol. Surg., 12
(10), 1775
–1787
(2017). https://doi.org/10.1007/s11548-017-1646-x Google Scholar
J. J. Lee and M. H. Jeong,
“Stereo camera head-eye calibration based on minimum variance approach using surface normal vectors,”
Sensors, 18
(11), 3706
(2018). https://doi.org/10.3390/s18113706 SNSRES 0746-9462 Google Scholar
Y. Zhang, Z. C. Qiu and X. M. Zhang,
“Calibration method for hand-eye system with rotation and translation couplings,”
Appl. Opt., 58
(20), 5375
–5387
(2019). https://doi.org/10.1364/AO.58.005375 APOPAI 0003-6935 Google Scholar
L. M. Zhou et al.,
“3D rigid pose tracking based on new distance function of line segments,”
Proc. SPIE, 11179 111794L
(2019). https://doi.org/10.1117/12.2540195 PSISDG 0277-786X Google Scholar
G. H. Golub,
“Singular value decomposition and least squares solutions,”
Numer. Math., 14
(5), 403
–420
(1970). Google Scholar
Q. Wang et al., The Sixth Edition of Linear Algebra, Department of Mathematics of Tongji University(2014). Google Scholar
BiographyLijun Zhong received his BE degree in computer science and technology from Central South University, Changsha, China, in 2008 and his ME degree in software engineering from Central South University, Changsha, China, in 2008. He is currently pursuing his PhD with the College of Aerospace Science and Engineering, National University of Defense Technology, Changsha, China. His research interests include image processing, computer vision, deep learning, and photogrammetry. Zhang Li is an assistant professor at the College of Aerospace Science and Engineering, National University of Defense Technology, China. He received his PhD in biomedical engineering from the Delft University of Technology in 2015. He authored more than 20 papers in high-ranking journals and conferences (e.g., TMI and TBME). His research interests include computer vision, particularly image registration. He is also interested in the applications of deep learning. Xiaohu Zhang received his PhD in aerospace science and technology from the National University of Defense Technology, Changsha, China, in 2006. He is a professor with the School of Aeronautics and Astronautics, Sun Yat-Sen University. His research interests include image processing, computer vision, space situational awareness, and photogrammetry. Yang Shang received his PhD in aerospace science and technology from the National University of Defense Technology, Changsha, China, in 2006. He is a professor with the College of Aerospace Science and Engineering, National University of Defense Technology, Changsha, China. His research interests include image processing, computer vision, vision navigation, and photogrammetry. Qifeng Yu received his PhD in precision optical measurement from the University of Bremen, Germany, in 1996. He is a professor with the College of Aerospace Science and Engineering, National University of Defense Technology, Changsha, China. He is also an academician of the Chinese Academy of Sciences. His research interests include image processing, computer vision, vision navigation, and photogrammetry. |