Non-rigid vehicle-borne LiDAR-assisted aerotriangulation

Li Zheng, Yuhao Li, Meng Sun, Zheng Ji, Manzhu Yu, Qingbo Shu

Research output: Contribution to journalArticlepeer-review

3 Scopus citations


VLS (Vehicle-borne Laser Scanning) can easily scan the road surface in the close range with high density. UAV (Unmanned Aerial Vehicle) can capture a wider range of ground images. Due to the complementary features of platforms of VLS and UAV, combining the two methods becomes a more effective method of data acquisition. In this paper, a non-rigid method for the aerotriangulation of UAV images assisted by a vehicle-borne light detection and ranging (LiDAR) point cloud is proposed, which greatly reduces the number of control points and improves the automation. We convert the LiDAR point cloud-assisted aerotriangulation into a registration problem between two point clouds, which does not require complicated feature extraction and match between point cloud and images. Compared with the iterative closest point (ICP) algorithm, this method can address the non-rigid image distortion with a more rigorous adjustment model and a higher accuracy of aerotriangulation. The experimental results show that the constraint of the LiDAR point cloud ensures the high accuracy of the aerotriangulation, even in the absence of control points. The root-mean-square error (RMSE) of the checkpoints on the x, y, and z axes are 0.118 m, 0.163 m, and 0.084m, respectively, which verifies the reliability of the proposed method. As a necessary condition for joint mapping, the research based on VLS and UAV images in uncontrolled circumstances will greatly improve the efficiency of joint mapping and reduce its cost.

Original languageEnglish (US)
Article number1188
JournalRemote Sensing
Issue number10
StatePublished - May 1 2019

All Science Journal Classification (ASJC) codes

  • General Earth and Planetary Sciences


Dive into the research topics of 'Non-rigid vehicle-borne LiDAR-assisted aerotriangulation'. Together they form a unique fingerprint.

Cite this