Watertight Scenes from Urban LiDAR and Planar Surfaces

The demand for large geometric models is increasing, especially of urban environments. This has resulted in production of massive point cloud data from images or LiDAR. Visualization and further processing generally require a detailed, yet concise representation of the scene's surfaces. Related...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Computer graphics forum Jg. 32; H. 5; S. 217 - 228
Hauptverfasser: van Kreveld, M., van Lankveld, T., Veltkamp, R. C.
Format: Journal Article
Sprache:Englisch
Veröffentlicht: Oxford, UK Blackwell Publishing Ltd 01.08.2013
Wiley
Schlagworte:
ISSN:0167-7055, 1467-8659
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The demand for large geometric models is increasing, especially of urban environments. This has resulted in production of massive point cloud data from images or LiDAR. Visualization and further processing generally require a detailed, yet concise representation of the scene's surfaces. Related work generally either approximates the data with the risk of over‐smoothing, or interpolates the data with excessive detail. Many surfaces in urban scenes can be modeled more concisely by planar approximations. We present a method that combines these polygons into a watertight model. The polygon‐based shape is closed with free‐form meshes based on visibility information. To achieve this, we divide 3‐space into inside and outside volumes by combining a constrained Delaunay tetrahedralization with a graph‐cut. We compare our method with related work on several large urban LiDAR data sets. We construct similar shapes with a third fewer triangles to model the scenes. Additionally, our results are more visually pleasing and closer to a human modeler's description of urban scenes using simple boxes.
Bibliographie:ark:/67375/WNG-TWGT6HBM-W
istex:681680EEE5B77A2A372E1598F59140FD0FA3CC28
ArticleID:CGF12188
SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 14
ObjectType-Article-2
content type line 23
ISSN:0167-7055
1467-8659
DOI:10.1111/cgf.12188