ArticlePDF Available

Abstract and Figures

Three-dimensional (3D) building models play an important role in digital cities and have numerous potential applications in environmental studies. In recent years, the photogrammetric point clouds obtained by aerial oblique images have become a major source of data for 3D building reconstruction. Aiming at reconstructing a 3D building model at Level of Detail (LoD) 2 and even LoD3 with preferred geometry accuracy and affordable computation expense, in this paper, we propose a novel method for the efficient reconstruction of building models from the photogrammetric point clouds which combines the rule-based and the hypothesis-based method using a two-stage topological recovery process. Given the point clouds of a single building, planar primitives and their corresponding boundaries are extracted and regularized to obtain abstracted building counters. In the first stage, we take advantage of the regularity and adjacency of the building counters to recover parts of the topological relationships between different primitives. Three constraints, namely pairwise constraint, triplet constraint, and nearby constraint, are utilized to form an initial reconstruction with candidate faces in ambiguous areas. In the second stage, the topologies in ambiguous areas are removed and reconstructed by solving an integer linear optimization problem based on the initial constraints while considering data fitting degree. Experiments using real datasets reveal that compared with state-of-the-art methods, the proposed method can efficiently reconstruct 3D building models in seconds with the geometry accuracy in decimeter level.
This content is subject to copyright.
remote sensing
Article
Combined Rule-Based and Hypothesis-Based Method for
Building Model Reconstruction from Photogrammetric
Point Clouds
Linfu Xie 1,2,3, Han Hu 4, Qing Zhu 4, Xiaoming Li 1, Shengjun Tang 1, You Li 1, Renzhong Guo 1, Yeting Zhang 2
and Weixi Wang 1,*


Citation: Xie, L.; Hu, H.; Zhu, Q.; Li,
X.; Tang, S.; Li, Y.; Guo, R.; Zhang, Y.;
Wang, W. Combined Rule-Based and
Hypothesis-Based Method for
Building Model Reconstruction from
Photogrammetric Point Clouds.
Remote Sens. 2021,13, 1107. https://
doi.org/10.3390/rs13061107
Academic Editor: Ben Gorte
Received: 19 January 2021
Accepted: 11 March 2021
Published: 14 March 2021
Publisher’s Note: MDPI stays neutral
with regard to jurisdictional claims in
published maps and institutional affil-
iations.
Copyright: © 2021 by the authors.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
1Research Institute for Smart Cities, School of Architecture and Urban Planning, Shenzhen University & Key
Laboratory of Urban Land Resources Monitoring and Simulation, MNR & Guangdong Key Laboratory of
Urban Informatics & Shenzhen Key Laboratory of Spatial Smart Sensing and Services,
Shenzhen 518060, China; linfuxie@szu.edu.cn (L.X.); lixming@szu.edu.cn (X.L.);
shengjuntang@szu.edu.cn (S.T.); liyou@szu.edu.cn (Y.L.); guorz@szu.edu.cn (R.G.)
2State Key Laboratory of Information Engineering in Surveying Mapping and Remote Sensing,
Wuhan University, Wuhan 430079, China; zhangyeting@263.net
3Department of Land Surveying & Geo-Informatics, The Hong Kong Polytechnic University, Hung Hom,
Kowloon 999077, Hong Kong
4Faculty of Geosciences and Environmental Engineering, Southwest Jiaotong University,
Chengdu 611756, China; han.hu@swjtu.edu.cn (H.H.); zhuq66@263.net (Q.Z.)
*Correspondence: wangwx@szu.edu.cn; Tel.: +86-13590317811
Abstract:
Three-dimensional (3D) building models play an important role in digital cities and have
numerous potential applications in environmental studies. In recent years, the photogrammetric
point clouds obtained by aerial oblique images have become a major source of data for 3D building
reconstruction. Aiming at reconstructing a 3D building model at Level of Detail (LoD) 2 and even
LoD3 with preferred geometry accuracy and affordable computation expense, in this paper, we
propose a novel method for the efficient reconstruction of building models from the photogrammetric
point clouds which combines the rule-based and the hypothesis-based method using a two-stage
topological recovery process. Given the point clouds of a single building, planar primitives and their
corresponding boundaries are extracted and regularized to obtain abstracted building counters. In
the first stage, we take advantage of the regularity and adjacency of the building counters to recover
parts of the topological relationships between different primitives. Three constraints, namely pairwise
constraint, triplet constraint, and nearby constraint, are utilized to form an initial reconstruction
with candidate faces in ambiguous areas. In the second stage, the topologies in ambiguous areas
are removed and reconstructed by solving an integer linear optimization problem based on the
initial constraints while considering data fitting degree. Experiments using real datasets reveal that
compared with state-of-the-art methods, the proposed method can efficiently reconstruct 3D building
models in seconds with the geometry accuracy in decimeter level.
Keywords: building models; 3D reconstruction; point clouds; photogrammetry
1. Introduction
Three-dimensional (3D) building models play an important role in constructing digital
cities and have numerous environmental applications in areas such as urban planning [
1
],
smart city [
2
], environmental analysis [
3
], and other civil engineering [
4
]. In recent years,
with the rapid development of aerial vehicles, cameras, and image-processing technologies,
aerial oblique images have become a major data source for 3D city modeling [
5
,
6
]. Due to
the time-consuming and labor-intensive nature of manual modeling processes, researchers
in the photogrammetry, computer vision, and graphic communities have developed auto-
matic building model reconstruction methods [
7
10
]. Due to the presence of occlusions in
Remote Sens. 2021,13, 1107. https://doi.org/10.3390/rs13061107 https://www.mdpi.com/journal/remotesensing
Remote Sens. 2021,13, 1107 2 of 23
complex city scenes and noise in forward-intersecting point clouds, some building features
(e.g., planes, edges) are degraded or even missed in the collected point clouds, which leads
to unreliable recovery of the topological relationships [11].
To address this problem, various data- and model-driven methods (or their combina-
tion) have been proposed in recent years and substantial improvements in quality have
been achieved [
9
,
12
14
]. However, for complex buildings with imperfect point coverage, it
is difficult to automatically and efficiently reconstruct building models with desired Level
of Detail (LoD) [
15
,
16
], hindering environmental simulation, spatial analysis, and other
model-based applications. To solve this problem, in this paper, a novel framework which
combines the rule-based and the hypothesis-based method is proposed for the efficient
reconstruction of high-quality polygonal building models. Starting with the point clouds
of a single building, the planar primitives and corresponding boundaries are extracted and
regularized to obtain abstracted building counters, followed by a two-stage reconstruction.
In the first stage, the regularity and adjacency of the building counters are used to recover
the topological relationship between different primitives and produce an initial reconstruc-
tion. In the second stage, the topologies of ambiguous areas are removed and reconstructed
by solving an integer linear optimization problem based on the initial reconstruction.
The major contributions of this paper are as follows:
(1)
A novel framework for 3D building reconstruction which combines the efficiency of
traditional rule-based methods and the integrity of recently developed hypothesis-
based methods.
(2)
A method for robust topology estimation that integrates the regularity and adjacency
relationships between building primitives in 3D.
(3)
An effective solution that enforces initial reconstruction results and constraints to
eliminate topological ambiguities.
The remainder of this paper is organized as follows. Section 2provides a brief re-view
of existing work on 3D building reconstruction from point clouds. In Section 3, the details
and key steps of the proposed approach are presented. The performance of the proposed
approach is evaluated in Section 4using real photogrammetric point clouds from aerial
oblique images. Discussion about the method and the experimental results are given in
Section 5. Finally, we draw our conclusions in Section 6.
2. Related Works
According to the City Geography Markup Language (CityGML) model format adopted
by the Open Geospatial Consortium, building models are divided into five LoDs, from
LoD0 to LoD4 [
17
]. With the rapid developments in point-cloud collection and processing
technologies, automatic building reconstruction methods for LoD0 and LoD1 building mod-
els are now relatively mature [
18
21
]. In past decades, researchers in the photogrammetry
and computer vision communities have expended great effort on the (semi-)automatic
reconstruction of LoD2 and even LoD3 building models [7,2224].
Point clouds used for city reconstruction are almost obtained by Light Detection
and Ranging (LiDAR) technology [
25
28
] or photogrammetry [
5
7
]. LiDAR point clouds
usually have more precise coordinates and less noise compared with those generated by
image matching. But the density of point clouds obtained through the Structure-from-
Motion (SfM) and Multi-View Stereo (MVS) pipeline is higher in areas with sufficient
textures. In general, methods for reconstructing 3D building models from point clouds
can be categorized as data-driven [
2
,
7
,
29
], model-driven [
30
,
31
], or a combination of the
two (also called hybrid-driven methods) [
12
,
32
,
33
]. Comparisons about them could be
found in References [
34
,
35
]. As summarized in several previous works, model-driven
methods that adopt top-down strategies require pre-defined hypothetical models, which
hinders their application in free-form building reconstruction [
35
]. In contrast, data-driven
methods that require top-down approaches have the potential to reconstruct complicated
buildings. This kind of reconstruction pipeline normally includes three crucial steps:
primitive extraction, regularization, and topology estimation. Given a point set, linear
Remote Sens. 2021,13, 1107 3 of 23
primitives such as planes [
10
], cylinders, spheres, and line segments [
36
] are first extracted
using methods based on model-fitting (e.g., RANdom Sample Consensus (RANSAC) [
37
] or
Hough transform [
21
,
38
]), region-growing [
39
], or clustering [
40
]. Then, these primitives or
their boundaries are regularized using Manhattan hard or soft constraints [
41
43
]. Finally,
at their core, data-driven methods [
44
] involve estimation and refinement of the adjacency
relations between different primitives to construct final models without any topological
conflicts [
10
,
12
]. Moreover, the hybrid-driven methods combine the primitive extraction
step from the data-driven methods at the first step, then these primitives are used to form
building models with pre-defined combination solutions.
As CityGML LoD2 building models mainly concern rooftop structures, air-borne
laser scanning point clouds have become the major data source for those automatic recon-
struction methods because of their accurate altitude measurements and fewer top-view
occlusions [
22
,
45
,
46
]. By projecting 3D rooftops onto a two-dimensional (2D) horizontal
plane, the topological relations between rooftop primitives are estimated by detecting ridge
edges and jump edges [
8
,
23
,
33
], and these relations are maintained by a binary space parti-
tioning (BSP) tree [
23
], an adjacency matrix [
16
,
47
], or a roof topology graph (RTG) [
13
,
48
].
Then, the subsets of primitives are used to form building models based on pre-defined
rules such as graph edit operations [
13
], planar partitioning, and merging [
23
]. These are
the so-called rule-based methods, which could be either data-driven or hybrid-driven.
In practice, photogrammetric point clouds acquired through SfM and MVS pipelines are
inherently noisier than those collected by laser scanning technology [
49
]. Cameras mounted
on unmanned aerial vehicles are inevitably hampered by occluded areas during the image
collection process, especially on the lower parts of buildings [
50
], which leads to unreliable
geometric accuracy or missing data in the photogrammetric point clouds. Although some
previous works have reported impressive results in the automatic reconstruction LoD2
building models, they may not be suitable for LoD3 building reconstruction from pho-
togrammetric point clouds due to the inferior data quality and difficulty in representing
more complicated topological relations in real 3D space [13,23,47,51].
Recently, researchers have tried to convert the model reconstruction problem into
an optimal subset selection problem with hard constraints [
10
,
14
,
52
]. Given primitives
detected from original point clouds, the object space is first divided into several segments
to form a candidate pool. Then, different hypotheses about the building models are quan-
tized by energy functions that measure their data fitness and rule fitness, with additional
topological constraints, such as watertight. After that, by searching for the maximal (or
minimal) values of the object functions, segments in the candidate pool are labeled as
either selected or not to establish the final building models. These kinds of methods are
referred as hypothesis-based methods in this paper, which could be either data-driven
or hybrid-driven methods. One virtue of these kinds of methods is that they are robust
when data are partially missing. In addition, with help from integer linear programming,
manifold and watertight hard constraints can be embedded to avoid topological conflicts
in the output models [
10
]. Hence, these methods have the potential to reconstruct LoD3
building models that contain more detailed structures [7].
However, with tile views, building façades are visible in the photogrammetric point
clouds generated by aerial oblique images, which result in more planar primitives. The
direct adoption of hypothesis-based reconstruction methods in such scenes may result in
unreasonable computation costs when solving integer linear programming problems [
10
].
As noted by Wang et al. [
35
], global regularities between different primitives in buildings
may reveal their topological relations. If this information could be properly explored and
utilized to estimate building topologies, even partially, the recovered topological relations
could stabilize the solution to decrease artifact and accelerate the problem-solving process
by reducing the size of the candidate pool.
In order to utilize the intrinsic structure of buildings in architectural design to produce
building models with preferred geometry accuracy while eliminating topological conflicts
Remote Sens. 2021,13, 1107 4 of 23
efficiently, a two-stage building reconstruction method that combines the traditional rule-
based and the recently arisen hypothesis-based methods is proposed.
3. Method
3.1. Overview of the Proposed Approach
Starting from the photogrammetric point clouds of aerial oblique images for single
buildings, Figure 1shows the overall workflow of the proposed method. The photogram-
metric point clouds could be produced from images with sufficient overlaps by existing
SfM and MVS pipeline, and single buildings could be extracted from the point clouds
manually or clipped according to existing 2D footprints.
Figure 1. Overall workflow of the proposed method.
In the pre-processing steps, the planar primitives in the point clouds are extracted with
simple parallel and orthogonal constraints using the existing RANSAC-based methods.
Then, the extracted 3D planes which share the same normal orientations are grouped
together. For each plane group, 3D points are projected to 2D space by translating their
centroid to the origin point and rotating the normal of the plane to the positive direction
of the Z-axis, and the consecutive boundary points of each plane are extracted using
alpha-shapes. After that, the boundary points of an individual plane are simplified by
shifting them along their refined normal vector to resist noise, and then grouped into
piecewise smooth segments. Finally, the orientations of each segments in the same plane
group are softly regularized to be parallel (or perpendicular) with each other or the normal
orientation of the other group planes. So, the initial point clouds are abstracted by a set of
Remote Sens. 2021,13, 1107 5 of 23
planar polygons with mutual regularity. For details of the pre-processing step, please refer
to the work by Xie et al. [43].
Si=1,2,···n={Pi,πi,Polyi}(1)
In Equation (1), S
i
represents the detected planar segments (nis the total number of
segments), P
i
is the point set which belong to segment S
i
,
πi
is the regularized base plane
of Si, and Polyiis a regularized boundary polygon of Piin plane πi.
Then, a two-stage reconstruction method is implemented. In the first stage, the adja-
cency relations between different polygons are estimated on the basis of spatial consistency
and mutual regularity rules, even if only partial, to gain an initial reconstruction of the
3D polygonal building model. For areas not reconstructed in the first stage, inspired by
the work of Nan and Wonka [
10
], hypotheses are posed regarding the final model based
on the pairwise intersection in finite distances, followed by the selection of the optimal
combination of candidates by solving a binary linear programming problem.
3.2. Adjacency Detection between Multiple Primitives
In this stage, the robust adjacency relations between different polygons are recovered
in areas with sufficient data support. Specifically, as shown in Figure 2, two types of
topological relations are identified: (1) the intersection of two planar primitives, which
indicates an edge in the model, and (2) the intersection of three planar primitives, which
indicates a vertex in the model.
Figure 2.
Graphic illustration of the two types of topological relations. In polygons A,B, and C, if
edge (Va
1
,Va
2
) matches edge (Vc
1
,Vc
2
), then the polygon pair (A,C) is considered to be pairwise
adjacent. If vertices Va
1
,Vb
1
, and Vc
1
are matched with each other, then polygon triplet (A,B,C) is
considered to intersect at the common point of the three supporting planes.
Although photogrammetric point clouds can be noisy or partly missing, large pla-
nar structures that are well sampled are still reliable for fitting planes and recovering
boundaries. Similar to the work by Arikan et al. [
53
], vertex–vertex matches (VVMs) and
vertex–edge matches (VEMs) between two polygons are searched to determine their pair-
wise adjacency. To obtain robust estimations, the maximal search radius, as described by
Arikan et al. [
53
], should be relatively low (in this paper, it is set to be twice the average
point spacing).
Edge–edge matches are derived from VVMs and/or VEMs. In this work, the adjacency
between two non-parallel polygons, Poly
i
and Poly
j
, is verified by finding at least one pair
of edges from the two polygons that satisfy the following criteria:
(1)
The two edges are parallel or collinear.
(2)
Two VVMs, or one VVM and one VEM, or two VEMs are found for them.
Remote Sens. 2021,13, 1107 6 of 23
Note that, if more than two edges are found to satisfy the first criteria, all of the
possible combination pairs should be tested if they satisfy the second criteria of edge–edge
matches. Then, to detect plane triplets of interest, an undirected graph G(V,E), as shown
in Figure 3, is generated by setting each polygon as a vertex (the blue dots), with the
edge between two vertices indicating matching relationships of the two polygons. As the
intersection of three non-parallel polygons can be defined without ambiguity, the shortest
closed cycles are searched in the graph G(V,E) [
51
], and those with the shortest walk of the
three indicate an underlying intersection of a triplet of polygons if none are parallel (the
red triangles).
Figure 3.
Graphic illustration of polygons and their matching relations. The blue dots represent
polygons while the black lines between two dots indicate that they are matched with each other. The
red triangles indicate the intersections of three non-parallel polygons.
3.3. Building Model Reconstruction with Initial Topology Constraints
In the previous stage, the adjacency relations between different planar primitives
were partially recovered. A set of adjacent planar polygon pairs and a set of intersecting
non-parallel polygon triplets are obtained. So, the topology relationship between different
planar polygons is divide into two parts: the confident part and the ambiguous part.
In this stage, we incorporate these relations to produce candidate faces and constraints
for generating the final building model. Unlike previous work which generates candidate
faces by simply intersecting detected planar segments, we embed the recovered topological
relations in this process to (1) generate more purposeful candidate faces, and (2) reduce the
size of unknown parameters in the energy function. After that, the recovered topology is
used to guide the candidate selection process as soft energy functions and hard constraints
to obtain better reconstructed models with less artifact and in remarkable running time.
3.3.1. Candidate Deduction with Topological and Spatial Hints
Given a set of boundary polygons with base planes, use of their pairwise intersections
is a simple strategy for generating redundant candidate faces. However, the drawback
is that the number of candidate faces increases substantially with increases in the initial
number of detected polygons. In addition, artifactual faces that are obviously invalid may
survive in the reconstructed models. Instead of pairwise intersection of detected planar
segments purely, in this paper, we conduct the candidate generation process based on three
assumptions:
(1)
For adjacent polygon pairs, the candidate faces in each polygon plane might be
bounded by their intersecting lines.
(2)
For adjacent non-parallel polygon triplets, the candidate faces in each polygon plane
might be bounded by the two other intersecting planes.
(3)
The potential intersection points of different polygons might not be far away from
their boundaries.
Remote Sens. 2021,13, 1107 7 of 23
The overall workflow of this process is shown in Figure 4. The planar primitives are
first pairwise intersected with each other within the scope of the enlarged bounding box to
form over-redundant candidate faces. Then, the candidate faces in each plane are processed
separately to eliminate part of them according to the detected topological relations in the
previous steps and the spatial hints. After that, the union set of invalid candidates are
removed. Finally, the remaining candidates in all planes are merged and those that do not
satisfy the two-manifold rules are treated as outliers.
Figure 4. Overall workflow of the proposed candidate face deduction method.
To swiftly eliminate invalid faces in a plane,
π
, we divide the candidate faces, F
P
,
in
π
into three categories, which are marked by green, orange, and gray solid circles
in
Figure 5
. The first category (F
PCover
) includes faces that share common areas with the
detected primitives in the 2D space, these faces are highly confident candidates. The second
category (F
PNear
) includes faces which are not far from the covering area of the detected
primitives and are treated as potential candidates. The rest of the faces (F
PInvalid
) in this
plane are labeled as invalid and should be rejected.
FP=nFCover
P,FNear
P,FInv alid
Po(2)
Remote Sens. 2021,13, 1107 8 of 23
Figure 5.
Illustration of the three constraints used to reduce the number of candidate faces. (
a
): Pairwise Constraint; (
b
):
Triplet Constraint; (
c
): Nearby Constraint. The red lines depict a simplified boundary polygon Ain its supporting plane
πA
,
and the blue lines which separate the plane into several candidate faces are the intersection of other planes in this plane.
The green, orange, and gray solid circles indicate the status of the occupied faces as Cover, Near, and Invalid, respectively.
Based on the above assumptions, three types of constraints are used to classify the
candidate faces obtained by brutal intersection, namely the pairwise constraint (PC), triplet
constraint (TC), and nearby constraint (NC).
Pairwise Constraint: All matched polygon pairs, in which the current one involved
should go through this process. As shown in Figure 5a, consider a polygon Awith support-
ing plane
πA
. If a polygon pair (A,B) is matched in the previous steps, e
A
is the matched
edge in polygon A, and the two supporting planes intersect at line l
AB
. Theoretically, if
the two vertices of e
A
are both on the convex hull of polygon A, the candidate faces in
πA
should be bounded by l
AB
. Then, the candidate faces in
πA
are labeled as covered (green
dot) and invalid (gray dot) respectively, according to their intersection relationship with
polygon A. Note that the same process would be done for polygon Bin its supporting plane
πB
. If both polygon Aand polygon Bare bounded by the intersection line l
AB
, a sharp edge
is implicitly reconstructed in the building model.
Triplet Constraint: All matched polygon triplets, in which the current one involved
should go through this process. As shown in Figure 5b, consider a polygon Awith
supporting planes
πA
, and a detected polygon triplet (A,B,C), whereby the projections
of supporting planes of polygon Band Cin
πA
are two lines, l
AB
and l
AC
. Thus, the
intersection of l
AB
and l
AC
divides
πA
into four parts. Ideally, if polygon Ais located in
only one of the four parts, we could bound the candidate faces in
πA
by the boundaries
of the intersecting lines. In practice, due to the presence of noise in point clouds and the
errors accumulated in the processing steps, we consider polygon Ato be in only one of the
four parts when the area percentage of Ain this part is larger than a given threshold (e.g.,
95%). Then, the candidate faces in
πA
are restricted to being in this part only. The same
Remote Sens. 2021,13, 1107 9 of 23
verification procedure is also performed for polygons Band Cin their supporting planes,
respectively. The best situation is that all three polygons are bounded by the two other
planes, and a corner point at which all three planes intersect is implicitly reconstructed to
bound the building model.
Nearby Constraint: As shown in Figure 5c, for a given polygon Awith supporting
plane
πA
, a haphazard pairwise intersection may result in large numbers of invalid can-
didate faces. Firstly, faces that share common areas with polygon Ain the 2D space (
πA
)
are labeled as Covered. Then, the remained faces that share at least one edge with the
candidate faces in the first category or at least one vertex that is close to (a distance lower
than a given threshold, e.g., 2 m) the vertices of all of the polygons in the first category are
labelled as Near. The rest of the faces in this plane are labeled as Invalid.
The union set of invalid candidate faces are rejected from the candidate pool and those
that do not satisfy the two-manifold rules are also labeled as Invalid, so the recovered
adjacency information in the Section 3.2 is embedded, so that the candidate faces in the
confident part are concise and the redundant candidates are mainly in the ambiguous part.
3.3.2. Face Selection with Initial Constraints
In this step, a subset of candidate faces is selected to establish the final building
model, which prefers some properties and satisfies certain constraints. To find the optimal
configuration of candidate faces, similar to the PolyFit framework [
10
], we use a binary
linear programming approach to quantify the favored properties of the reconstructed
model while imposing certain hard constraints. The optimization model contains the
binary variables shown in Equation (3) below, which was first developed by Nan and
Wonka [10]:
xf,e,es {0, 1},fF,eE,es E(3)
where xfand xeindicate faces in the face pool (F) constructed by all of the candidate faces
or an edge in the edge pool (E) that includes all of the associated edges of the faces in the
face pool that is selected or not, and x
es
indicates whether the edge is sharp or not. In this
work, four properties are favored:
Property 1: Faces supported by a large number of points are favored.
Epts =1
fF
numhnp
projplanef(p)f,dis(p,planef)<εoi
num[p](4)
In Equation (4), pstands for the points in detected primitives, fdonates a face of the
candidate face pool (F), plane
f
represents the regularized 3D plane that supports face f,
proj
plane-f
(p) is the 2D projection of a 3D point ponto the supporting plane of polygon A(
πA
),
dis(p,plane
f
) is the unsigned perpendicular distance from point p to
πA
,
ε
is a pre-defined
distance tolerance, and num() is the size of the set.
Property 2: Faces with a high proportion of their area covered by supporting points
are favored.
Ecover =
f
(1area(fP)
area(f))(5)
This property is quantified by the ratio of overlapped area in candidate face with
the boundary polygon of the supporting points [
10
]. In Equation (5), Pis a 2D polygon
consisting of the boundary of the supporting points for face fand area(
) is the area of the
closed polygon. Both of the areas are calculated in 2D space (the supporting plane of face f).
Equations (4) and (5) are used to determine the degree to which the data fit the candidate
faces, whereby the lower the values, the better the data fit.
Property 3: Intersecting faces in a PC or TC are preferably selected.
Θ(f) = 1+η,fPC TC
1, otherwise (6)
Remote Sens. 2021,13, 1107 10 of 23
In Equation (6),
Θ
(
) is the confidence coefficient of face fand
η
is a non-negative con-
stant (set as 1 in all the experiments in this paper) that increases the confidence coefficient
values for certain faces.
Property 4: Sharp edges associated with two polygons in a PC or TC are favored.
Eedge =
e
xes ·(1length(eeP1eP2)
length(e))(7)
This property is measured by the overlapped ratio of the sharp edge with the share
section of the two matched edges. In Equation (7), estands for an edge which connects two
faces, e
P1
and e
P2
are the projections of two matched edges (if they matched as PC or TC)
on the intersecting line of their corresponding planes, and length(
) represents the unsigned
length of the segments. Longer overlapping ratio is preferred. Hence, the following energy
function is defined:
E=Eplane +ωedge ·Eedge (8)
Eplane=xf·(ωpts ·Epts +ωcover ·Ecover)/Θ(f)(9)
The first term in Equation (8) measures the degree to which the properties fit the
candidate faces, and the second term measures that for their related edges,
ω
stands for
the corresponding weight. As given in Equation (9), the data fitting degree of a face is
composed of two parts, one for point number (E
pts
), the other one for point coverage
(E
cover
). Larger number of inlier points and bigger area of face coverage is preferred. The
denominator
Θ
(f) are used to give larger opportunities for faces involved in a PC or TC. In
the paper, the weights of the three energy forms, Eedge,Eptd , and Ecovrt, are set as identical.
In addition, three constraints are imposed to strengthen the topologic and geometric
features of the building model:
Constraint 1: An edge must be selected when one of its associated faces is selected.
xfasso(ei)=1xei=1 (10)
Constraint 2: When an edge is selected, it connects only two candidate faces.
xei=1num[nxf=1
fasso(ei)o]= 2 (11)
Constraint 3: The edges associated with two polygons in a PC or TC should be sharp.
xei=1, eiPC TC xesi=1 (12)
Properties 1 and 2 and Constraints 1 and 2 are similar to those reported by Nan
and Wonka [
10
] to ensure a watertight polygonal building model, whereas Properties 3
and 4 and Constraint 3 enforce the recovered topology described in Section 3.2 to realize
favorable models. The above objectives and constraints are formulated as a binary linear
programming problem that can be minimized using existing solvers [
54
]. Lastly, the
selected faces are combined to yield the final building model.
4. Experimental Analysis
4.1. Test Data Description and Experimental Settings
To test the performance of the proposed methods, we made qualitative and quanti-
tative comparisons of the reconstruction quality and computational cost of our method
with those of some state-of-the-art (SOTA) methods. The photogrammetric point clouds
of the aerial oblique images of typical buildings in Shenzhen, China, were incorporated
in the experiments. The original images are captured from five directions (one vertical
and four tilt views), so the building façades are (partially) visible. The image overlap is
approximately 60% to 80%. Image orientation and dense image matching are accomplished
by existing solutions in Context Capture. After that, point clouds of single buildings are
Remote Sens. 2021,13, 1107 11 of 23
manually extracted by drawing 2D bounding boxes. The major components of buildings
are visible, and the roof structures are well-sampled. But there are also some missing and
imperfect areas due to occlusions and unfavorable lighting conditions in the lower parts of
the buildings. The point clouds are shown in Figure 6, while Table 1lists basic information
related to the test data.
Figure 6.
Photogrammetric point clouds used in the experiments. The row numbers correspond
to the building IDs. From left to right, the columns show the original point clouds, the segmented
planar primitives, and the extracted outer boundaries of each primitive.
Remote Sens. 2021,13, 1107 12 of 23
Table 1. Basic information related to the test data.
Building ID Number of
Points
Average
Spacing (m)
Footprint Area
(m2)Detected Planes
1 44,034 0.21 234 18
2 60,675 0.15 432 20
3 203,317 0.09 720 17
4 523,233 0.11 2124 33
5 611,982 0.16 672 37
6 548,766 0.20 5978 45
Starting with the single point clouds of a building, we used the region-growing
method to identify the planar segments. Then, the boundaries of the segments were
traced and regularized to concisely abstract the initial building. As plane detection and
regularization were beyond the scope of this study, we set the planar segments and their
regularized boundaries as inputs in our experiments, which could be generated by an
existing method [
43
]. Figure 6shows the original point clouds, the detected planes, and
the regularized boundaries. As shown in Figure 6, the regularized boundaries of detected
planar primitives could preferably represent building outlines at large primitives which
are well-sampled. Meanwhile, in areas with data insufficiency (e.g., under sampling or
noisy), the recovered boundaries are rather ambiguous, and even some gaps occurred (as
pointed out by the circles).
Because PolyFit [
10
] is the method most closely related to ours, we first qualitatively
and quantitatively compared our method with PolyFit with respect to geometric quality
and computational efficiency. Since the ground truth 3D models are not available in
this area, geometric quality of reconstructed models was determined based on a visual
comparison, cloud-to-mesh (C2M) distance statistics, and mesh-to-cloud (M2C) distance
statistics. C2M distances are calculated by projecting the original points to their nearest
faces in the 3D model, and M2C distances are estimated by first sampling the 3D model
and then calculating the cloud-to-cloud distances between the sampled and original point
clouds. For C2M distances, a lower value means better data fitting degree, meanwhile,
for M2C distances, a lower value means fewer artifacts in the model. For both of the two
values, the lower is the better. Besides, computational efficiency is determined by the
recorded running time of each step in generating 3D building models. Comparisons with
two other SOTA building model reconstruction methods are also presented.
4.2. Comparison with PolyFit
Figures 712 show the intermediate candidates generated by the proposed and PolyFit
methods, as well as the final reconstructed models. Table 2provides basic information about
the candidates and final results, along with the computational costs of the reconstruction
process. At first glance, we can observe that the candidate faces generated by the proposed
method are more concise than those produced by PolyFit. Some sharp edges and even some
building corners were already reconstructed by the proposed adjacency-based topology
detection method. As shown in Table 2, the number of candidate faces generated by PolyFit
ranged from 1163 to 15,885, whereas those generated by the proposed method ranged
between 138 and 2964, which again indicates the efficiency of the process of generating
candidate faces by our method. Although the proposed method has to first estimate the
adjacency relations before computing the candidate faces, the running time in this step
(denoted as ADT in Table 2) is relatively fast compared with the computing times of
candidate generation (CGT) and model generation (MGT). The total running times (TT) of
PolyFit and our proposed method, as shown in Table 2for all six tested buildings, reveals
that our method is substantially faster. For simple buildings with fewer planar faces (e.g.,
Buildings #1, 2#, and 3#), although our method used only about 84.2% to 35.6% of the
running time that PolyFit did, the computing times are both in the second level. As a
building became more complicated, the running time for PolyFit increased dramatically
Remote Sens. 2021,13, 1107 13 of 23
because of the geometric growth in the number of candidate faces. For Building #5, PolyFit
used more than 10 min, whereas our method only took 26.428 s, only 4.2% of the total time
that PolyFit took. And for Building #6, PolyFit took about 37 min, whereas our method
accomplished this task in 37.383 s, nearly 60 times faster than PolyFit.
Figure 7.
Comparison of candidate faces and resulting models generated by the proposed method and PolyFit [
10
] for
Building #1.
Figure 8.
Comparison of candidate faces and resulting models generated by the proposed method and PolyFit [
10
] for
Building #2.
Remote Sens. 2021,13, 1107 14 of 23
Figure 9.
Comparison of candidate faces and resulting models generated by the proposed method and PolyFit [
10
] for
Building #3.
Figure 10.
Comparison of candidate faces and resulting models generated by the proposed method and PolyFit [
10
] for
Building #4.
Remote Sens. 2021,13, 1107 15 of 23
Figure 11.
Comparison of candidate faces and resulting models generated by the proposed method and PolyFit [
10
] for
Building #5.
Figure 12.
Comparison of candidate faces and resulting models generated by the proposed method and PolyFit [
10
] for
Building #6.
Remote Sens. 2021,13, 1107 16 of 23
Table 2.
Quantitative statistics on data and computation cost. BID: Building ID; Can. No.: Number
of Candidate faces; Res. No.: Number of resulting faces; ADT: Time for adjacency estimation; CGT:
Time for candidate generation; MGT: Time for resulting Model Generation; TT: Total time.
BID Method Can. No. Res. No. ADT (s) CGT (s) MGT (s) TT (s)
#1 Our 138 99 0.018 0.577 0.049 0.644
PolyFit 1190 114 - 0.646 1.164 1.810
#2 Our 242 163 0.034 0.777 0.040 0.851
PolyFit 1584 169 - 0.752 0.989 1.741
#3 Our 196 159 0.019 2.953 0.043 3.015
PolyFit 1163 159 - 3.106 0.475 3.581
#4 Our 689 533 0.085 11.668 0.019 11.772
PolyFit 6809 578 - 12.014 60.983 72.997
#5 Our 1187 707 0.222 14.266 11.940 26.428
PolyFit 8117 784 - 13.425 619.913 633.338
#6 Our 2964 1489 0.660 18.500 18.886 37.386
PolyFit 15,885 1558 - 17.041 2210.079 2227.120
In addition, in Buildings #2, #4, and #5, there are obvious artifacts in the final build-
ing models generated by PolyFit as compared with the original point clouds shown in
Figure 6. This is a side effect of candidate generation based on the intersections of all
potential planar segments. In contrast, by utilizing constraints derived from adjacency
information, these artifacts are avoided or alleviated in both candidate generation and
candidate selection stages. To verify the objective data fidelity of the reconstructed building
models quantitatively, we calculated the mean values of the C2M and M2C distances, as
shown in Figure 13. The mean C2M distances for the 3D models generated by PolyFit
and the proposed method are all less than 0.30 m, and those for the same buildings are
similar. However, for Buildings #2, #4, and #5, the M2C distances generated by PolyFit
are significantly greater than those generated by the proposed method. Because greater
M2C distances indicates greater data distortion in the final building models, we can infer
that the 3D models generated by our method have better geometric accuracy, which also
accords with the visual judgments shown in Figures 8,10 and 11.
Figure 13.
Mean C2M and M2C distances from Buildings #1 to #6 generated by three-dimensional (3D) models from
PolyFit [10] and the proposed method.
Remote Sens. 2021,13, 1107 17 of 23
4.3. Comparison with Other SOTA Methods
To further evaluate the proposed method, we also conducted an experiment with two
SOTA 3D building reconstruction methods, namely the 2.5D dual contouring method (2.5D
DC) [
55
] and the structuring method [
56
]. As shown in Figure 14, all three methods were
able to reconstruct the overall structure of buildings from point clouds. The structuring
method also preserved some sharp features, such as the edges and corners of buildings in
the models, although some artifactual meshes appeared because of insufficient sampling
and data noise. The 2.5D DC method was able to reconstruct building roofs and protect
planar structures to a certain degree, but it failed to reconstruct Building #6, which has a
typical 3D structure. The mean values of M2C and C2M distances are shown in
Figure 15
.
From Building #1 to Building #4, the C2M distances for all three methods are almost at
the same level, ranging from 0.05 to about 0.15 m. Meanwhile, the C2M distance of the
proposed method in Building #5 is larger than the two others. The reason is that some
small primitives in the original point clouds are not detected and recovered. For Building
#6, the C2M value of 2.5D DC is distinctively large since the façades of the building are not
well-reconstructed. The mean values of M2C of the three methods are between 0.2 to 0.5
m. Note that, in these three methods, the proposed method is the only one which imposes
watertight constraints, so some holes in the point clouds are filled in the output models (e.g.,
the bottom of the building), which inevitably leads to a higher value of M2C statistics. The
situations in Figure 15 reveal that the models recovered by structuring and the proposed
method could represent true 3D structures of buildings. Compared with those of the 2.5D
DC and structuring methods, the 3D models produced by our proposed method were more
concise, with the building surfaces briefly represented by several planar faces. As given in
Table 3, the models generated by the proposed method are composed of less faces, over
90% to 99% of faces are reduced when compared with those generated by 2.5D DC and
structuring, yielding a concise representation of the building structure which would benefit
subsequent applications such as computational simulation in reducing computational costs.
Although some small details were smoothed, the models generated by our method have
better potential for further interactive editing, if needed, since the main structuring of
buildings have been concisely represented without topological conflicts.
Table 3.
Numbers of faces in the reconstructed models by our method, 2.5D DC method [
55
], and
structuring method [56].
Building ID Our 2.5D DC Structuring
1 99 1452 10,972
2 163 2913 14,086
3 159 3797 126,674
4 533 18,170 32,077
5 707 30,527 128,315
6 1489 28,075 138,474
Remote Sens. 2021,13, 1107 18 of 23
Figure 14.
Comparison of models generated by our proposed method, 2.5D DC method [
55
], and structuring method [
56
].
Remote Sens. 2021,13, 1107 19 of 23
Figure 15.
Mean C2M and M2C distances from Buildings #1 to #6 generated by 3D models from 2.5D DC method [
55
],
structuring method [56], and the proposed method.
5. Discussion
As shown in Figures 712, the candidates generated by PolyFit are quite superfluous,
which makes it hard to recognize the building features (edges, corners) from the candidate
faces; meanwhile, the candidates generated by the proposed method have incorporated
some recovered topological information, so some of the building features could be visually
discovered. Consider the information in Figure 14 and Table 3, one virtue of the proposed
method is the concise representation of the original building point clouds with preferred
geometry accuracy. In fact, this is also the virtue of PolyFit. Since our method combines the
initial recovered topological information, which is obtained by the rule-based method, in
the candidate generation and selection process, some unwanted artifacts are alleviated or
avoided. The M2C values in Figure 13 also reveal this virtue. Besides, the running time of
our method is substantially faster than the traditional hypothesis-based methods, such as
PolyFit, especially when the complexity of the building increases. The main difference lies
in the MGT process, where a lower ambiguity needs fewer computation times. For a dataset
with lower noise and better coverage, it is more likely to extract intact building primitives
and to obtain the high-quality regularized boundaries. In these cases, topology relations
between primitives are easier to be detected, leading to better initial reconstructions and
fewer candidates (see the examples in Figures 79). Thus, the running times in the candidate
selection process are less. In extreme cases with favorable data coverage and sampling
density, if all the topological relations of building primitives have been recovered, the
candidate faces generated accordingly should equal to the final output model, because
there is only one selection combination from which conform all the properties in the
energy equations and constraints. Conversely, if the building primitives are not well
detected because of data quality (or some other reasons), the adjacency detection process
may only work in a small part of areas. Then, the number of generated candidates of
our algorithm would increase, leading to longer MGTs and, maybe, some unexpected
artifactual structures in the building models. In the most unfortunate cases where no
Remote Sens. 2021,13, 1107 20 of 23
topology has been recovered in the first stage of our algorithm, the proposed method
would degenerate to a pure hypothesis-based reconstruction method, similar to PolyFit.
6. Conclusions
In this work, we proposed a novel method for efficient reconstructing building models
from photogrammetric point clouds obtained from aerial oblique images in a two-stage
topology recovery process which combined the rule-based and the hypothesis-based meth-
ods. Given the point clouds of a single building, planar primitives and their corresponding
boundaries are extracted and regularized to obtain abstracted building counters. In the
first stage, the adjacency relations between different polygons are estimated based on their
spatial consistency and mutual regularity to form initial reconstruction. In the second
stage, the candidate faces of the building model are generated by the pairwise planar inter-
sections along with three constraints, namely pairwise constraint, triplet constraint, and
nearby constraint, derived from the recovered topology. Lastly, the optimal combination of
candidate faces is selected by solving a binary linear programming problem that shapes the
favored properties under certain constraints. An experimental comparison of our method
with three SOTA methods revealed that the proposed method could efficiently reconstruct
3D building models in several seconds and produce concise models with preferred data
fidelity and geometric accuracy at the decimeter level. Detailed comparison with PolyFit
indicated the high efficiency of the proposed method in reconstructing complex building
models and showed promising prospect for this method in 3D city environmental applica-
tions. The advantage of the proposed methods is the ability to handle point clouds with
various noises in the reconstruction process. However, when the point clouds have large
holes or extensively missing areas, the recovered models by our method may degenerate
to those constructed by PolyFit, which is a disadvantage. Besides, if the density of the
input point clouds is too low, the boundaries extracted from the building primitives may
become unstable as well as the initial topology estimation. In future works, methods
for incorporating multi-scale primitives and 2D features from oriented images should be
developed to handle the situation with low point density and to achieve more detailed
reconstruction.
Author Contributions:
Conceptualization, L.X. and W.W.; methodology, L.X. and Q.Z.; software, L.X.
and H.H.; validation, S.T., Y.L. and Y.Z.; formal analysis, Q.Z. and X.L.; supervision, R.G. All authors
wrote the manuscript. All authors have read and agreed to the published version of the manuscript.
Funding:
This work was supported by National Key R&D Program of China (2019YFB210310,
2019YFB2103104), the National Natural Science Foundation of China (No.42001407, No.41971341,
No. 41971354, 41801392), the Open Fund of Key Laboratory of Urban Land Resources Monitoring
and Simulation, MNR (KF-2019-04-042, KF-2019-04-019), the Guangdong Basic and Applied Basic
Research Foundation (2019A1515110729, 2019A1515010748, 2019A1515011872), and the Open Re-
search Fund of State Key Laboratory of Information Engineering in Surveying Mapping and Remote
Sensing, Wuhan University (No. 20E02).
Acknowledgments:
We would like to thank Liangliang Nan, Florent Lafarge and Qianyi Zhou for
providing the source codes or executable programs for experimental comparison.
Conflicts of Interest: The authors declare no conflict of interest.
Abbreviations
LoD Level of Detail
CityGML City Geography Markup Language
LiDAR Light Detection And Ranging
SfM Structure-from-Motion
MVS Multi-View Stereo
RANSAC RANdom Sample Consensus
Remote Sens. 2021,13, 1107 21 of 23
BSP Binary Space Partitioning
RTG Roof Topology Graph
VVM Vertex–Vertex Match
VEM Vertex–Edge Match
PC Pairwise Constraint
TC Triplet Constraint
NC Nearby Constraint
C2M Cloud to Mesh
M2C Mesh to Cloud
References
1.
Tan, Z.; Lau, K.K.-L.; Ng, E. Planning strategies for roadside tree planting and outdoor comfort enhancement in subtropical
high-density urban areas. Build. Environ. 2017,120, 93–109. [CrossRef]
2.
Wang, S.; Cai, G.; Cheng, M.; Junior, J.M.; Huang, S.; Wang, Z.; Su, S.; Li, J. Robust 3D reconstruction of building surfaces from
point clouds based on structural and closed constraints. ISPRS J. Photogramm. Remote Sens. 2020,170, 29–44. [CrossRef]
3.
Badach, J.; Voordeckers, D.; Nyka, L.; Van Acker, M. A framework for air quality management zones—Useful GIS-based tool for
urban planning: Case studies in Antwerp and Gda ´nsk. Build. Environ. 2020,174, 106743. [CrossRef]
4.
Biljecki, F.; Stoter, J.; LeDoux, H.; Zlatanova, S.; Çöltekin, A. Applications of 3D city models: State of the art review. ISPRS Int. J.
Geo-Inf. 2015,4, 2842–2889. [CrossRef]
5.
Toschi, I.; Ramos, M.M.; Nocerino, E.; Menna, F.; Remondino, F.; Moe, K.; Poli, D.; Legat, K.; Fassi, F. Oblique photogrammetry
supporting 3D urban reconstruction of complex scenarios. ISPRS Arch. 2017,XLII-1/W1, 519–526. [CrossRef]
6.
Yu, D.; Ji, S.; Liu, J.; Wei, S. Automatic 3D building reconstruction from multi-view aerial images with deep learning. ISPRS J.
Photogramm. Remote Sens. 2021,171, 155–170. [CrossRef]
7.
Liu, X.; Zhang, Y.; Ling, X.; Wan, Y.; Liu, L.; Li, Q. TopoLAP: Topology recovery for building reconstruction by deducing the
relationships between linear and planar primitives. Remote Sens. 2019,11, 1372. [CrossRef]
8.
Awrangjeb, M.; Gilani, S.A.N.; Siddiqui, F.U. An effective data-driven method for 3D building roof reconstruction and robust
change detection. Remote Sens. 2018,10, 1512. [CrossRef]
9.
Jarz ˛abek-Rychard, M.; Borkowski, A. 3D building reconstruction from ALS data using unambiguous decomposition into
elementary structures. ISPRS J. Photogramm. Remote Sens. 2016,118, 1–12. [CrossRef]
10. Nan, L.; Wonka, P. PolyFit: Polygonal Surface Reconstruction from Point Clouds. In Proceedings of the 2017 IEEE International
Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2372–2380.
11. Verdie, Y.; Lafarge, F.; Alliez, P. LOD generation for urban scenes. ACM Trans. Graph. 2015,34, 1–14. [CrossRef]
12.
Xiong, B.; Elberink, S.O.; Vosselman, G. A graph edit dictionary for correcting errors in roof topology graphs reconstructed from
point clouds. ISPRS J. Photogramm. Remote Sens. 2014,93, 227–242. [CrossRef]
13.
Xu, B.; Jiang, W.; Li, L. HRTT: A hierarchical roof topology structure for robust building roof reconstruction from point clouds.
Remote Sens. 2017,9, 354. [CrossRef]
14.
Ochmann, S.; Vock, R.; Klein, R. Automatic reconstruction of fully volumetric 3D building models from oriented point clouds.
ISPRS J. Photogramm. Remote Sens. 2019,151, 251–262. [CrossRef]
15.
Gruen, A.; Schubiger, S.; Qin, R.; Schrotter, G.; Xiong, B.; Li, J.; Ling, X.; Xiao, C.; Yao, S.; Nuesch, F. Semantically enriched high
resolution LoD3 building model generation. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci.
2019
,XLII-4/W15, 11–18.
[CrossRef]
16.
Wen, X.; Xie, H.; Liu, H.; Yan, L. Accurate reconstruction of the LoD3 building model by integrating multi-source point clouds
and oblique remote sensing imagery. ISPRS Int. J. Geo-Inf. 2019,8, 135. [CrossRef]
17. Gröger, G.; Plümer, L. CityGML–Interoperable semantic 3D city models. ISPRS J. Photogramm. 2012,71, 12–33. [CrossRef]
18.
Zhang, K.; Yan, J.; Chen, S.-C. Automatic construction of building footprints from airborne LiDAR data. IEEE Trans. Geosci.
Remote Sens. 2006,44, 2523–2533. [CrossRef]
19.
Zhao, Z.; Duan, Y.; Zhang, Y.; Cao, R. Extracting buildings from and regularizing boundaries in airborne LiDAR data using
connected operators. Int. J. Remote Sens. 2016,37, 889–912. [CrossRef]
20.
Chen, Q.; Wang, L.; Waslander, S.L.; Liu, X. An end-to-end shape modeling framework for vectorized building outline generation
from aerial images. ISPRS J. Photogramm. Remote Sens. 2020,170, 114–126. [CrossRef]
21.
Widyaningrum, E.; Gorte, B.; Lindenbergh, R. Automatic building outline extraction from ALS point clouds by ordered points
aided hough transform. Remote Sens. 2019,11, 1727. [CrossRef]
22. Vosselman, G. Building reconstruction using planar faces in very high density height data. ISPRS Arch. 1999,32, 87–94.
23.
Sohn, G.; Huang, X.; Tao, V. Using a binary space partitioning tree for reconstructing polyhedral building models from airborne
LiDAR data. Photogramm. Eng. Remote Sens. 2008,74, 1425–1438. [CrossRef]
24.
Yang, B.; Huang, R.; Li, J.; Tian, M.; Dai, W.; Zhong, R. Automated reconstruction of building LoDs from airborne LIDAR point
clouds using an improved morphological scale space. Remote Sens. 2016,9, 14. [CrossRef]
25.
Kurdi, F.T.; Awrangjeb, M.; Munir, N. Automatic filtering and 2D modeling of airborne laser scanning building point cloud. Trans.
GIS 2021,25, 164–188. [CrossRef]
Remote Sens. 2021,13, 1107 22 of 23
26.
Kurdi, F.T.; Awrangjeb, M. Automatic evaluation and improvement of roof segments for modelling missing details using Lidar
data. Int. J. Remote Sens. 2020,41, 4702–4725. [CrossRef]
27.
Song, J.; Xia, S.; Wang, J.; Chen, D. Curved buildings reconstruction from airborne LiDAR data by matching and deforming
geometric primitives. IEEE Trans. Geosci. Remote Sens. 2021,59, 1660–1674. [CrossRef]
28.
Kulawiak, M.; Lubniewski, Z. Improving the accuracy of automatic reconstruction of 3D complex buildings models from airborne
LiDAR point clouds. Remote Sens. 2020,12, 1643. [CrossRef]
29.
Kim, K.; Shan, J. Building roof modeling from airborne laser scanning data based on level set approach. ISPRS J. Photogramm.
Remote Sens. 2011,66, 484–497. [CrossRef]
30.
Poullis, C.; You, S. Photorealistic large-scale urban city model reconstruction. IEEE Trans. Vis. Comput. Graph.
2008
,15, 654–669.
[CrossRef]
31.
Henn, A.; Gröger, G.; Stroh, V.; Plümer, L. Model driven reconstruction of roofs from sparse LiDAR point clouds. ISPRS J.
Photogramm. Remote Sens. 2013,76, 17–29. [CrossRef]
32.
Lafarge, F.; Descombes, X.; Zerubia, J.; Pierrot-Deseilligny, M. Automatic building extraction from DEMs using an object approach
and application to the 3D city modeling. ISPRS J. Photogramm. Remote Sens. 2008,63, 365–381. [CrossRef]
33.
Cao, R.; Zhang, Y.; Liu, X.; Zhao, Z. 3D building roof reconstruction from airborne LiDAR point clouds: A framework based on a
spatial database. Int. J. Geogr. Inf. Sci. 2017,31, 1359–1380. [CrossRef]
34. Tarsha Kurdi, F.; Landes, T.; Grussenmeyer, P.; Koehl, M. Model-driven and data-driven approaches using Lidar data: Analysis
and comparison. ISPRS Arch. 2007,XXXVI W49A, 1682–1687.
35.
Wang, R.; Peethambaran, J.; Chen, D. LiDAR point clouds to 3D urban models: A review. IEEE J. Sel. Top. Appl. Earth Obs. Remote
Sens. 2018,11, 606–627. [CrossRef]
36.
Lin, Y.; Wang, C.; Cheng, J.; Chen, B.; Jia, F.; Chen, Z.; Li, J. Line segment extraction for large scale unorganized point clouds.
ISPRS J. Photogramm. Remote Sens. 2015,102, 172–183. [CrossRef]
37.
Schnabel, R.; Wahl, R.; Klein, R. Efficient RANSAC for point-cloud shape detection. Comput. Graph. Forum
2007
,26, 214–226.
[CrossRef]
38.
Tarsha Kurdi, F.; Landes, T.; Grussenmeyer, P. Hough-transform and extended RANSAC algorithms for automatic detection of
3D building roof planes from LiDAR data. In Proceedings of the ISPRS Workshop on Laser Scanning 2007 and SilviLaser 2007,
Espoo, Finland, 12–14 September 2007; Volume 36, pp. 407–412.
39.
Lari, Z.; Habib, A. An adaptive approach for the segmentation and extraction of planar and linear/cylindrical features from laser
scanning data. ISPRS J. Photogramm. Remote Sens. 2014,93, 192–212. [CrossRef]
40. Lu, X.; Yao, J.; Tu, J.; Li, K.; Li, L.; Liu, Y. Pairwise linkage for point cloud segmentation. ISPRS Ann. 2016,III-3, 201–208.
41.
Sampath, A.; Shan, J. Building boundary tracing and regularization from airborne LiDAR point clouds. Photogramm. Eng. Remote
Sens. 2007,73, 805–812. [CrossRef]
42.
Li, M.; Nan, L.; Smith, N.; Wonka, P. Reconstructing building mass models from UAV images. Comput. Graph.
2016
,54, 84–93.
[CrossRef]
43.
Xie, L.; Zhu, Q.; Hu, H.; Wu, B.; Li, Y.; Zhang, Y.; Zhong, R. Hierarchical regularization of building boundaries in noisy aerial
laser scanning and photogrammetric point clouds. Remote Sens. 2018,10, 1996. [CrossRef]
44.
Chen, D.; Wang, R.; Peethambaran, J. Topologically aware building rooftop reconstruction from airborne laser scanning point
clouds. IEEE Trans. Geosci. Remote Sens. 2017,55, 7032–7052. [CrossRef]
45.
Chen, D.; Zhang, L.; Mathiopoulos, P.T.; Huang, X. A Methodology for automated segmentation and reconstruction of urban 3D
buildings from ALS point clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014,7, 4199–4217. [CrossRef]
46.
Benciolini, B.; Ruggiero, V.; Vitti, A.; Zanetti, M. Roof planes detection via a second-order variational model. ISPRS J. Photogramm.
Remote Sens. 2018,138, 101–120. [CrossRef]
47.
Sampath, A.; Shan, J. Segmentation and reconstruction of polyhedral building roofs from aerial lidar point clouds. IEEE Trans.
Geosci. Remote Sens. 2010,48, 1554–1567. [CrossRef]
48.
Elberink, S.O.; Vosselman, G. Building reconstruction by target based graph matching on incomplete laser data: Analysis and
limitations. Sensors 2009,9, 6101–6118. [CrossRef]
49.
Hu, H.; Chen, C.; Wu, B.; Yang, X.; Zhu, Q.; Ding, Y. Texture-aware dense image matching using ternary census transform. ISPRS
Ann. 2016,III-3, 59–66.
50.
Wu, B.; Xie, L.; Hu, H.; Zhu, Q.; Yau, E. Integration of aerial oblique imagery and terrestrial imagery for optimized 3D modeling
in urban areas. ISPRS J. Photogramm. Remote Sens. 2018,139, 119–132. [CrossRef]
51.
Perera, G.S.N.; Maas, H.-G. Cycle graph analysis for 3D roof structure modelling: Concepts and performance. ISPRS J. Photogramm.
Remote Sens. 2014,93, 213–226. [CrossRef]
52. Bauchet, J.-P.; Lafarge, F. Kinetic shape reconstruction. ACM Trans. Graph. 2020,39, 1–14. [CrossRef]
53.
Arikan, M.; Schwärzler, M.; Flöry, S.; Wimmer, M.; Maierhofer, S. O-snap: Optimization-based snapping for modeling architecture.
ACM Trans. Graph. 2013,32, 6. [CrossRef]
54.
Gamrath, G.; Anderson, D.; Bestuzheva, K.; Chen, W.; Eifler, L.; Gasse, M.; Gemander, P.; Gleixner, A.; Gottwald, L.; Halbig, K.;
et al. The SCIP Optimization Suite 7.0; Optimization Online; Zuse Institut: Berlin, Germany, 2020.
Remote Sens. 2021,13, 1107 23 of 23
55.
Zhou, Q.-Y.; Neumann, U. 2.5D Dual Contouring: A robust approach to creating building models from aerial LiDAR point clouds.
In Constructive Side-Channel Analysis and Secure Design, Proceedings of the 3rd International Workshop, COSADE, Darmstadt, Germany,
3–4 May 2012; Springer: Berlin, Germany, 2012.
56. Lafarge, F.; Alliez, P. Surface reconstruction through point set structuring. Comput. Graph. Forum 2013,32, 225–234. [CrossRef]
... as an example, Figure 11a represents the built 3D model, Figure 11b represents the vertices on the model (each point is a sphere), Figure 11c represents the wireframe of the model, and Figure 11d The second aspect concerns the accuracy of the models. This work mainly uses the orthogonal distance from the original building points to their nearest faces of the building 3D model (P2M) and the shortest distance from the building model vertices to the original building points (O2P) to assess the model accuracy [15,16]. Taking building 1 we mentioned in Section 3.1 as an example, Figure 12a represents the combination of the input point clouds and the built 3D model, Figure 12b represents the P2M distance of each point The second aspect concerns the accuracy of the models. ...
... Taking building 1 we mentioned in Section 3.1 as an example, Figure 12a represents the combination of the input point clouds and the built 3D model, Figure 12b represents the P2M distance of each point The second aspect concerns the accuracy of the models. This work mainly uses the orthogonal distance from the original building points to their nearest faces of the building 3D model (P2M) and the shortest distance from the building model vertices to the original building points (O2P) to assess the model accuracy [15,16]. Taking building 1 we mentioned in Section 3.1 as an example, Figure 12a represents the combination of the input point clouds and the built 3D model, Figure 12b represents the P2M distance of each point in the point clouds, Figure 12c represents the combination of the input point clouds and the building model vertices, and Figure 12d represents the O2P distance of each building model vertex. ...
... Remote Sens. 2024,16, 792 ...
Article
Full-text available
The three-dimensional (3D) reconstruction of buildings using photogrammetric point clouds is important for many applications, ranging from digital city construction to urban energy consumption analysis. However, problems such as building complexity and point cloud flaws may lead to incorrect modeling, which will affect subsequent steps such as texture mapping. This paper introduces a pipeline for building surface reconstruction from photogrammetric point clouds, employing a hybrid method that combines connection evaluation and framework optimization. Firstly, the plane segmentation method divides building point clouds into several pieces, which is complemented by a proposed candidate plane generation method aimed at removing redundancies and merging similarities. Secondly, the improved connection evaluation method detects potential skeleton lines from different planes. Subsequently, a framework optimization method is introduced to select suitable undirected polygonal boundaries from planes, forming the basis for plane primitives. Finally, by triangulating all plane primitives and filling holes, a building surface polygonal model is generated. Experiments conducted on various building examples provide both qualitative and quantitative evidence that the proposed hybrid method outperforms many existing methods, including traditional methods and deep learning methods. Notably, the proposed method successfully reconstructs the main building structures and intricate details, which can be further used to generate textural models and semantic models. Experimental results validate that the proposed method can be used for the surface reconstruction from photogrammetric point clouds of planar buildings.
... It has also been demonstrated that model-driven reconstruction methods can be used to semi-automatically reconstruct 3D city models from large-format aerial imagery as illustrated in Figure 2.1 [50]. Reconstruction of 3D buildings can also be performed with rule-based methods and photogrammetric point clouds [51]. A common and fast technique for 3D building reconstruction models is building footprint extrusion. ...
... Subsequent plane assembling techniques then either construct a connectivity graph between the detected planes [8,11,69] to find out the vertices and edges composing the output mesh, or compute a space decomposition by plane slicing operations before extracting a mesh from it [5,9,23,41]. In particular, PolyFit [41] has been used in several building reconstruction pipelines [22,35,64]. However, due to the complexity of its binary linear optimization, these approaches do not scale well to complex structures composed of a few dozens or hundred planar components. ...
... In addition, there are lots of works which focus on vectorizing urban buildings in the form of levels of detail (LODs). Most of these methods build city models with levels of detail ranging from LOD2 to LOD3 (Arefi et al., 2008;Han et al., 2021b;Verdie et al., 2015;Xie et al., 2021;Zhu et al., 2018Zhu et al., , 2020, or specially reconstruct 2.5D models (Bauchet and Lafarge, 2019;Duan and Lafarge, 2016;Lafarge and Mallet, 2012;Li et al., 2016;Poullis and You, 2009), while our goal is to provide a complete and detailed 3D vectorized model for more complicated general scenes including urban buildings and indoor scenes. ...
Article
Full-text available
We present a novel high-quality vectorized modeling pipeline for large-scale scenes. With a reconstructed dense point cloud and its corresponding multi-view source images and camera parameters as input, our system can efficiently reconstruct a geometrically complete and detail-preserved vectorized model. Unlike most existing planar shape assembling methods which cannot handle large-scale vectorized modeling well due to the limitation of memory and computation, we can achieve complete high-quality vectorized modeling for complicated large-scale scenes in a time and memory efficient way. Our pipeline first carries out a 3D semantic segmentation on the dense point cloud, by performing 2D semantic labeling on the source images with a semantic segmentation network and fusing the 2D semantic labels into the point cloud. According to the fused dense 3D semantic labels, we then divide the scene into main structure including the grounds, walls and ceilings, and isolated objects that do not belong to the main structure. After the scene division completes, vectorized modeling is performed successively on the main structure and isolated objects to extract their polygonal models respectively instead of vectorizing the whole scene, to improve both time and memory efficiencies. Additionally, the previously vectorized main polygonal structures are used as priors to refine the segmentation and guide the vectorization of the objects to ensure the geometrical completeness and topological consistency of the entire vectorized model. Especially, during the vectorization procedure, a well designed binary space partition tree is designed to better slice the space so that high-quality polygonal mesh with more geometric details can be reconstructed with both time and memory efficiencies. Experiments with quantitative and qualitative evaluations on large-scale scenes demonstrate the accuracy and efficiency of the proposed vectorization pipeline. We also compare our method with state-of-the-art planar shape reconstruction methods to show its effectiveness in reconstructing large-scale vectorized models.
... More specifically, Dong et al. [7] locate autonomous vehicles in underground mines by combining image recognition technology and data from other sensors. Xie et al. [8] extract planar primitives from point clouds and reconstruct 3D building models based on topological relationships. Li et al. [9] decomposite 3D space by partitioning different buildings based on planar primitives. ...
Article
Full-text available
Planar reconstruction detects planar segments and deduces their 3D planar parameters (normals and offsets) from the input image; this has significant potential in the fields of digital preservation of cultural heritage, architectural design, robot navigation, intelligent transportation, and security monitoring. Existing methods mainly employ multiple-view images with limited overlap for reconstruction but lack the utilization of the relative position and rotation information between the images. To fill this gap, this paper uses two views and their relative camera pose to reconstruct indoor scene planar surfaces. Firstly, we detect plane segments with their 3D planar parameters and appearance embedding features using PlaneRCNN. Then, we transform the plane segments into a global coordinate frame using the relative camera transformation and find matched planes using the assignment algorithm. Finally, matched planes are merged by tackling a nonlinear optimization problem with a trust-region reflective minimizer. An experiment on the Matterport3D dataset demonstrates that the proposed method achieves 40.67% average precision of plane reconstruction, which is an improvement of roughly 3% over Sparse Planes, and it improves the IPAA-80 metric by 10% to 65.7%. This study can provide methodological support for 3D sensing and scene reconstruction in sparse view contexts.
... To circumvent the problem of occlusion, sparsity and/or nonuniform point density, classical building reconstruction approaches (Nan and Wonka, 2017;Xie et al., 2021) usually employ primitive fitting algorithms while typical deep learning (DL) − based approaches resolve it via point cloud completion (PCC) networks Pan et al., 2021;Cai et al., 2022). However, most PCC models are built with input point sets that are shape-wise correct, uniformly dense and only incomplete, whereas airborne PCs on the other hand are not just incomplete but also have point density variations and noise. ...
Article
Full-text available
The benefits of having digital twins of urban buildings are numerous. However, a major difficulty encountered in their creation from airborne LiDAR point clouds is the effective means of accurately reconstructing significant occlusions amidst point density variations and noise. To bridge the noise/sparsity/occlusion gap and generate high fidelity 3D building models, we propose APC2Mesh which integrates point completion into a 3D reconstruction pipeline, enabling the learning of dense geometrically accurate representation of buildings. Specifically, we leveraged complete points generated from occluded ones as input to a linearized skip attention-based deformation network for 3D mesh reconstruction. In our experiments, conducted on 3 different scenes, we demonstrate that: (1) Compared to SoTA methods like NDF, Points2Poly, Point2Mesh, and a few others, APC2Mesh ranked second in positional RMSE and first in directional RSME, with error magnitudes of 0.0134 m and 0.1581, respectively. This indicates the efficacy of APC2Mesh in handling the challenges of airborne building points of diverse styles and complexities. (2) The combination of point completion with typical deep learning-based 3D point cloud reconstruction methods offers a direct and effective solution for reconstructing significantly occluded airborne building points. As such, this neural integration holds promise for advancing the creation of digital twins for urban buildings with greater accuracy and fidelity. Our source code is available at https://github.com/geospatial-lab/APC2Mesh.
... Some works extend this idea to enhance the quality of the results in specific situations, like for ALS data with only roofs [6] or concentrate on recovery of a broken topology [24]. Similarly, Ref. [25] proposes a two-stage topological recovery process to create candidate faces under three constraints and, then, remove the redundant faces by optimization. There are also deep-learning-based methods under this hybrid strategy. ...
Article
Full-text available
Existing contour-based building-reconstruction methods face the challenge of producing low-poly results. In this study, we introduce a novel iterative contour-based method to reconstruct low-poly meshes with only essential details from mesh soups. Our method focuses on two primary targets that determine the quality of the results: reduce the total number of contours, and generate compact surfaces between contours. Specifically, we implemented an iterative pipeline to gradually extract vital contours by loss and topological variance, and potential redundant contours will be removed in a post-processing procedure. Based on these vital contours, we extracted the planar primitives of buildings as references for contour refinement to obtain compact contours. The connection relationships between these contours are recovered for surface generation by a contour graph, which is constructed using multiple bipartite graphs. Then, a low-poly mesh can be generated from the contour graph using our contour-interpolation algorithm based on polyline splitting. The experiments demonstrated that our method produced satisfactory results and outperformed the previous methods.
... Plane-based methods. A typical pipeline of plane-based methods (Li et al., 2016;Nan and Wonka, 2017;Bouzas et al., 2020;Wang et al., 2020;Xie et al., 2021;Yang et al., 2022;Chen et al., 2022;Bauchet and Lafarge, 2020) takes the initial planes detected by RANSAC (Schnabel et al., 2007) or region growing (Gatzke and Grimm, 2006;Lafarge and Mallet, 2012) as inputs first. Then, the pipeline assembles the input planes through two key steps: (1) candidate plane or 3D cell partitioning, and (2) candidate selection. ...
Article
[Free PDF:🌐 https://frankxue.com/#publications 🌐] Compact building models are demanded by global smart city applications, while high-definition urban 3D data is increasingly accessible by dint of the advanced reality capture technologies. Yet, existing building reconstruction methods encounter crucial bottlenecks against high-definition data of large scales and high-level complexity, particularly in high-density urban scenes. This paper proposes a Building Section Skeleton (BSS) to reflect architectural design principles about parallelism and symmetries. A BSS atom describes a pair of intrinsic parallel or symmetric points; a BSS segment clusters dense BSS atoms of a pair of symmetric surfaces; the polyhedra of all BSS segments further echo the architectural forms and reconstructability. To prove the concepts of BSS for automatic compact reconstruction, this paper presents a BSS method for building reconstruction that consists of one stage of BSS segments hypothesizing and another stage of BSS segments merging. Experiments and comparisons with four state-of-the-art methods have been conducted on 15 diverse scenes encompassing more than 60 buildings. Results confirmed that the BSS method achieves frontiers in compactness, robustness, geometric accuracy, and efficiency, simultaneously, especially for highdensity urban scenes. On average, the BSS method reconstructed each scene into 623 triangles with a root-mean-square deviation (RMSD) of 0.82 m, completing the process in 110 seconds. First, the proposed BSS is an expressive 3D feature reflecting architectural designs in high-density cities, and can open new avenues to city modeling and other urban remote sensing and photogrammetry studies. Second, for practitioners in smart city development, the BSS method for building reconstruction offers an accurate and efficient approach to compact building and city modeling. The source code and tested scenes are available at https://github.com/eiiijiiiy/sobss.
Article
Full-text available
The reconstruction of buildings using inhomogeneous and unstructured point clouds is a challenging task for photogrammetry and computer vision research communities. A new approach for 3D building surface modeling, based on closed constraints, is proposed. First, a region growth algorithm is applied to fit the input point clouds by a set of candidate planes. Then, additional candidate planes are generated from the initial planes according to a rigid transformation followed by expanding the original primitive set to the candidate model set through generation rules. Furthermore, an energy function is employed to combine the data fitting errors with the structural constraints at the model selection stage. Finally, the 3D building surface model is generated from the candidate set through energy minimization. More precisely speaking, we adopt the surface optimization scheme that enforces the 3D polygonal surfaces of the building to be consistent with a priori geometric structures. Our approach was assessed using multi-source datasets with different densities, noise levels covering diverse and complex structures. The experimental results demonstrated that the proposed approach achieves better accuracy and robustness than those of several state-of-the-art methods.
Article
Full-text available
The study presented in this paper introduced a new fully automatic three-dimensional building reconstruction method that can generate first level of detail (LoD 1) building models from multi-view aerial images without any assistance from other data. The accuracy and completeness of our reconstructed models have approached that of manually delineated models to a large extent. The presented method consists of three parts: (1) efficient dense matching and Earth surface reconstruction, (2) reliable building footprint extraction and polygon regularization, and (3) highly accurate height inference of building roofs and bases. First, our novel deep learning-based multi-view matching method, composed of a convolutional neural network, gated recurrent convolutions, and a multi-scale pyramid matching structure, is used to reconstruct the digital surface model (DSM) and digital orthophoto map (DOM) efficiently without generating epipolarly rectified images. Second, our three-stage 2D building extraction method is introduced to deliver reliable and accurate building contours. Deep-learning based seg-mentation, assisted with DSM, is used to segment buildings from backgrounds; and the generated building maps are fused with a terrain classification algorithm to reach better segmentation results. A polygon regularization algorithm and a level set algorithm are thereafter employed to transfer the binary segmentation maps to structured vector-form building polygons. Third, a novel method is introduced to infer the height of building roofs and bases using adaptive local terrain filtering and neighborhood buffer analysis. We tested our method on a large experimental area that covered 2284 aerial images and 782 various types of buildings. Our results as far as correctness and completeness exceeded the results of other similar methods in a between-method comparison by at least 15% for individual 3D building models with many of them comparable to manual delineation results.
Article
Full-text available
Abstract This article suggests a new approach to automatic building footprint modeling using exclusively airborne LiDAR data. The first part of the suggested approach is the filtering of the building point cloud using the bias of the Z-coordinate histogram. This operation aims to detect the points of roof class from the building point cloud. Hence, eight rules for histogram interpretation are suggested. The second part of the suggested approach is the roof modeling algorithm. It starts by detecting the roof planes and calculating their adjacency matrix. Hence, the roof plane boundaries are classified into four categories: (1) outer boundary; (2) inner plane boundaries; (3) roof detail boundaries; and (4) boundaries related to the missing planes. Finally, the junction relationships of roof plane boundaries are analyzed for detecting the roof vertices. With regard to the resulting accuracy quantification, the average values of the correctness and the completeness indices are employed in both approaches. In the filtering algorithm, their values are respectively equal to 97.5 and 98.6%, whereas they are equal to 94.0 and 94.0% in the modeling approach. These results reflect the high efficacy of the suggested approach.
Article
Full-text available
This project was partially funded by Luxcarta Technology. We thank our anonymous reviewers for their input, Qian-Yi Zhou and Arno Knapitsch for providing us datasets from the Tanks and Temples benchmark (Meeting Room,Horse,M60,Barn,Ignatius,Courthouse and Church), and Pierre Alliez, Mathieu Desbrun and George Drettakis for their help and advice. We are also grateful to Liangliang Nan and Hao Fang for sharing comparison materials. Datasets Full thing,Castle,Tower of Piand Hilbert cube originate from Thingi 10K,Hand, Rocker Arm, Fertility and LansfromAim@Shape, and Stanford Bunny and Asian Dragon from the Stanford 3D Scanning Repository.
Article
Full-text available
Airborne light detection and ranging (LiDAR) data are widely applied in building reconstruction, with studies reporting success in typical buildings. However, the reconstruction of curved buildings remains an open research problem. To this end, we propose a new framework for curved building reconstruction via assembling and deforming geometric primitives. The input LiDAR point clouds are first converted into contours where individual buildings are identified. After recognizing geometric units (primitives) from building contours, we get initial models by matching the basic geometric primitives to these primitives. To polish assembly models, we employ a warping field for model refinements. Specifically, an embedded deformation (ED) graph is constructed via downsampling the initial model. Then, the point to model displacements is minimized by adjusting node parameters in the ED graph based on our objective function. The presented framework is validated on several highly curved buildings collected by various LiDAR in different cities. The experimental results, as well as accuracy comparison, demonstrate the advantage and effectiveness of our method. The new insight attributes to an efficient reconstruction manner. Moreover, we prove that the primitive-based framework significantly reduces the data storage to 10%-20% of classical mesh models.
Article
Full-text available
Due to high requirements of variety of 3D spatial data applications with respect to data amount and quality, automatized, efficient and reliable data acquisition and preprocessing methods are needed. The use of photogrammetry techniques—as well as the light detection and ranging (LiDAR) automatic scanners—are among attractive solutions. However, measurement data are in the form of unorganized point clouds, usually requiring transformation to higher order 3D models based on polygons or polyhedral surfaces, which is not a trivial process. The study presents a newly developed algorithm for correcting 3D point cloud data from airborne LiDAR surveys of regular 3D buildings. The proposed approach assumes the application of a sequence of operations resulting in 3D rasterization, i.e., creation and processing of a 3D regular grid representation of an object, prior to applying a regular Poisson surface reconstruction method. In order to verify the accuracy and quality of reconstructed objects for quantitative comparison with the obtained 3D models, high-quality ground truth models were used in the form of the meshes constructed from photogrammetric measurements and manually made using buildings architectural plans. The presented results show that applying the proposed algorithm positively influences the quality of the results and can be used in combination with existing surface reconstruction methods in order to generate more detailed 3D models from LiDAR scanning.
Article
Full-text available
ABSTRACT Despite the large number of studies conducted during the last three decades concerning 3D building modelling starting from Light detection and ranging (Lidar) data, two persistent problems still exist. The first one is the absence of some roof details, which will not only disappear in the building roof model due to their small areas regarding the point density but are also considered as undesirable noise among the modelling procedures. The second problem consists in that the involved segmentation algorithms do not perform well in the presence of noise in the building point cloud data. These two problems generate undesirable deformation in the final 3D building model. This paper proposes a new automatic approach for detecting and modelling the missing roof details in addition to improving the building roof segments. In this context, the error map matrix, which presents the deviations of points to their fitting planes, is considered. Moreover, this matrix is analysed in order to deduce the mask of missing roof details. At this stage, a new numeric factor is defined for estimating the roof segmentation accuracy in addition to the validity of the roof segmentation result. Then, the building point cloud is enhanced in order to decrease the negative noise influence and, consequently, to improve the building roof segments. Finally, the functionality and the accuracy of the proposed approach are tested and discussed.
Article
Full-text available
There is a growing recognition of the importance of proper urban design in the improvement of air flow and pollution dispersion and in reducing human exposure to air pollution. However, a limited number of studies have been published so far focusing on the development of standard procedures which could be applied by urban planners to effectively evaluate urban conditions with respect to air quality. To fill this gap, a new approach for the determination of urban Air Quality Management Zones (AQMZs) was proposed and presented based on two case studies: Antwerp, Belgium and Gdańsk, Poland. The main objectives of the study were to 1) formulate a theoretical framework for the management of urban ventilation potential and human exposure to air pollution and to 2) develop methods for its implementation by means of a geographic information system (GIS). As a result of the analysis, the typologies that may be associated with decreased ventilation potential and the areas that require close monitoring due to potential human exposure to air pollution were identified for both cities. It is advocated that delimiting these typologies – combined with investigating local climate, wind and topography conditions and air pollution characteristics – could constitute a preliminary step in the urban planning process aimed at air quality improvement. These methods can be further applied to other urban areas in order to indicate where detailed studies are required and to facilitate the development of planning guidelines. Moreover, the directions for further research and urban planning strategies were discussed.
Article
The identification and annotation of buildings has long been a tedious and expensive part of high-precision vector map production. The deep learning techniques such as fully convolution network (FCN) have largely promoted the accuracy of automatic building segmentation from remote sensing images. However, compared with the deep-learning-based building segmentation methods that greatly benefit from data-driven feature learning, the building boundary vector representation generation techniques mainly rely on handcrafted features and high human intervention. These techniques continue to employ manual design and ignore the opportunity of using the rich feature information that can be learned from training data to directly generate vectorized boundary descriptions. Aiming to address this problem, we introduce PolygonCNN, a learnable end-to-end vector shape modeling framework for generating building outlines from aerial images. The framework first performs an FCN-like segmentation to extract initial building contours. Then, by encoding the vertices of the building polygons along with the pooled image features extracted from segmentation step, a modified PointNet is proposed to learn shape priors and predict a polygon vertex deformation to generate refined building vector results. Additionally, we propose 1) a simplify-and-densify sampling strategy to generate homogeneously sampled polygon with well-kept geometric signals for shape prior learning; and 2) a novel loss function for estimating shape similarity between building polygons with vastly different vertex numbers. The experiments on over 10,000 building samples verify that PolygonCNN can generate building vectors with higher vertex-based F1-score than the state-of-the-art method, and simultaneously well maintains the building segmentation accuracy achieved by the FCN-like model.
Article
The paper describes a unified automatic procedure for the detection of roof planes in gridded height data. The procedure exploits the Blake-Zisserman (BZ) model for segmentation in both 2D and 1D, and aims to detect, to model and to label roof planes. The BZ model relies on the minimization of a functional that depends on first- and second-order derivatives, free discontinuities and free gradient discontinuities. During the minimization, the relative strength of each competitor is controlled by a set of weight parameters. By finding the minimum of the approximated BZ functional, one obtains: (1) an approximation of the data that is smoothed solely within regions of homogeneous gradient, and (2) an explicit detection of the discontinuities and gradient discontinuities of the approximation. Firstly, input data is segmented using the 2D BZ. The maps of data and gradient discontinuities are used to isolate building candidates and planar patches (i.e. regions with homogeneous gradient) that correspond to roof planes. Connected regions that can not be considered as buildings are filtered according to both patch dimension and distribution of the directions of the normals to the boundary. The 1D BZ model is applied to the curvilinear coordinates of boundary points of building candidates in order to reduce the effect of data granularity when the normals are evaluated. In particular, corners are preserved and can be detected by means of gradient discontinuity. Lastly, a total least squares model is applied to estimate the parameters of the plane that best fits the points of each planar patch (orthogonal regression with planar model). Refinement of planar patches is performed by assigning those points that are close to the boundaries to the planar patch for which a given proximity measure assumes the smallest value. The proximity measure is defined to account for the variance of a fitting plane and a weighted distance of a point from the plane. The effectiveness of the proposed procedure is demonstrated by means of its application to urban digital surface models characterized by different spatial resolutions. Results are presented and discussed along with some promising developments.