Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Geometric-based approach for linking various building measurement data to a 3D city model

  • Yoshiki Ogawa ,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – review & editing

    ogawa@csis.u-tokyo.ac.jp

    Affiliation Center for Spatial Information Science, the University of Tokyo, Tokyo, Japan

  • Go Sato,

    Roles Data curation, Formal analysis, Investigation, Methodology, Resources, Software, Validation, Visualization, Writing – original draft

    Affiliation Department of Civil Engineering, the University of Tokyo, Tokyo, Japan

  • Yoshihide Sekimoto

    Roles Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Visualization, Writing – review & editing

    Affiliation Center for Spatial Information Science, the University of Tokyo, Tokyo, Japan

Abstract

Currently, the Ministry of Land, Infrastructure, Transport, and Tourism (Japan) is in the process of developing an open 3D city model known as PLATEAU. Abundant measurement data related to buildings, including maps produced by private companies and mobile mapping system point clouds, have been collected to enhance the value of the 3D city model. To achieve this, it is necessary to identify the buildings for which measurement data is available. In this study, we propose and evaluate an efficient matching method for various building measurement data, primarily using geometric properties. In Numazu city, PLATEAU IDs were assigned to 88,525 Zenrin buildings as part of a private map. The results indicate that 90.6% of the polygons were matched. For aerial images, 93.6% of the extracted buildings matched the PLATEAU buildings, although only 70.9% of the PLATEAU data was extracted from the images. Using the level of detail 1 and 2 models, 46 textured building files were created from the mobile mapping system point cloud. In addition, the cover ratio for the laser profiling point cloud was mostly greater than 40%, which was higher than that of the mobile mapping system.

1. Introduction

In recent years, numerous countries around the world have developed 3D city models. Wide area semantic-based 3D city models have been developed in 21 cities in nine countries worldwide [1] and the PLATEAU project aims to develop 3D city models as open data for cities throughout Japan [2]. These models are expected to be beneficial for urban activity monitoring, disaster prevention, city planning, and more. The value of these models can be further enhanced by incorporating information on buildings, such as textures and building use.

This study focuses on three types of information used to augment building data: 2D footprint data, aerial photographs, and 3D point clouds. In Japan, a private company produces 2D footprints, also known as residential maps, based on field surveys, measurements, and drawings, which include information on building names and uses. Aerial images, which contain information on the time-series changes of buildings and ceiling surfaces, are provided by the Geospatial Information Authority of Japan (GSI) and private companies. 3D point clouds, containing information on the color and shape of buildings, are measured by aerial laser surveying or specialized vehicles called mobile mapping system (MMS). Because these data were created independently from the 3D city model, it was necessary to link them to the 3D city model.

This study presents numerous technical advancements. Previous studies have evaluated 2D footprint similarity by utilizing the distance between representative points and overlapping areas to merge multiple Geographical Information System (GIS) databases that do not share the same geometry [3, 4]. These studies commonly cited the difficulty of matching one-to-many polygons as an issue. Furthermore, a study created a 3D buffer from the corresponding wall elements and extracted the point cloud [5] in the matching of 3D point clouds to models, but automating the entire point cloud matching process, including generating a 3D city model with wall textures and providing a WebAPI service, has yet to be achieved. Thus, it is essential to develop a WebAPI to facilitate the open utilization of the matching algorithm and to meet social needs and technological advancements.

Considering social needs and technical advancements, the research question of this study is to systematically organize an automatic linking method that can link various types of exterior measurement data, which are voluminous and diverse, to a 3D urban model.

To effectively utilize the three types of building measurement data (2D footprint, aerial images, and 3D point clouds) from various sources, we developed a matching method for reliable 3D building models (PLATEAU) and implemented it as a Web API. The matching was performed using three typical examples: residential maps, aerial images, and 3D point clouds (Fig 1).

thumbnail
Fig 1. Conceptual illustrations for the three types of matching targeted in this study.

https://doi.org/10.1371/journal.pone.0296445.g001

The remainder of this paper is organized as follows: Chapter 2 describes existing technologies and previous studies related to this study. Section 3 describes the datasets used in this study, and Section 4 describes the methodology used. Section 5 presents the results and discussion of the study, and Section 6 concludes this thesis and presents future research directions.

The WebAPI developed in this study is accessible at the following URL: https://github.com/Project-PLATEAU/UC22-008-Building-matching-WebAPI.

2. Related work

2.1 Wide area 3D city model

The development of wide-area 3D city models is being promoted in numerous countries worldwide. This study is based on Geography Markup Language (CityGML) 2.0, which defines the level of detail (LOD) concept for the representation of geographic features, such as buildings [6]. The definition of LOD in CityGML 2.0 is as follows: LOD1 defines a building shape by a series of prisms and does not represent the outer façade, whereas LOD2 represents detailed roof shapes and outer façades. For instance, the shape of an arcade in a shopping district was added to LOD2. Building openings, such as doors and windows, are added in LOD3, and the interior of the building, including rooms and interior installations such as desks and chairs, is represented in LOD4.

The 3DCityDB, a software package for handling CityGML, lists 21 cities in nine countries worldwide for which wide-area CityGML 3D models have been developed, including Helsinki, Singapore, and Tokyo [1]. The Helsinki 3D, a semantic-based model in the CityGML format, is an example of 3D city model development [7]. The model was utilized for comparative studies of urban energy consumption and the visualization of future development models. Another example of wide area 3D city models is "Virtual Singapore" [8], which covers the entire area of Singapore and employs the CityGML data format. In Japan, Project PLATEAU, a 3D city model in the CityGML format, is being developed, and approximately 200 cities, including Tokyo, are targeted for development by FY2023 [2]. The target number of cities to be developed is set at approximately 500 by FY2027, corresponding to approximately 30% of the total number of municipalities in Japan.

The integration of diverse datasets is critical in creating wide-area 3D city models. Seto et al. 2020 [9] successfully integrated and visualized complex and diverse datasets obtained in their study area, including aerial images, 3D building models, 3D point clouds, and data on people flow and traffic, while reducing data volume. However, data development for 3D point clouds is currently limited to a single location. In Sofia, Bulgaria, Dimitrov and Petrova-Antonova 2021 [10] integrated DSM/DEM (TIFF files), residential map data (point format), and building data (2D polygons). The study successfully created an LOD1 building model in the CityGML format without being hindered owing to topography.

2.2 2D footprint matching

Previous studies that integrated 2D footprint data include Tong et al. 2009 [3] and Ruiz-Lendínez et al. 2017 [4]. Tong et al. 2009 [3] employed polygons extracted from paper maps and QuickBird images as the GIS polygons, whose shapes did not match. Polygon similarity was assessed based on the distance between the representative points and the overlapping area. Similarly, Ruiz-Lendínez et al. 2017 [4] used data created by the Spanish National Geographic Institute (BCN25) and the Andalusian Institute of Statistical Mapping (MTA10) as two GIS polygons whose shapes did not match. They used a weighted linear combination of polygon similarity values obtained from seven criteria: the number of convex vertices, number of concave vertices, perimeter length, area, minimum second moment, Arkin Graph Area, and minimum bounding rectangle. Both studies cited the difficulty in handling one-to-many matching as a challenge in 2D polygon matching.

A tree structure is effective for locating 2D polygons in close proximity to each other. Creating an R-tree using PostGIS, a component of 3DCityDB, accelerates the spatial retrieval of 2D polygons [11]. Quadtree-based partitioning, such as Open Location Code [12], is a simpler spatial partitioning method compared to R-tree. Quadtrees possess advantages such as low overhead and applicability to distributed databases [13, 14].

Converting aerial images into 2D polygons is an instance segmentation problem in image recognition. State-of-the-art models for instance segmentation include Mask R-CNN [15], MS R-CNN [16], and RefineMask [17]. Super-resolution techniques are beneficial for object detection in aerial images. State-of-the-art models for super-resolution, such as SRGAN [18], ESRGAN [19], and SwinIR [20], are based on GAN techniques. In Japan, the Geospatial Information Authority of Japan (GSI) provided an open-source satellite image dataset, which Chen et al. 2023 [21] utilized to study the extraction of building shapes. However, the GSI dataset has some drawbacks, such as color differences and low resolution, which were addressed by enhancing the Mask R-CNN, color adjustment, and super-resolution techniques by extending ESRGAN. Because the extracted building geometry is independent of the existing geographic information database, a matching technique between 2D polygons is required to use the building information extracted using this method in combination with existing geographic information.

2.3 3D point cloud matching

Studies linking airborne laser scanner (ALS) point clouds to LOD1 models for mapping 3D point clouds to 3D urban models exist. Park and Guldmann 2019 [22] estimated building heights for the LOD1+ model (an extended white box model in which one building is represented by a combination of white boxes with multiple heights) by extracting only the point cloud of building roof surfaces from the ALS point cloud. Albeaik et al. 2017 [23] corrected a low-resolution noisy ALS point cloud to create an LOD1 building 3D model.

Building texture images based on airborne data were created in the following studies: Lee and Yang 2019 [24] utilized oblique images captured from aircraft to generate textured images of the sides of buildings to improve the quality of 3D urban models. Some studies that utilized oblique images to create 3D city models employed the structure from motion technique [25].

Several studies have created building texture images based on data from MMS. Yang 2019 [26] created a building texture image by developing an image acquisition and distortion correction system using MMS. Kelly et al. 2017 [27] and Femiani et al. 2018 [28] utilized Google Street View (GSV) images [29] to generate building-side images. Kelly et al. 2017 [27] extracted individual buildings from MMS images using edge scores, determined the representative colors of buildings using the most frequent values, and estimated the locations of structures, such as windows and doors, to produce a textured 3D city model for 1011 buildings in 37 blocks. Liu et al. 2020 [30] set up a loss function based on the symmetric rectangular shape of windows and doors for object detection in images of building sides. By contrast, Dai et al. 2021 [31] performed a study on images of suburban areas instead of urban areas. Femiani et al. 2018 [28] used angle information to crop buildings from GSV images. Tian and Wang 2023 [32] proposed a method to find the best match between the 2D photos captured at unknown distances and the actual object in the captured stereo images through perspective projection and model matching. The matching fitness function was optimized using a genetic algorithm. The experimental results were obtained using two different objects and the accuracy was found to be better for the industrially manufactured product (sunscreen) compared to the agricultural product (pear) with irregular shape variations. Scientific research addresses the thermal image mapping onto 3D models for visualization and analysis. Antón and Amaro-Mellado 2021 [33] developed an open-source software graphical method to produce 3D thermal data from infrared thermography (IRT) images for temperature visualization and subsequent analysis.

E. Oniga 2012 [34] proposed an algorithm for the semiautomatic texture generation based on color information, RGB values of every point captured by terrestrial laser scanning technology, and 3D surfaces defining building facades generated using commercial 3D software. The operator needed to define the limiting value, i.e., the minimum distance between a point and the closest surface. In this study, the threshold used to choose the points to be projected was set automatically. Beil et al. 2021 [5] coupled a 3D city model provided by CityGML with a point cloud. To link the MMS point cloud to CityGML, a 3D buffer was created from the corresponding CityGML elements, and the point cloud was extracted. However, Beil et al. 2021 [5] have not yet automated all point-cloud matching tasks, including the output of 3D city models with wall surface textures, and provided services such as WebAPI. Open3D [35] is an open-source software package that can be used for point-cloud extraction, and functions in Open3D can be used to select and save only points within a certain distance of the building in terms of the XY plane. However, using the functions included in Open3D, it is not possible to extract a point cloud based on the distance between a point and a surface, considering the Z coordinate.

In this study, 3D city models with wall surface textures were made from open MMS 3D point cloud data using only free and open-source software rather than using commercial 3D software. The texture was added to the 3D city model provided by CityGML. All point-cloud matching tasks were automated and the service was provided as WebAPI.

3. Dataset

3.1 3D city model (PLATEAU)

The 3D city model data used in this study were the PLATEAU data provided by the Ministry of Land, Infrastructure, Transport, and Tourism. This study utilized LOD1 and LOD2 data from Numazu city, and the survey year for the PLATEAU data in Numazu city was 2015.

3.2 2D footprint (residential maps)

2D footprint data is the map data that has attributes related to the name and use of each building based on field surveys, measurement information, and surveys. In this study, we used 2D polygon data sold by a private company (Zenrin Corporation), mainly utilized in Japan. Data updates are performed once every year, and the data employed in this study were obtained in FY2021. The residential maps were used in this study as a 69.9 MB polygon file in GPKG format.

3.3 Aerial images

In this study, building footprints extracted from the orthorectified image dataset provided by the Geospatial Information Authority of Japan (GSI) in the method by Chen et al. 2023 [21] were matched to the 3D city model. The orthorectified image dataset is a distortion-free image of aerial photographs captured by the GSI, which can overlap with various types of geospatial information by adding precise location information. The aerial images of Numazu City, with a resolution of 20 cm, used in this study were captured in 2010 and 2012. For the matching process in this study, we used a 2D GPKG polygon file (138.8 MB), from which polygons were extracted from the orthorectified image dataset of Numazu City.

3.4 3D point cloud

3D point clouds are produced by aerial laser surveying or special vehicles called MMS and contain information on the color and shape of buildings. The 3D point cloud data files were downloaded from the G-Spatial Information Center website [36]. The point cloud data were generated in 2019, and the area to be downloaded was selected using a 400 m x 300 m mesh. The 3D point cloud utilized in this study pertains to a single mesh near the northern exit of Numazu Station, and both the data (201.4 MB) obtained by an aerial laser survey (LP) and the data (3.2 GB) obtained by a mobile measuring vehicle (MMS) were employed.

4. Methodology

4.1 Processes implemented in this study

In this study, we implemented a 3D city model building ID matching and suggestion process for 2D footprints and a wall ID assignment process for each point that comprises a 3D point cloud. Fig 2 depicts the overall processing of 2D footprints and 3D point clouds.

thumbnail
Fig 2. Overall processing of 2D footprints and 3D point clouds.

https://doi.org/10.1371/journal.pone.0296445.g002

4.2 2D footprint matching

The matching results of the 2D footprints to the 3D city model were obtained based on the criterion that the distance between the centers of gravity should be within 10 m and the overlap ratio should be 40% or more. The overlap ratio is the intersection area of two polygons divided by the area of the smaller polygon. This criterion was set to enable the correct assignment or rejection of IDs for all data for one city block near the north exit of Numazu Station. By using this city block, it was confirmed that this criterion can handle the cases of one-to-many or one-to-none matching, which was considered difficult in previous studies. If no matching building is found in the aforementioned process, a building ID is suggested for a building with an area difference of 20% or less from the 2D polygons and located within 10 m of the center of gravity. If more than one matching buildings is found, the nearest one is chosen by comparing the distance between the centers of gravity.

Fig 3 presents conceptual illustrations of the matching based on this criterion. In the figure, (a1) shows an example in which the intersecting area exceeds 40% of the input 2D building polygons, and (a2) shows an example in which the intersecting area exceeds 40% of the building database polygons. By defining the overlap ratio as the intersection area of the two polygons divided by the area of the smaller polygon, matching was successful even when one polygon was larger than the other, as shown in (a2). Cases, where matching was not performed using this criterion, include (b1), where the polygons do not intersect and their centers of gravity are far apart, and (b2), where the area of intersection is less than 40% of the area of either of the two polygons.

thumbnail
Fig 3. Conceptual illustrations of matching based on distance and overlap ratio.

Red denotes input 2D building polygons, green denotes building DB polygons, and yellow represents intersections. [a1] An example of matching where the intersection exceeds 40% of the input 2D building polygon. [a2] An example where the intersection exceeds 40% of the building DB polygons and matches. [b1] Not matched because neither polygon intersects, and the center of gravity is extremely far. [b2] Not matched because the intersection area was less than 40% of the polygons in both polygons.

https://doi.org/10.1371/journal.pone.0296445.g003

4.3 3D point cloud matching

The cover ratio is calculated as the number of points close to the wall surface divided by the total surface area of the building, representing the proportion of the wall surface covered by the point cloud. After 1 m3 voxel sampling, the number of points within 1 m of the wall was computed. Buildings with high-quality wall images that met the condition of a cover ratio of at least 20% were chosen as targets for wall texturing. Points close to the wall surface were defined as those within 1 m of the wall surface, considering the sampling interval. A projective transformation of the wall surface was performed to generate a wall surface image from a 3D point cloud. Fig 4 illustrates an example of a point cloud downsampled in cubic units of 1 m per side and nearby buildings.

thumbnail
Fig 4. Example of point cloud downsampled in cubic units of 1 m per side and nearby buildings.

https://doi.org/10.1371/journal.pone.0296445.g004

Fig 5 presents the creation of a wall surface image from a 3D point cloud. To create the wall surface image from the 3D point cloud, we first gathered points near the wall surface in the xz plane by rotating around the z-axis, projecting to the xz plane, and interpolating to the x and z coordinates of the wall surface using nearest-neighbor interpolation. In LOD2, many nonvertical wall surfaces were present. However, by estimating the normal vector of the surface from the three points that were not on the same line and constituted the LOD2 wall surface, the plane to be projected onto was determined, corresponding to the xz plane in LOD1. The wall texture was created by performing the same rotational shift, projective transformation, and nearest-neighbor interpolation as in LOD1. Initially, the resolution of the wall image was set such that 1 cm corresponded to one pixel. When the long side length exceeded 512 cm, the longer side length and width were set to 512 pixels. The areas without points are shown in gray color.

thumbnail
Fig 5. Method for creating wall surface images from 3D point clouds.

[a] Projection transformation of a group of points on the wall surface (yellow) to create a textured image. [b] Delete points projected outside the surface and points whose distance from the surface is greater than the threshold value (*). [c] Nearest-neighbor interpolation. [d] Areas with no points are shown in gray.(*) The threshold is defined as the maximum distance between the wall and points inside the building that have the designated wall ID as the closest wall ID, preventing the wall from being imaged by the backside point cloud.

https://doi.org/10.1371/journal.pone.0296445.g005

The imaging function collects all the points in the vicinity of the building and chooses the points to be projected based on the texture mapping method option. When using "- all: all points are projected," all points near the building are used. When using "- nearest: the closest point to each surface is mapped," only the point nearest to the surface to be projected is used. When using "- smart: Auto-detect maximum depth," a threshold is set as the maximum distance between the wall and points inside the building that have the designated wall ID as the closest wall ID. Points closer to the wall than the threshold are selected as the mapping targets.

4.4 Integrated execution of matching in WebAPI

The WebAPI was constructed on an AWS EC2 r5a.large instance (with 16 GB memory, 2vCPUs, and a 2.5 GHz clock frequency). All three matching methods were executed, and their execution times were measured in the WebAPI environment. The WebAPI was designed to enable both command line interface (CUI) and graphical user interface access to the matching methods.

The created Web API consists of three endpoints, as listed in Table 1. The first endpoint is 2D footprint matching, which sends a query for 2D footprint data expressed in GeoJSON polygons and returns the corresponding 3D city model building data in GeoJSON format. Experimental results for 2D footprint matching were generated from this endpoint. The second endpoint is 3D point cloud matching. When a 3D point cloud data file in LAS format is sent as query data, a list of building IDs of the 3D city model corresponding to the point cloud and information necessary for cover ratio calculation are returned in 2D GeoJSON format. Experimental results for the cover ratio calculation were generated from this endpoint. The third endpoint is 3D texture mapping, which sends the 3D point cloud data and the ID of the building data of the 3D city model as query data and returns the 3D city model data with the texture generated from the point cloud data mapped to the building as a Wavefront OBJ format file. Experimental results on textured 3D city models were generated from this endpoint.

Further explanation on the WebAPI is accessible at the following URL: https://project-plateau.github.io/UC22-008-Building-matching-WebAPI/.

5. Results and discussion

5.1 2D footprint matching

The results of 2D footprints are listed in Table 2. Using the distance between the centers of gravity and overlap ratio, we matched building IDs for 90.6% of the footprints, rejected 8.2% (no matching), and classified 1.2% as footprints requiring visual judgment. In Ruiz-Lendínez et al. 2017 [4], 90.6% of the matches were smaller than the value of 97.9%. However, considering the differences in datasets used and the fact that the nonmatching footprints include successful cases of one-to-none matching, i.e., correct rejects, the method employed in this study is more accurate. Despite the simplicity and small number of features used, no significant difference existed in accuracy compared to Ruiz-Lendínez et al. 2017 [4].

The matching and suggestion results for the residential map polygons are shown in Fig 6, depicting all results within the ranges shown. Notably, the suggestion in the figure is for a 3D city model (PLATEAU) polygon that exhibits similar shapes in the northeast direction but is not matched owing to a low overlap ratio, highlighting the importance of incorporating a suggestion function.

thumbnail
Fig 6. Results of the matching and suggestion process for 2D footprints.

https://doi.org/10.1371/journal.pone.0296445.g006

The results for aerial image processing are listed in Table 3. Although the number of buildings extracted from aerial images was only 70.9% of the total number of polygons in the 3D city model (PLATEAU), 93.6% of the extracted buildings matched the PLATEAU building IDs, and 6.3% were successfully rejected (no matching). Fig 7 shows the process results, including matching and suggestion of polygons extracted from aerial photographs. The matching (black shaded line) and suggestion (green shaded line) results were all valid within the range shown in the figure, and the rejects corresponded to buildings that existed in aerial images but not in PLATEAU and buildings whose building outlines were misdetected. The former is effective in tracking time-series changes in buildings, whereas the latter is useful in detecting errors in the building contour extraction model.

thumbnail
Fig 7. Results of the matching and suggestion process for polygons extracted from aerial images.

https://doi.org/10.1371/journal.pone.0296445.g007

5.2 3D point cloud matching

Fig 8 shows the cover ratio calculation results for the MMS point cloud, whereas Fig 9 shows the cover ratio calculation results for the LP point cloud. The cover ratio of buildings located north-south near the center of the figure and close to the main street where the MMS traveled was generally over 20% for the MMS point cloud and over 40% for the LP point cloud measured from the sky. This suggests that the cover ratio difference was determined by the measurement method. Fig 10 shows the results of generating wall images for multiple buildings in the MMS point cloud, comparing the outcomes of wall image generation depending on the level of model detail. Comparing the LOD 1 and LOD 2 results, the LOD 2 building model more closely resembled the actual building, resulting in fewer gray pixels (no points) on the walls.

thumbnail
Fig 8. Result of the cover ratio calculation for the MMS point cloud.

Green: Buildings with a cover ratio of 40% or more. Yellow: Buildings with a cover ratio between 20% and 40%. Red: Buildings with a cover ratio of less than 20%.

https://doi.org/10.1371/journal.pone.0296445.g008

thumbnail
Fig 9. Result of the cover ratio calculation for the LP point cloud.

Green: Buildings with a cover ratio of 40% or more. Yellow: Buildings with a cover ratio between 20% and 40%. Red: Buildings with less than 20% cover ratio.

https://doi.org/10.1371/journal.pone.0296445.g009

thumbnail
Fig 10. Comparison of wall surface image generation results by level of detail.

(Upper) LOD1. (Lower) LOD2.

https://doi.org/10.1371/journal.pone.0296445.g010

When generating wall images for LOD2 buildings, projecting only points with the corresponding wall ID onto the wall surface resulted in image generation issues owing to geometry overlap. For such buildings, the optimal results were obtained by projecting all building points that were within the threshold distance from the surface. Fig 11 compares the texture changes for a building with a complex LOD2 geometry as a result of changes in the set of projected points, as well as the texture mapping results for an LOD2 building with different mapping settings. Even with the improved thresholded projection algorithm, projecting using the LP point cloud resulted in the majority of the wall surface being covered in gray owing to the small number of points.

thumbnail
Fig 11. Texture mapping results for an LOD2 building with different mapping settings.

[a] LOD2 building model and MMS point cloud. [b] "all": All points were projected. [c] "nearest": The point closest to each surface is mapped. [d] "smart": Maximum depth is auto-detected.

https://doi.org/10.1371/journal.pone.0296445.g011

5.3 Integrated execution of matching in WebAPI

Table 4 summarizes the matching results of various measurement data in the WebAPI. For the 2D footprints (residential map), 90.6% of the 88,525 polygons in Numazu city matched with building IDs, whereas 8.2% were rejected (no matching) and 1.2% required visual judgment (classified as gray). Limiting the visual judgment to 1.2% can significantly reduce the time required for confirmation in practical terms. The time required to match the entire city was 104 s, considered practical in terms of execution speed.

thumbnail
Table 4. Summary of the matching results of the various measurement data used in this study.

https://doi.org/10.1371/journal.pone.0296445.t004

For the aerial photographs, only 70.9% of the PLATEAU buildings were extracted in the preprocessing step for matching. Out of the 80,645 extracted polygons, 93.6% had building IDs matched, whereas 6.3% were rejected (no matching) owing to various reasons such as incorrect extraction and new construction or demolition of buildings. Visual confirmation was required for both suggestions and rejections. The matching time for the entire city was 212 s, and the matching speed for the aerial photographs was considered practical.

The cover ratio of 3D point cloud data, defined as the ratio of point clouds to building surface, was calculated. Using the MMS points, a textured 3D city model was automatically generated for 46 buildings, with a cover ratio of more than 20% achieved successfully. The coverage calculation for the MMS point cloud required 92.5 s, whereas wall ID estimation and wall image creation required 845 s, with a total of 16 min for the two tasks.

6. Conclusions

This study aimed to effectively utilize various types of building measurement data from different sources by developing an efficient matching method for various measurement data on buildings using mainly geometric properties and implementing it as a Web API.

Specifically, for the 2D building polygons, we defined the overlap ratio between the 2D polygons and applied it to 88,525 residential buildings in Numazu City. Moreover, the developed 2D matching method was applied to aerial images to perform super-resolution processing, building extraction processing, and matching of the extracted building polygons. Although the number of buildings extracted from the aerial photographs was only 70.9% of the total number of PLATEAU buildings, 93.6% of the extracted buildings matched the PLATEAU building IDs, which holds significant practical potential.

The 3D point cloud data were linked to 3D geometries such as PLATEAU LOD1 and LOD2, and a textured 3D city model was automatically created for 46 buildings. Although the processing time was longer than that for 2D data, the cover ratio, defined as the percentage of point clouds on the building surface, was generally over 20% for MMS measured along roads and over 40% for LP measured from the sky. This clarified the coverage status depending on the measurement method.

In this study, 3D point cloud data and aerial photographs, which are generally available as open data from national and local governments, were matched to a 3D building model. In contrast to GSV, which was widely used to create building textures in previous studies, the building texture information source utilized in this study was open data. Therefore, a 3D urban model with textures created through this approach could be released to the public and used in various fields in the future. Consequently, the textured 3D city model created by this method could be publicly released and utilized in numerous fields.

Finally, we discuss the limitations of this study. One limitation of matching 2D polygons is that it handles cases in which polygons representing the same building are initially parallel-shifted and have a small or nonexistent overlap. In the proposed method, the suggestion process linked polygons if the distance between the centers of gravity was less than 10 m and the difference in area was less than 20%. However, in the case of large parallel shift distances, the criterion of a distance between the centers of gravity of 10 m or less was not met.

Aerial images provide valuable information on the roofs of buildings, which is difficult to obtain from ground-based observations. Based on the results of the 2D polygon matching proposed in this study, ways to increase the information on roofs in 3D city models must be investigated.

One limitation of point-cloud data matching is the use of LP point clouds. Although the LP point cloud has the advantage of high coverage compared to the MMS point cloud, ways to use the public LP point cloud to enhance the information of the 3D urban model must be considered.

Acknowledgments

We would like to express our gratitude to Mr. Omata from the Center for Spatial Information Science, University of Tokyo, and Mr. Endo from the Association for Promotion of Infrastructure Geospatial Information Distribution for their invaluable assistance in analyzing the data in this study.

References

  1. 1. Kolbe, T.H., 2023. 3DCityDB in action [Online]. https://www.3dcitydb.org/3dcitydb/3dcitydb-in-action/.
  2. 2. Ministry of Land, Infrastructure, Transport and Tourism, 2023. Plateau [Online]. https://www.mlit.go.jp/report/press/content/001602513.pdf.
  3. 3. Tong X., Shi W., Deng S., 2009. A probability-based multi-measure feature matching method in map conflation. In: Int. J. Remote Sens. 30, 5453–5472.
  4. 4. Ruiz-Lendínez J.J., Ureña-Cámara M.A., Ariza-López F.J., 2017. A polygon and point-based approach to matching geospatial features. In: ISPRS Int. J. Geo Inf. 6, 399.
  5. 5. Beil, C., Kutzner, T., Schwab, B., Willenborg, B., Gawronski, A., Kolbe, T.H., 2021. Integration of 3D point clouds with semantic 3D city models–providing semantic information beyond classification. In: ISPRS Ann. Photogramm. Remote Sens. Spatial Inf. Sci. VIII–4/W2, 105–112.
  6. 6. Open Geospatial Consortium, 2012. OGC City Geography Markup Language (CityGML) Encoding Standard.
  7. 7. Ruohomäki, T., Airaksinen, E., Huuska, P., Kesäniemi, O., Martikka, M., Suomisto, J., 2018. Smart city platform enabling digital twin. In: International Conference on Intelligent Systems (IS) pp. 155–161.
  8. 8. National Research Foundation, Government of Singapore, 2021. Virtual Singapore [Online]. https://www.nrf.gov.sg/programmes/virtual-singapore.
  9. 9. Seto, T., Sekimoto, Y., Asahi, K., Endo, T., 2020. Constructing a digital city on a web-3D platform: Simultaneous and consistent generation of metadata and tile data from a multi-source raw dataset. In: Proceedings of the 3rd ACM SIGSPATIAL International Workshop on Advances in Resilient and Intelligent Cities pp. 1–9.
  10. 10. Dimitrov, H., Petrova-Antonova, D., 2021. 3D city model as a first step towards the digital twin of Sofia city. In: Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. XLIII–B4, 23–30.
  11. 11. Open Source Geospatial Foundation, 2012. Spatial indexing [Online]. http://postgis.net/workshops/postgis-intro/indexing.html.
  12. 12. Google; 2021. Open location code [Online]. https://github.com/google/open-location-code.
  13. 13. Lee, K., Ganti, R.K., Srivatsa, M., Liu, L., 2014. Efficient spatial query processing for big data. In: Proceedings of the 22nd ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems pp. 469–472.
  14. 14. Liu, J., Li, H., Gao, Y., Yu, H., Jiang, D., 2014. A hash-based index for spatial data management in distributed memory, In: 22Nd international conference on geoinformatics pp. 1–4.
  15. 15. He, K., Gkioxari, G., Dollár, P., Girshick, R., 2017. Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision 2961–2969.
  16. 16. Huang, Z., Huang, L., Gong, Y., Huang, C., Wang, X., 2019. Mask scoring r-cnn. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition pp. 6402–6411.
  17. 17. Zhang, G., Lu, X., Tan, J., Li, J., Zhang, Z., Li, Q., et al., 2021. Refinemask: Towards high-quality instance segmentation with fine-grained features. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 6857–6865.
  18. 18. Ledig, C. et al., 2017. Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition pp. 105–114.
  19. 19. Wang, X. et al., 2019. Esrgan: enhanced super-resolution generative adversarial networks. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops pp. 63–79.
  20. 20. Liang, J., Cao, J., Sun, G., Zhang, K., Van Gool, L., Timofte, R., 2021. Swinir: Image restoration using swin transformer. In: Proceedings of the IEEE/CVF International Conference on Computer Vision pp. 1833–1844.
  21. 21. Chen S., Ogawa Y., Zhao C., Sekimoto Y., 2023. Large-scale individual building extraction from open-source satellite imagery via super-resolution-based instance segmentation approach. In: ISPRS J. Photogramm 195, 129–152.
  22. 22. Park Y., Guldmann J.M., 2019. Creating 3D city models with building footprints and LIDAR point cloud classification: A machine learning approach. In: Comput. Environ. Urban Syst. 75, 76–89.
  23. 23. Albeaik, S., Alrished, M., Aldawood, S., 2017. Alsubaiee, S., Alfaris, A. Virtual cities: 3d urban modeling from low-resolution lidar data. In: Proceedings of the 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, pp. 1–4.
  24. 24. Lee J., Yang B., 2019. Developing an optimized texture mapping for photorealistic 3D buildings. In: Trans. GIS 23, 1–21.
  25. 25. Li S., Xiao X., Guo B., Zhang L., 2020. A novel OpenMVS-based texture reconstruction method based on the fully automatic plane segmentation for 3D mesh models. In: Remote Sens. 12, 3908.
  26. 26. Yang B., 2019. Developing a mobile mapping system for 3D GIS and smart city planning. In: Sustainability 11, 3713.
  27. 27. Kelly T., Femiani J., Wonka P., Mitra N.J., 2017. BigSUR: Large-scale structured urban reconstruction. In: ACM Trans. Graph 36, 1–16.
  28. 28. Femiani, J., Para, W.R., Mitra, N., Wonka, P., 2018. Facade Segmentation in the Wild. In: arXiv Preprint ArXiv:1805.08634.
  29. 29. Anguelov D., Dulong C., Filip D., Frueh C., Lafon S., Lyon R., et al., 2010. Google Street View: Capturing the world at street level. Computer 43, 32–38.
  30. 30. Liu H., Xu Y., Zhang J., Zhu J., Li Y., Hoi S.C.H., 2020. DeepFacade: A deep learning approach to facade parsing with symmetric loss. In: IEEE Trans. Multimedia 22, 3153–3165.
  31. 31. Dai M., Ward W.O.C., Meyers G., 2021. Densley Tingley D D, Mayfield M. Residential building facade segmentation in the urban environment. In: Build. Environ. 199, 107921.
  32. 32. Tian H., Wang J., 2023. A Variable Photo-Model Method for Object Pose and Size Estimation with Stereo Vision in a Complex Home Scene. In: Sensors 23(15), 6924. pmid:37571707
  33. 33. Antón D., Amaro-Mellado J.L., 2021. Engineering graphics for thermal assessment: 3D thermal data visualization based on infrared thermography, GIS, and 3D point cloud processing software. In: Symmetry 13(2), 335.
  34. 34. Oniga E., 2012. A new approach for the semi-automatic texture generation of the buildings facades, from terrestrial laser scanner data. In: The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences 39, 161–166.
  35. 35. Zhou, Q.Y., Park, J., Koltun, V., 2018. Open3D: A Modern Library for 3D Data Processing. arXiv Preprint ArXiv:1801.09847.
  36. 36. Association for Promotion of Infrastructure Geospatial Information Distribution; 2021. VIRTUAL Shizuoka, Shizuoka Prefecture, Southeast of Mt. Fuji, Izu area point cloud [Online]. https://www.geospatial.jp/ckan/dataset/shizuoka-19-20-pointcloud.