Next Article in Journal
Macronutrients, Amino and Fatty Acid Composition, Elements, and Toxins in High-Protein Powders of Crickets, Arthrospira, Single Cell Protein, Potato, and Rice as Potential Ingredients in Fermented Food Products
Previous Article in Journal
Vertical Electrical Sounding (VES) Technique to Map Potential Aquifers of the Guigou Plain (Middle Atlas, Morocco): Hydrogeological Implications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Crack Segmentation on Earthen Heritage Site Surfaces

School of Information Science and Technology, Northwest University, Xi’an 710127, China
*
Authors to whom correspondence should be addressed.
Submission received: 15 November 2022 / Revised: 8 December 2022 / Accepted: 12 December 2022 / Published: 14 December 2022
(This article belongs to the Topic AI Enhanced Civil Infrastructure Safety)

Abstract

:
Earthen heritage sites are historical relics left by ancient human activity, with earthen as the primary building material, and have significant historical, scientific, and artistic value. However, many sites have experienced extensive deterioration caused by environmental forces and human factors. A crack is a kind of typical damage to the walls of earthen heritage sites. Studies of the crack-formation process can effectively predict trends in damage, which will play a critical role in the maintenance of earthen heritage sites. This study is the first of its kind to propose a deep learning method to study the cracks on earthen heritage sites at the pixel-level, adopt the idea of transfer learning, and employ a mixed-crack image dataset for training three deep learning models. The precision, recall, IoU, and F1 metrics were used to evaluate the performance of the trained models. The experimental results showed that FPN-vgg16 appeared to have the highest level of applicability to detect cracks on earthen heritage sites among all networks, due to the highest F1 score of 84.40% and the highest IoU score of 73.11%. The results illustrated that the proposed method in this paper can effectively be used to analyze the rammed earth surface crack images, with great potential in related research fields.

1. Introduction

Earthen heritage sites are historical relics with earthen as the main building material, left by ancient human activities, which are typically found in dryland regions because of the wide availability of earthen materials and favorable environmental conditions. Rammed earth has been widely used to build dwellings and temples in ancient Europe and Asia. In China, these earthen heritage sites are relatively concentrated in the arid regions of Northwest China, mainly including ancient cities, the Great Wall, passes, beacon towers, earthen pagodas, and tombs. Earthen heritage constitutes 10% of sites on the World Heritage List [1] and has important historical, scientific, and artistic value. However, many sites have experienced extensive deterioration mediated by environmental and human activities, which threatens to reduce their value [2,3]. Many heritage sites composed of adobe materials the earthen structures with significant value remain in the Gobi Desert and other desert regions in Northwest China [4]. Suoyang Ancient City is a typical rammed earth heritage site, which was built during the Han (206BCE–220CE) and Tang (618–907CE) dynasties [5]. It is located in Gansu Province, northwest China, along the ancient Silk Road (Figure 1a). This ancient city site, situated in arid areas, has all kinds of harsh environmental conditions, such as strong wind, sandstorms, and drought. The damage to the walls of earthen heritage sites caused by the harsh environment include scouring, flaking, and cracking (Figure 1c–e). Many investigations have been conducted at the Suoyang Ancient City site and some typical damage was spotted. This study aimed to find a long-term method for researching the deterioration of earthen heritage sites by studying the cracks on the rammed earth surfaces of the sites.
Cracks comprise the damage to the main structures of earthen heritage sites. Detecting the crack formation process can effectively predict the trend in the damage to earthen sites and thus help officials adopt appropriate methods to maintain the sites’ value. The traditional method of crack detection was measuring the length and width of the cracks through visual inspection by well-trained cultural relic workers with the help of rulers. Although the current state of the crack can be recorded via manual inspection, long-term observation of the crack and the objectivity and reliability of the crack observation cannot be effectively guaranteed. The replacement of observers will lead to the instability of observations. Therefore, many researchers focus on developing automated and efficient crack-detection techniques for the long-term monitoring of earthen heritage sites.
The application of computer vision-based crack detection technology for modern buildings, such as cement roads, brick walls, and bridges, has been widely studied. Threshold segmentation [6,7], morphology [8], wavelet transform [9], and filter-based algorithms [7,10,11] have been applied in the field of crack detection. Wang et al. [12] summarized traditional computer vision-based crack detection and subdivided them into four categories, namely an integrated algorithm, morphological approach, percolation-based method, and practical technique, and then analyzed and compared the experimental results of the four types of methods. Using these methods to analyze a photo of the building surface, one can identify the cracks, locate the cracks, and measure the cracks’ width, length, and orientation. By adjusting the parameters of these algorithms, high inspection detection accuracy can be achieved. The traditional computer vision-based crack detection algorithm usually focuses on the understanding and analysis of the partitive features of the image, and parameters are designed for a specific dataset. However, when there is interference, such as stains, potholes, etc., the partitive image will be difficult to segment. The main characteristics of rammed earth heritage are that cracks are distributed on the irregular surface and mixed with irregular textures, such as shadows and holes. It is difficult for computer vision methods to segment cracks.
In order to overcome the drawbacks of traditional computer vision-based algorithms, some studies have suggested applying machine learning to image processing tasks. Lecun et al. first proposed that Convolutional Neural Networks (CNN) could be applied to handwritten character recognition [13]. More studies have proposed image classification [14,15,16], object detection [17,18,19], semantic segmentation [20,21,22], and crack detection [23,24,25] tasks. Instead of extracting features, Convolutional Neural Networks can automatically learn the features of an image and identify cracks from images. Deep learning semantic segmentation models have been extensively studied for crack detection in modern buildings with positive results.
Classical machine learning algorithms, such as Support Vector Machines [26] and Random Forests [27], require the manual extraction of crack features. Even though this sort of algorithm relies heavily on manual extraction, they produce some fruitful study findings. The accuracy of the algorithm will be diminished to varying degrees due to the variations in external environmental factors, such as background noises and lighting conditions, when extracting manually. These factors may lead to unexpected practical problems and limitations in the application of classical machine learning algorithms.
As a branch of machine learning, the architecture of deep learning can extract features related to it and thus can effectively avoid the limitations existing in traditional algorithms. Research on crack detection technology based on deep learning, including image classification, object recognition, and semantic segmentation, has been proposed. Image classification algorithms can automatically classify the images into two categories, crack images and non-crack images. Object recognition algorithms can find and locate objects that contain cracks in images and use bounding boxes to identify and locate them.
Semantic segmentation is basic research in the field of computer vision processing and has been widely used in many fields, such as autonomous driving, medical image analysis, etc. Compared with image classification and object recognition, semantic segmentation is more accurate in positioning, and its positioning output can be distinguished using different colors. It can identify the width and length of cracks at the pixel-level. Obviously, in crack detection for earthen heritage sites, it is important not only to detect whether there are cracks on the surface but also determine the size of the cracks, which is of great significance in studying the causes and development of the cracks and proposing scientific protection plans.
Ni et al. [28] used GoogLeNet and ResNet classification architecture to find crack patches, employed Otsu’s thresholding to segment crack patches, and adopted median filtering to eliminate the influence. Kang et al. [29] proposed the faster region proposal convolutional neural network (Faster R-CNN) algorithm to detect crack regions. Wei et al. [30] used Mask R-CNN to perform crack segmentation at the pixel-level. Several studies have employed two settings: first, the classification architecture was used to locate crack patches, and then, the crack map was extracted at the pixel-level. The approach above is known as Hybrid Semantic Segmentation.
Another method was performed with encoder-decoder structures without identifying a crack patch or crack region beforehand. A typical encoder-decoder structure consists of an encoder network and a corresponding decoder network. The encoder network has a backbone architecture that contains convolution, pooling, and activation layers. The role of the decoder network is to map the low-resolution feature maps to full-resolution feature maps for pixel-wise classification.
Several crack segmentation methods with encoder-decoder structures have been developed to detect surface cracks, such as U-Net [31], SegNet [32,33], LinkNet [34], ResNet [35], and several variants [36,37]. In the semantic segmentation setting, the key to achieving improved accuracy is to provide the contextual information flow in architecture [38].
U-Net was initially released in the context of medical image analysis for biomedical image segmentation [31]. In several studies, the U-Net architecture has been proposed for crack detection. Sizyakin et al. [39] used morphological filtering to improve the binary map with localized cracks from the U-Net neural network. Lau et al. [40] put forward a U-Net-based architecture that replaces the encoder with a pre-trained ResNet-34 backbone. Lin et al. [41] advanced a U-Net architecture that employs the Full Attention Strategy, which comprises synthesis of the attention mechanism and the outputs from each encoding layer in a skip connection. Cao et al. [42] used the squeeze-and-excite module to empower U-Net, which they call SE-U-Net. This architecture can adaptively learn the shallow information matched with deep features in the skip path. The preceding studies demonstrated its superiority over the methods based on traditional machine learning algorithms.
SegNet [32] is made up of an encoder network and a corresponding decoder network. The encoder network architecture is topologically identical to the VGG16 [43] network with 13 convolutional layers. Song et al. [44] deployed SegNet architecture to segment structure surface cracks. By using SegNet as a reference, Chen et al. [45] recommended an encoder-decoder structural model for inspecting concrete pavement, asphalt pavement, and bridge deck cracks. Zou et al. [46] proposed the DeepCrack network based on the encoder-decoder architecture of SegNet and pairwisely fused the convolutional features that are generated in the encoder network and decoder networks. LinkNet employs an improved codec network architecture, and the input of each encoder layer is also bypassed by the output of its corresponding decoder [34]. This method can recover the spatial information lost by downsampling operations in the encoder and reduces the processing time. Training deeper neural networks is more challenging. He et al. presented a residual learning framework to make it simpler to train networks [35]. ResNet, which has a depth of up to 152 layers, is less complex than VGG. Combining the Residual connections and Inception architecture [47], Szegedy et al. [36] proposed that Inception-ResNet improves training speed with low total parameter counts and computational costs. Loverdos et al. [48] evaluated the effectiveness of several deep learning networks (U-Net, DeepLabV3+, U-Net (SM), LinkNet (SM), and FPN (SM)) in brick segmentation and crack detection with masonry walls. The analysis results showed that machine learning is superior to image-processing techniques for segmentation. Deep learning demonstrates enormous potential in crack detection.
However, there is not much research being done on the application of convolutional neural network models for the detection of cracks in earthen heritage sites because deep learning models require huge amounts of training data to make them work, and the adoption of deep learning techniques relies significantly on the volume of data available. The complexity of building a deep learning model stems from the difficulty in obtaining a significant number of cracked photographs of earthen heritage sites.
This study focused on contributing to this growing area of research by exploring an effective and efficient method for crack detection in earthen heritage sites. By evaluating different network performances, pixel-level crack segmentation models were recommended for the study of earthen heritage site surfaces. It also aimed to establish a hybrid dataset. Different open-source image datasets were mixed, and then, a hybrid dataset was constructed. The hybrid dataset combines the images of cracks on the surface of cement pavement, asphalt pavement, and earthen heritage sites.
In this paper, a method for crack detection in earthen heritage sites is proposed. Mixed datasets were used to evaluate different network performances by using key metrics and manually labeled ground truth for evaluation purposes.
The rest of the paper is structured as follows. Section 2 briefly presents the mixed datasets and proposed network. Section 3 presents experimental results. The results are analyzed and discussed in Section 4, and Section 5 draws some conclusions and presents some hints for future work.

2. Materials and Methods

2.1. Overview

This study evaluated the reliability and precision of the application of different methods to cases of cracks in the earthen surfaces of the Suoyang Ancient City and proposed a deep learning method that can effectively detect and predict rammed earthen heritage cracks. The workflow of the proposed method approach is illustrated in Figure 2.
Step1: Gather open-source crack image datasets, which include road cracks with occlusions and shadow interference. Every image has a manually annotated binary ground truth image.
Step2: Take some photographs of the crack on the surface of the Suoyang Ancient City wall. These photos were cropped to a uniform size, and 177 photos were selected and manually annotated to generate a dataset.
Step3: Combine the open-source crack dataset and the Suoyang Ancient City crack dataset, training 3 models with the mixed crack dataset to optimize model parameters.
Step4: Analyze results using key metrics and predictions, devoted to the segmentation of cracks on the surface of earthen heritage sites.

2.2. Image Dataset

Many open-source crack image datasets were widely used, such as CrackTree206 [49], Crack500 [50], and DeepCrack [51], etc. The CrackTree206 dataset contains 206 800 × 600 pixel photos of road cracks, with occlusions and shadow interference. It was widely used because it was released early with many samples. The Crack500 dataset contains 500 images with a resolution of around 2000 × 1500. Each image is cropped into 16 non-overlapping image regions and saved as 3368 crack images, with 1896 images in the training data, 1124 images in the testing data, and 348 images in the validation data. The DeepCrack dataset contains 537 crack images with three textures and two scenes, where the training data contains 300 images and the test data contains 237 images. All above datasets were manually annotated and generated binary ground truth images.
We collected many images of the cracks on the Suoyang Ancient City wall surface and removed some images with excessive interference. A dataset, named SYCrack, including 177 images with a uniform size of 256 × 256 pixels images, was generated. This dataset contains two categories: images with cracks and images without cracks. All of these images were manually annotated and rendered as binary ground truth images. Then, 80% of these images were chosen as training datasets, with 20% as testing datasets.
To improve the robustness of models, CrackTree206, Crack500, Deep Crack, and SYCrack were merged to form a Mixed Crack Dataset (MCD). The MCD, with a total of 2538 raw images and annotated images, contains concrete cracks, asphalt cracks, and earthen cracks.

2.3. Proposed Network

The semantic segmentation network consists of an encoder and a decoder, and different pre-trained Convolutional Neural Network architectures are used as backbones. As stated before, related works show that U-net, FCN, and Linknet have outstanding performance in image segmentation, especially in detecting cracks in some specific buildings. Their application, however, is limited only to images of those buildings. A deep learning network that was trained on a specific material dataset cannot achieve the same accuracy with different materials [52]. This means that the model trained on the cement road datasets is not necessarily suitable for the cracks in walls of earthen heritage sites. Özgenel et al. [53] proposed that a pre-trained CNN network can extract the low-level features of cracks. Therefore, with only the requirement for a small number of training samples and fast convergence networks, models can be achieved for new materials.

2.4. Transfer Learning

Deep learning requires enough data to complete the model training. The current dataset of earthen cracks cannot meet the needs of model training, and thus, the transfer learning method was adopted.
Transfer learning involves making the best possible use of the knowledge in the annotated domain to find the similarity of the target problem to assist in knowledge acquisition and learning in the target field. The transfer learning task, based on similarity, applies the model trained in the source domain to the target domain. Transfer learning can solve the insufficiency of high-quality training data in the target domain, and even in the case of a small amount of data, it can also solve the problem of model training.
In this research, we adopted the concept of transfer learning and used the road cracks similar to the surface cracks of earthen heritage sites to combine the pictures of the crack images of the earthen heritage sites for deep learning model training. This was verified using the Suoyang Ancient City crack image testing dataset to compare the deep learning models and select the best the crack identification model.

3. Verification and Results

3.1. Experimental Environment

All models were created in the Python 3.8 environment using the Tensorflow framework, and training and testing were carried out in the following experimental setting. Table 1 and Table 2 detail the software and hardware environments, respectively.

3.2. Experimental Data

The open-source dataset and crack dataset of the Suoyang Ancient City were mixed. The mixed crack data contain 2538 raw images and annotated ground truth images. Figure 3 shows samples of the training dataset.

3.3. Analysis of the Pre-Trained Deep Learning Models

The deep neural network model architectures used in this paper were Unet, Linknet, and FPN, with four backbones for each architecture, and all backbones have pre-trained weights on ImageNet [54]. The number of parameters is a performance indicator of the model, which means the requirement for computing resources. The larger number of parameters, the more storage resources and computing time required. Table 3 shows the statistics of the total number of parameters for each model.
As can be seen from Table 3, the parameters of FPN were the fewest, indicating a simple model structure and a lower memory footprint. The number of model parameters varied by more than three times due to different backbone networks.

3.4. Training Configuration

All models were trained for 100 epochs until the F1 score on the validation set stopped increasing. The data were supplied to the network with a batch size of 4. The optimization algorithm employed was Adam, which is efficient for gradient descent and requires less memory. The learning rate was set as 0.001. The Binary Cross-Entropy loss function (see Equation (1)) was implemented here, as the function works well for classification objectives, such as segmentation at the pixel-level [55]. Binary Cross-Entropy is defined as:
L B C E y , y ^ = ( y l o g y ^ + 1 y l o g 1 y ^ )
where y is the ground truth and y ^ is the prediction. Additionally, y can have values of 0 (background) or 1 (crack), whereas y ^ can have values between 0 and 1.
The evaluation of the network is based on the values of Precision, Recall, IoU, and F1 score. These metrics are given as:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1   s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
I o U = G r o u n d T r u t h P r e d i c t i o n G r o u n d T r u t h P r e d i c t i o n
The T P , F P , T N , and F N terms are defined as follows:
  • T r u e   P o s i t i v e   T P : the pixel is a crack (Positive) and is identified as a crack (True);
  • F a l s e   P o s i t i v e   F P : the pixel is intact (Positive) and is identified as a crack (False);
  • T r u e   N e g a t i v e   T N : the pixel is intact (Negative) and is identified as intact (True);
  • F a l s e   N e g a t i v e   F N : the pixel is a crack (Negative) and is identified as intact (False).
Precision and Recall are the two most common metrics. Additionally, they form the basis of the F1 score. Precision is the percentage of correct positive predictions versus the total number of positive predictions. Recall indicates the completeness of the positive predictions. Normally, poor precision is caused by high recall scores and vice versa. The F1 score combines the accuracy and recall metrics and is defined as the harmonic mean of precision and recall.
The F1 score is a useful model performance metric when working on classification models with an imbalanced dataset. The intersection over Union (IoU) is a metric used to measure the accuracy of an annotation based on a task. The F1 score and IoU score were used as the main indicators to evaluate the model.

4. Results and Discussion

All models were trained for 100 epochs with a batch size of 4 for all models. Table 4 shows the metrics presented for the validation set.
Metrics of U-Net, As shown in Figure 4, Precision achieved the highest score at epoch 36, but the Recall score was only 40.38%, which means more cracked pixels were classified as intact. After completing training based on 100 epochs, the Precision score was 88.08%, but the Recall value was only 74.59%. Metrics of Linknet and FPN were also analyzed, shown in Figure 5 and Figure 6, revealing similar results. Furthermore, the authors checked the predicted outputs, as shown in Figure 7, the crack annotation was incomplete for the predicted result and there was noise in the background. In other words, neither the Precision nor Recall score achieved good results. Moreover, the IoU value was 67.43%, which also shows that the prediction result did not match the ground truth very well. Among the results predicted by U-Net-densenet201, U-Net-resnet152, FPN-resnet152, and FPN-inceptionv3 models, some cracked pixels were classified as intact, which indicates failure in detecting some cracked pixels. In the results predicted by the Linknet-densenet201, Linknet-inceptionv3, FPN-densenet201 models, there was too much noise in the background.
U-Net-vgg16, Linknet-vgg16, and FPN-vgg16 all performed well without noise in the background. As can be seen in Figure 4, Figure 5 and Figure 6, the results of completing training based on 100 epochs had very close F1 scores. According to Table 4, FPN-vgg16 obtained the highest F1 score of 84.40% and the highest IoU score of 73.11%, which indicates the highest level of applicability in crack detection of earthen heritage sites among all networks in this study. FPN-vgg16 demonstrated unexpected performance and has fewer training parameters than other networks when F1 and IoU scores are compared and the prediction result image is analyzed.

5. Conclusions and Future Work

5.1. Conclusions

This paper proposes a method for crack detection in earthen heritage sites by evaluating the performances of different networks. Pixel-level crack segmentation models were recommended to detect cracks on earthen heritage site surfaces. The three models evaluated in this study are the most promising crack detection networks, and the four backbones show different abilities in prediction. Through the comparison of F1 and IoU scores and the analysis of the prediction result picture, FPN-vgg16 shows surprising performance and has fewer training parameters than other networks.
We collected some crack pictures at the site of Suoyang City, and cropped, screened, and annotated 177 pictures of cracks. A hybrid dataset, a mixture of different opensource image datasets, was constructed, which combines the images of cracks on the surface of cement pavement, asphalt pavement, and earthen heritage site surfaces, which thus solves the problem of a lack of data on cracks on earthen heritage site surfaces. To the best of the authors’ knowledge, this study is the first to propose pixel-level crack segmentation for surface cracks in earthen heritage site surfaces.
Monitoring the change in cracks on earthen heritage site surfaces is a long-term research topic. The crack detection method that we proposed can help conservators automatically and effectively extract crack information.

5.2. Future Work

According to the statistics generated from the experiments, the research methods and transfer learning ideas proposed in this paper have been proven to be effective. The models can only segment rammed earthen surface cracks, but the position, width, length, and depth of cracks are still uncertain, and thus, further optimization is needed. Moreover, the dataset of cracks on rammed earthen heritage sites is still small. More in-depth work is needed.
  • Since the image data of surface cracks in earthen heritage sites is still relatively small, there are few studies on the deep learning of surface cracks in earthen heritage sites. Therefore, it is very important to collect a large amount of image data of surface cracks on earthen heritage site surfaces and build a larger dataset. In the future, UAVs can be used to collect crack images, and image cropping and annotating can be performed using the trained model to establish a dataset.
  • By sharing the dataset of cracks on the surface of heritage sites and uploading crack pictures in real-time, a larger database of crack pictures was established, which is convenient for researchers to study the cracks.
  • This study focused on the identification of cracks. Future work can be done regarding the location, length, width, and depth of cracks.
  • The current solutions, which cannot be deployed in lightweight embedded systems, were limited due to the complexity of the network. For practical application, models that can be deployed based on UAVs will be able to detect cracks more efficiently.

Author Contributions

Conceptualization, Y.Z. and Z.Z.; methodology, Y.Z. and Z.Z.; software, Y.Z. and Q.L.; validation, W.Z.; data curation, W.Z. and Q.L.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported in part by the National Key Research and Development Program of China under Grant No. 2019YFC1520904.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Richards, J.; Zhao, G.; Zhang, H.; Viles, H. A Controlled Field Experiment to Investigate the Deterioration of Earthen Heritage by Wind and Rain. Herit. Sci. 2019, 7, 51. [Google Scholar] [CrossRef] [Green Version]
  2. Shao, M.; Li, L.; Wang, S.; Wang, E.; Li, Z. Deterioration Mechanisms of Building Materials of Jiaohe Ruins in China. J. Cult. Herit. 2013, 14, 38–44. [Google Scholar] [CrossRef]
  3. Pu, T.; Chen, W.; Du, Y.; Li, W.; Su, N. Snowfall-Related Deterioration Behavior of the Ming Great Wall in the Eastern Qinghai-Tibet Plateau. Nat. Hazards 2016, 84, 1539–1550. [Google Scholar] [CrossRef]
  4. Li, L.; Shao, M.; Wang, S.; Li, Z. Preservation of Earthen Heritage Sites on the Silk Road, Northwest China from the Impact of the Environment. Environ. Earth Sci. 2011, 64, 1625–1639. [Google Scholar] [CrossRef]
  5. Richards, J.; Viles, H.; Guo, Q. The Importance of Wind as a Driver of Earthen Heritage Deterioration in Dryland Environments. Geomorphology 2020, 369, 107363. [Google Scholar] [CrossRef]
  6. Oliveira, H.; Correia, P.L. Automatic Road Crack Segmentation Using Entropy and Image Dynamic Thresholding. In Proceedings of the 2009 17th European Signal Processing Conference, Glasgow, UK, 24–28 August 2009; pp. 622–626. [Google Scholar]
  7. Talab, A.M.A.; Huang, Z.; Xi, F.; HaiMing, L. Detection Crack in Image Using Otsu Method and Multiple Filtering in Image Processing Techniques. Optik 2016, 127, 1030–1033. [Google Scholar] [CrossRef]
  8. Mishra, R.; Chandrakar, C.L.; Mishra, R. Surface defects detection for ceramic tiles using image processing and morphological techniques. Int. J. Res. Eng. Appl. Sci. 2012, 2, 1307–1322. [Google Scholar]
  9. Abdel-Qader, I.; Abudayyeh, O.; Kelly, M.E. Analysis of Edge-Detection Techniques for Crack Identification in Bridges. J. Comput. Civ. Eng. 2003, 17, 255–263. [Google Scholar] [CrossRef] [Green Version]
  10. Salman, M.; Mathavan, S.; Kamal, K.; Rahman, M. Pavement Crack Detection Using the Gabor Filter. In Proceedings of the 16th International IEEE Conference on Intelligent Transportation Systems (ITSC 2013), The Hague, The Netherlands, 6–9 October 2013; pp. 2039–2044. [Google Scholar]
  11. Medina, R.; Llamas, J.; Gómez-García-Bermejo, J.; Zalama, E.; Segarra, M. Crack Detection in Concrete Tunnels Using a Gabor Filter Invariant to Rotation. Sensors 2017, 17, 1670. [Google Scholar] [CrossRef]
  12. Wang, P.; Huang, H. Comparison Analysis on Present Image-Based Crack Detection Methods in Concrete Structures. In Proceedings of the 2010 3rd International Congress on Image and Signal Processing, Yantai, China, 16–18 October 2010; pp. 2530–2533. [Google Scholar]
  13. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-Based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  14. Pal, K.K.; Sudeep, K.S. Preprocessing for Image Classification by Convolutional Neural Networks. In Proceedings of the 2016 IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bangalore, India, 20–21 May 2016; pp. 1778–1781. [Google Scholar]
  15. Jmour, N.; Zayen, S.; Abdelkrim, A. Convolutional Neural Networks for Image Classification. In Proceedings of the 2018 International Conference on Advanced Systems and Electric Technologies (IC_ASET), Hammamet, Tunisia, 22–25 March 2018; pp. 397–402. [Google Scholar]
  16. Basha, S.H.S.; Dubey, S.R.; Pulabaigari, V.; Mukherjee, S. Impact of Fully Connected Layers on Performance of Convolutional Neural Networks for Image Classification. Neurocomputing 2020, 378, 112–119. [Google Scholar] [CrossRef]
  17. Kang, K.; Ouyang, W.; Li, H.; Wang, X. Object Detection from Video Tubelets with Convolutional Neural Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 817–825. [Google Scholar]
  18. Galvez, R.L.; Bandala, A.A.; Dadios, E.P.; Vicerra, R.R.P.; Maningo, J.M.Z. Object Detection Using Convolutional Neural Networks. In Proceedings of the TENCON 2018–2018 IEEE Region 10 Conference, Jeju, Republic of Korea, 28–31 October 2018; pp. 2023–2027. [Google Scholar]
  19. Li, K.; Ma, W.; Sajid, U.; Wu, Y.; Wang, G. Object Detection with Convolutional Neural Networks. arXiv 2019, arXiv:1912.01844. [Google Scholar]
  20. Kampffmeyer, M.; Salberg, A.-B.; Jenssen, R. Semantic Segmentation of Small Objects and Modeling of Uncertainty in Urban Remote Sensing Images Using Deep Convolutional Neural Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 680–688. [Google Scholar]
  21. Mohammadimanesh, F.; Salehi, B.; Mahdianpari, M.; Gill, E.; Molinier, M. A New Fully Convolutional Neural Network for Semantic Segmentation of Polarimetric SAR Imagery in Complex Land Cover Ecosystem. ISPRS J. Photogramm. Remote Sens. 2019, 151, 223–236. [Google Scholar] [CrossRef]
  22. Yang, H.; Yu, B.; Luo, J.; Chen, F. Semantic Segmentation of High Spatial Resolution Images with Deep Neural Networks. GIScience Remote Sens. 2019, 56, 749–768. [Google Scholar] [CrossRef]
  23. Cha, Y.-J.; Choi, W.; Büyüköztürk, O. Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks: Deep Learning-Based Crack Damage Detection Using CNNs. Comput. Aided Civ. Infrastruct. Eng. 2017, 32, 361–378. [Google Scholar] [CrossRef]
  24. Zhao, X.; Li, S. Convolutional Neural Networks-Based Crack Detection for Real Concrete Surface. In Proceedings of the Sensors and Smart Structures Technologies for Civil, Mechanical, and Aerospace Systems 2018, Denver, CO, USA, 5–8 March 2018; Sohn, H., Ed.; SPIE: Bellingham, WA, USA; p. 143.
  25. Dung, C.V.; Anh, L.D. Autonomous Concrete Crack Detection Using Deep Fully Convolutional Neural Network. Autom. Constr. 2019, 99, 52–58. [Google Scholar] [CrossRef]
  26. Shanmugamani, R.; Sadique, M.; Ramamoorthy, B. Detection and Classification of Surface Defects of Gun Barrels Using Computer Vision and Machine Learning. Measurement 2015, 60, 222–230. [Google Scholar] [CrossRef]
  27. Shi, Y.; Cui, L.; Qi, Z.; Meng, F.; Chen, Z. Automatic Road Crack Detection Using Random Structured Forests. IEEE Trans. Intell. Transport. Syst. 2016, 17, 3434–3445. [Google Scholar] [CrossRef]
  28. Ni, F.; Zhang, J.; Chen, Z. Zernike-moment Measurement of Thin-crack Width in Images Enabled by Dual-scale Deep Learning. Comput. Aided Civ. Infrastruct. Eng. 2019, 34, 367–384. [Google Scholar] [CrossRef]
  29. Kang, D.; Benipal, S.S.; Gopal, D.L.; Cha, Y.-J. Hybrid Pixel-Level Concrete Crack Segmentation and Quantification across Complex Backgrounds Using Deep Learning. Autom. Constr. 2020, 118, 103291. [Google Scholar] [CrossRef]
  30. Wei, F.; Yao, G.; Yang, Y.; Sun, Y. Instance-Level Recognition and Quantification for Concrete Surface Bughole Based on Deep Learning. Autom. Constr. 2019, 107, 102920. [Google Scholar] [CrossRef]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. ISBN 978-3-319-24573-7. [Google Scholar]
  32. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  33. Kendall, A.; Badrinarayanan, V.; Cipolla, R. Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding. In Proceedings of the British Machine Vision Conference 2017, London, UK, 4–7 September 2017; British Machine Vision Association: London, UK, 2017; p. 57. [Google Scholar]
  34. Chaurasia, A.; Culurciello, E. LinkNet: Exploiting Encoder Representations for Efficient Semantic Segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  36. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. AAAI 2017, 31. [Google Scholar] [CrossRef]
  37. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  38. Chen, L.-C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  39. Sizyakin, R.; Voronin, V.V.; Gapon, N.; Pižurica, A. A Deep Learning Approach to Crack Detection on Road Surfaces. In Proceedings of the Artificial Intelligence and Machine Learning in Defense Applications II, Online, 21–25 September 2020; Dijk, J., Ed.; SPIE: London, UK; p. 20.
  40. Lau, S.L.H.; Chong, E.K.P.; Yang, X.; Wang, X. Automated Pavement Crack Segmentation Using U-Net-Based Convolutional Neural Network. IEEE Access 2020, 8, 114892–114899. [Google Scholar] [CrossRef]
  41. Lin, F.; Yang, J.; Shu, J.; Scherer, R.J. Crack Semantic Segmentation Using the U-Net with Full Attention Strategy. arXiv 2021, arXiv:2104.14586. [Google Scholar]
  42. Cao, X.; Yao, B.; Chen, B.; Wang, Y. Multi-Defect Detection for Magnetic Tile Based on SE-U-Net. In Proceedings of the 2020 IEEE International Symposium on Product Compliance Engineering-Asia (ISPCE-CN), Chongqing, China, 6–8 November 2020; pp. 1–6. [Google Scholar]
  43. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  44. Song, C.; Wu, L.; Chen, Z.; Zhou, H.; Lin, P.; Cheng, S.; Wu, Z. Pixel-Level Crack Detection in Images Using SegNet. In Multi-disciplinary Trends in Artificial Intelligence; Chamchong, R., Wong, K.W., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2019; Volume 11909, pp. 247–254. ISBN 978-3-030-33708-7. [Google Scholar]
  45. Chen, T.; Cai, Z.; Zhao, X.; Chen, C.; Liang, X.; Zou, T.; Wang, P. Pavement Crack Detection and Recognition Using the Architecture of SegNet. J. Ind. Inf. Integr. 2020, 18, 100144. [Google Scholar] [CrossRef]
  46. Zou, Q.; Zhang, Z.; Li, Q.; Qi, X.; Wang, Q.; Wang, S. DeepCrack: Learning Hierarchical Convolutional Features for Crack Detection. IEEE Trans. on Image Process. 2019, 28, 1498–1512. [Google Scholar] [CrossRef]
  47. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  48. Loverdos, D.; Sarhosis, V. Automatic Image-Based Brick Segmentation and Crack Detection of Masonry Walls Using Machine Learning. Autom. Constr. 2022, 140, 104389. [Google Scholar] [CrossRef]
  49. Zou, Q.; Cao, Y.; Li, Q.; Mao, Q.; Wang, S. CrackTree: Automatic Crack Detection from Pavement Images. Pattern Recognit. Lett. 2012, 33, 227–238. [Google Scholar] [CrossRef]
  50. Yang, F.; Zhang, L.; Yu, S.; Prokhorov, D.; Mei, X.; Ling, H. Feature Pyramid and Hierarchical Boosting Network for Pavement Crack Detection. IEEE Trans. Intell. Transport. Syst. 2020, 21, 1525–1535. [Google Scholar] [CrossRef] [Green Version]
  51. Liu, Y.; Yao, J.; Lu, X.; Xie, R.; Li, L. DeepCrack: A Deep Hierarchical Feature Learning Architecture for Crack Segmentation. Neurocomputing 2019, 338, 139–153. [Google Scholar] [CrossRef]
  52. Alipour, M.; Harris, D.K. Increasing the Robustness of Material-Specific Deep Learning Models for Crack Detection across Different Materials. Eng. Struct. 2020, 206, 110157. [Google Scholar] [CrossRef]
  53. Özgenel, Ç.F.; Sorguç, A.G. Performance Comparison of Pretrained Convolutional Neural Networks on Crack Detection in Buildings. In Proceedings of the International Symposium on Automation and Robotics in Construction, Berlin, Germany, 20–25 July 2018. [Google Scholar]
  54. Iakubovskii, P. Segmentation Models 2019. Available online: https://github.com/qubvel/segmentation_models (accessed on 10 May 2022).
  55. Jadon, S. A Survey of Loss Functions for Semantic Segmentation. In Proceedings of the 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Viña del Mar, Chile, 27–29 October 2020; pp. 1–7. [Google Scholar]
Figure 1. Suoyang Ancient City. (a) Google Maps image of Suoyang Ancient City. The photos show (b) some typical damage to wall surfaces, including (c) scouring, (d) flaking, and (e) cracking.
Figure 1. Suoyang Ancient City. (a) Google Maps image of Suoyang Ancient City. The photos show (b) some typical damage to wall surfaces, including (c) scouring, (d) flaking, and (e) cracking.
Applsci 12 12830 g001
Figure 2. Workflow of the proposed method for detecting and predicting cracks in rammed earthen heritage sites.
Figure 2. Workflow of the proposed method for detecting and predicting cracks in rammed earthen heritage sites.
Applsci 12 12830 g002
Figure 3. Samples of cracks in the training dataset.
Figure 3. Samples of cracks in the training dataset.
Applsci 12 12830 g003
Figure 4. The metrics Precision, Recall, IoU, and F1 score as obtained from U-Net for different backbones: (a) vgg16, (b) resnet152, (c) densenet201, (d) inceptionv3.
Figure 4. The metrics Precision, Recall, IoU, and F1 score as obtained from U-Net for different backbones: (a) vgg16, (b) resnet152, (c) densenet201, (d) inceptionv3.
Applsci 12 12830 g004
Figure 5. The metrics Precision, Recall, IoU, and F1 score as obtained from Linknet for different backbones: (a) vgg16, (b) resnet152, (c) densenet201, (d) inceptionv3.
Figure 5. The metrics Precision, Recall, IoU, and F1 score as obtained from Linknet for different backbones: (a) vgg16, (b) resnet152, (c) densenet201, (d) inceptionv3.
Applsci 12 12830 g005aApplsci 12 12830 g005b
Figure 6. The metrics Precision, Recall, IoU, and F1 score as obtained from FPN for different backbones: (a) vgg16, (b) resnet152, (c) densenet201, (d) inceptionv3.
Figure 6. The metrics Precision, Recall, IoU, and F1 score as obtained from FPN for different backbones: (a) vgg16, (b) resnet152, (c) densenet201, (d) inceptionv3.
Applsci 12 12830 g006
Figure 7. The original image, ground truth, and predictions with all models for different images from the validation dataset.
Figure 7. The original image, ground truth, and predictions with all models for different images from the validation dataset.
Applsci 12 12830 g007aApplsci 12 12830 g007b
Table 1. Software environment setups employed in this paper.
Table 1. Software environment setups employed in this paper.
SoftwareDetails
OSWindows 10 Professional
Programming languagePython 3.8
Deep learning frameworkTensorflow-gpu 2.3.0
Dependent libraryCuda, Keras, PIL etc.
Table 2. Hardware environment setups employed in this paper.
Table 2. Hardware environment setups employed in this paper.
HardwareDetails
CPUIntel(R) Xeon(R) CPU E5-2670 v3 @ 2.30GHz
CPU RAM(GB)256
GPUNVIDIA GeForce RTX 3070
GPU memory (GB)8
CUDA cores5888
CUDA version11.2
Table 3. Total number of parameters for each model.
Table 3. Total number of parameters for each model.
ModelBackbonesParameters
Unetvgg1623,752,273
resnet15267,295,194
densenet20126,378,577
inceptionv329,933,105
Linknetvgg1620,325,137
resnet15263,517,466
densenet20122,549,393
inceptionv326,268,401
FPNvgg1617,572,545
resnet15261,645,898
densenet20121,523,905
inceptionv325,029,281
Table 4. The metrics are presented for the validation set.
Table 4. The metrics are presented for the validation set.
ModelBackbonesPrecision%Recall%F1 Score%IoU%Best Epoch *
Unetvgg1689.4185.5283.6572.0084
resnet15293.0279.8582.2869.9893
densenet20192.6094.3883.6872.0180
inceptionv395.2180.6682.6570.5184
Linknetvgg1690.1985.9782.7670.7499
resnet15290.4083.0782.6370.4376
densenet20189.8081.5681.8369.3396
inceptionv388.0380.2080.8867.9988
FPNvgg1688.8385.8284.4073.1193
resnet15290.5083.3583.3171.4689
densenet20187.8482.6282.5970.4490
inceptionv390.0585.3781.7269.2594
* Epoch where the highest F1 score was obtained for the validation set.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Zhang, Z.; Zhao, W.; Li, Q. Crack Segmentation on Earthen Heritage Site Surfaces. Appl. Sci. 2022, 12, 12830. https://0-doi-org.brum.beds.ac.uk/10.3390/app122412830

AMA Style

Zhang Y, Zhang Z, Zhao W, Li Q. Crack Segmentation on Earthen Heritage Site Surfaces. Applied Sciences. 2022; 12(24):12830. https://0-doi-org.brum.beds.ac.uk/10.3390/app122412830

Chicago/Turabian Style

Zhang, Yuan, Zhiyong Zhang, Wu Zhao, and Qiang Li. 2022. "Crack Segmentation on Earthen Heritage Site Surfaces" Applied Sciences 12, no. 24: 12830. https://0-doi-org.brum.beds.ac.uk/10.3390/app122412830

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop