Skip to content

CrackDenseLinkNet: A deep convolutional neural network for semantic segmentation of cracks on concrete surface images (DenseNet and LinkNet combination with the compound loss)

License

Notifications You must be signed in to change notification settings

preethamam/CrackDenseLinkNet-DeepLearning-CrackSegmentation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

99 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CrackDenseLinkNet (CDLN)

Title

CrackDenseLinkNet: A deep convolutional neural network for semantic segmentation of cracks on concrete surface images

Abstract

Cracks are the defects formed by cyclic loading, fatigue, shrinkage, creep, and so on. In addition, they represent the deterioration of the structures over some time. Therefore, it is essential to detect and classify them according to the condition grade at the early stages to prevent the collapse of structures. Deep learning-based semantic segmentation convolutional neural network (CNN) has millions of learnable parameters. However, depending on the complexity of the CNN, it takes hours to days to train the network fully. In this study, an encoder network DenseNet and modified LinkNet with five upsampling blocks were used as a decoder network. The proposed network is referred to as the ‘‘CrackDenseLinkNet’’ in this work. CrackDenseLinkNet has 19.15 million trainable parameters, although the input image size is 512 x 512 and has a deeper encoder. CrackDenseLinkNet and four other state-of-the-art (SOTA) methods were evaluated on three public and one private datasets. The proposed CNN, CrackDenseLinkNet, outperformed the best SOTA method, CrackSegNet, by 2.2% of F1-score on average across the four datasets. Lastly, a crack profile analysis demonstrated that the CrackDenseLinkNet has lesser variance in relative errors for the crack width, length, and area categories against the ground-truth data.

For more details, please check out: CrackDenseLinkNet Paper.

CrackDenseLinkNet Python modules installation and usage

Configuration

Our network was built and tested on the following system requirements:

Ubuntu 22.04 or higher
CUDA toolkit 11.7
PyTorch 2.0 or higher
Python 3.8 or higher

Setting up the Anaconda Environment

Please make sure you have installed Anaconda and have Python version > 3.8.5 before following the next steps

conda env create -f crackdenselinknet.yml
conda activate crackdenselinknet

This should install all the necessary packages required to run the below-documented scripts. However, by any chance, if still you get package errors then just install that package through pip or conda.

Usage

Place the Liu+Xincong+CrackSegNet+CDLN within the CrackDenseLinkNet folder, or if the datasets are placed somewhere else, please change the DATA_DIR, VALIND_DIR, VALID_DIR in the train.py and DATA_DIR in the test.py files. If you want to test o multiple model/weight files, please use the test_loop.py and change the DATA_DIR accordingly. To execute the CrackDenseLinkNet program either run the train.py, test.py, and/or test_loop.py in the IDE or use the shell scripts in the scripts folder.

CrackDenseLinkNet and other methods trained model/weights files

The trained weights files of the FCN, DeepCrack, FPHB, CrackSegNet, and CrackDenseLinket methods on the four combined datasets (Liu+Xincong+CrackSegNet+CDLN) are provided below in the table:

Method Dataset Link
FCN Liu+Xincong+CrackSegNet+CDLN Weights
DeepCrack Liu+Xincong+CrackSegNet+CDLN Weights
CrackSegNet Liu+Xincong+CrackSegNet+CDLN Weights
FPHB Liu+Xincong+CrackSegNet+CDLN Weights
CrackDenseLinket Liu+Xincong+CrackSegNet+CDLN Weights

CrackDenseLinkNet, FCN, DeepCrack, and CrackSegNet datasets combined for training, validation, and testing

In this study, four datasets were used to evaluate the crack segmentation capability of five CNN-based semantic segmentation networks. Datasets FCN [1], DeepCrack [2], and CrackSegNet [3] are available to public. In addition, the proposed method dataset CrackDenseLinkNet was created around the University of Southern California campus. The complexity of images varies gradually from CrackDenseLinkNet, FCN, DeepCrack, and CrackSegNet. Figure below shows the four datasets’ sample images.

CrackDenseLinkNet dataset consists of 250 testing images of a concrete surface. In this dataset, the cracks have a stronger contrast, texture, and wider cracks relative to the other datasets. About 249 images have high textural content out of 250. The FCN dataset is primarily of the concrete surface and consists of 154 testing crack samples that vary in size, out of which 6 and 148 images have low and high texture, respectively. In this dataset, crack widths are thin and thicker. Furthermore, it consists of longitudinal, transverse, and surface cracks. All the images are of high quality. However, it was observed that some of the groundtruth images were wrongly labeled (labels are thicker than the actual width of the cracks).

Similar to the FCN dataset, the DeepCrack dataset consists of 237 testing samples of various longitudinal, transverse, and surface cracks. About 78 and 22% of the images have concrete and asphalt material surfaces, respectively. The crack width varies largely across the images of this dataset. Around 21 and 216 images in this dataset have low and high textures, respectively. Lastly, CrackSegNet consists of 184 testing images of the concrete surface, 5 and 179 images of low and high texture, respectively. The crack width of this dataset is relatively thinner compared to the other datasets. This is the most challenging dataset to train the semantic segmentation CNN methods, as it contains highly blurry images, textural noise, and paint artifacts. This dataset was purposefully included to assess the limitations of all five methods in comparison.

All data
Datasets’ sample images: (a) FCN, (b) DeepCrack, (c) CrackSegNet, and (d) CrackDenseLinkNet.
Download Link: Liu+Xincong+CrackSegNet+CDLN

CrackDenseLinkNet dataset only

CrackDenseLinkNet dataset consists of 707, 79, and 250 training, validation, and testing images of a concrete surface with their corresponding pixel-wise semantic annotations/labels. In this dataset, the cracks have a stronger contrast, texture, and wider cracks relative to the other datasets used in this study. About 249 images have high textural content out of 250. A few of the testing images are displayed in the figure below. The download link is provided below.

Original Images Groundtruth Images
Original Groundtruth
Sample images and their groundtruth from the CrackDenseLinkNet dataset. Download Link: CrackDenseLinkNet

Citation

CrackDenseLinkNet code and dataset are available to the public. If you use this code/dataset in your research, please use the following BibTeX entry to cite:

@article{manjunatha2023crackdenselinknet,
author = {Preetham Manjunatha and Sami F Masri and Aiichiro Nakano and Landon Carter Wellford},
title ={{CrackDenseLinkNet}: a deep convolutional neural network for semantic segmentation of cracks on concrete surface images},
journal = {Structural Health Monitoring},
volume = {0},
number = {0},
pages = {14759217231173305},
year={2023},
publisher={SAGE Publications Sage UK: London, England},
doi = {10.1177/14759217231173305},
URL = {https://doi.org/10.1177/14759217231173305},
eprint = {https://doi.org/10.1177/14759217231173305},
}

References

[1] Yang X, Li H, Yu Y, et al. Automatic pixel-level crack detection and measurement using fully convolutional network. Comput-Aided Civil Infrastruct Eng 2018; 33(12): 1090–1109.

[2] Liu Y, Yao J, Lu X, et al. Deepcrack: a deep hierarchical feature learning architecture for crack segmentation. Neurocomputing 2019; 338: 139–153.

[3] Ren Y, Huang J, Hong Z, et al. Image-based concrete crack detection in tunnels using deep fully convolutional networks. Constr Build Mater 2020; 234: 117367.

About

CrackDenseLinkNet: A deep convolutional neural network for semantic segmentation of cracks on concrete surface images (DenseNet and LinkNet combination with the compound loss)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published