PDF(18104 KB)
PDF(18104 KB)
PDF(18104 KB)
基于偏振成像与深度学习的浑浊水体水下结构表观缺陷检测
Detection of Apparent Defects of Underwater Structures in Turbid Waters Based on Polarization Imaging and Deep Learning
浅水环境常呈现浑浊特征,导致光学图像出现模糊、色偏、对比度低等问题。浑浊水体中散射粒子会遮蔽水下结构表观缺陷信息,造成缺陷识别率低、检测效率低和分类不准等问题。针对这些挑战,提出一种基于偏振成像和深度学习的轻量级三阶段水下缺陷检测方法,借助偏振复原模型、超分辨率重建模型和缺陷检测模型3个子模型实现缺陷检测。偏振复原模型用于实现浑浊水体的清晰成像和水下缺陷图像复原,CAA-SRGAN超分辨率重建模型用于获取高分辨率水下缺陷图像,CBAM-YOLOv7缺陷检测模型用于检测水下结构常见的裂缝、孔洞和剥落缺陷,最终形成适用于浑浊水体水下结构的PCC-YOLOv7缺陷检测模型。分别通过与现有先进图像复原、超分辨率重建和目标检测方法进行对比分析,结果显示3个子模型的输出结果在各自评价指标中均有提升。PCC-YOLOv7缺陷检测模型对平均精确度指标(mAP0.5、mAP0.75、mAP0.5~0.95)的提升幅度均值达33.5%。本文所构建的模型相较于现有模型,对浑浊水下检测场景有着更强的适配性,能够为浑浊水体中水下结构表观缺陷检测工作提供切实可行的方法。
[Objective] In underwater engineering inspection, the turbid shallow water environment severely hinders the performance of machine vision-based methods for detecting surface defects in underwater structures. To address the challenge of defect detection in turbid water, this study proposes a lightweight three-stage underwater defect detection method that integrates polarization imaging and deep learning techniques. A defect detection model, named PCC-YOLOv7, is developed. [Methods] First, polarization imaging technology was combined with a polarization restoration model to analyze the polarization characteristics of light waves. This approach effectively suppressed scattering interference in turbid water, thereby achieving clear imaging of turbid environments and restoring defect images. Consequently, defect details obscured by scattering particles were reconstructed. Second, the CAA-SRGAN (Coordinate Attention ACON-Super Resolution Generative Adversarial Network) model was introduced. By employing an improved attention mechanism and a generative adversarial network structure, super-resolution processing was performed on the restored images. This yielded high-resolution underwater defect images, providing a high-quality data foundation for subsequent precise detection. Finally, a defect detection model based on CBAM-YOLOv7 was established, where the convolutional block attention module (CBAM) was utilized to enhance the network’s focus on defect features. Leveraging the advanced YOLOv7 object detection framework, common underwater structural defects, including cracks, holes, and spalling can be rapidly and accurately identified. These three sub-models worked collaboratively to form a comprehensive detection system. [Results] For image restoration, the polarization restoration model exhibited superior performance in metrics such as image clarity and color fidelity compared to current restoration methods. The CAA-SRGAN model generated images with notable improvements in detail texture preservation and resolution enhancement. The CBAM-YOLOv7 defect detection model achieved higher accuracy in both defect localization and classification. A comprehensive evaluation of the PCC-YOLOv7 defect detection model revealed an average improvement of 33.5% in mean average precision (mAP0.5, mAP0.75, and mAP0.5-0.95). Compared to existing models, PCC-YOLOv7 significantly enhanced defect detection performance in turbid underwater environments, effectively improving both recognition rate and detection efficiency. [Conclusions] The PCC-YOLOv7 defect detection model innovatively integrates polarization imaging technology with deep learning. Through the collaborative operation of three functionally complementary sub-models, it successfully addresses the challenge of detecting surface defects in underwater structures in turbid water. Compared to existing models, the proposed model demonstrates enhanced adaptability to turbid underwater detection scenarios. It enables stable and efficient detection of surface defects in underwater structures under complex turbid conditions, providing a practical technical solution for the safety assessment and maintenance of underwater structures. Future work may focus on further optimizing the model structure and extending its application to more underwater scenarios.
浑浊水体 / 水下结构 / 缺陷检测 / 偏振成像 / 深度学习 / 超分辨率重建
turbid water / underwater structure / defect detection / polarization imaging / deep learning / super-resolution reconstruction
| [1] |
钮新强, 谭界雄, 田金章. 混凝土面板堆石坝病害特点及其除险加固[J]. 人民长江, 2016, 47(13):1-5.
(
|
| [2] |
盛金保, 厉丹丹, 蔡荨, 等. 大坝风险评估与管理关键技术研究进展[J]. 中国科学:技术科学, 2018, 48(10):1057-1067.
(
|
| [3] |
陈永灿, 陈嘉杰, 王皓冉, 等. 大直径长引水隧洞水下检测机器人系统关键技术[J]. 清华大学学报(自然科学版), 2023, 63(7): 1015-1031.
(
|
| [4] |
陶显, 侯伟, 徐德. 基于深度学习的表面缺陷检测方法综述[J]. 自动化学报, 2021, 47(5):1017-1034.
(
|
| [5] |
|
| [6] |
|
| [7] |
|
| [8] |
|
| [9] |
|
| [10] |
李培育, 张雅丽. 基于改进SRGAN模型的人脸图像超分辨率重建[J]. 计算机工程, 2023, 49(4):199-205.
传统生成对抗网络模型重建人脸图像时出现过多失真,难以在减少失真的情况下有效提高人脸图像真实感。针对该问题,在生成对抗网络SRGAN模型的基础上,提出一种改进的人脸图像超分辨率重建方法。为提高重建像素点与周围像素点的相关性,将双注意力机制模块嵌入到SRGAN模型的生成器和判别器中,在空间域和通道域中获取更精准的特征依赖关系。同时应用自适应激活函数ACON取代原SRGAN网络中的激活函数,通过动态学习ACON激活函数参数为每个神经元设计不同激活形式,从而提高网络特征表达能力。使用改进SRGAN的人脸图像超分辨率重建算法在CelebA测试集上进行重建实验,结果表明:该算法较原算法PSNR值提高0.675 dB,SSIM值提高0.016,LPIPS值优化0.036,有效减少了重建人脸图像中眼睛等重点部位的失真情况;与其他非生成对抗网络的主流算法相比,LPIPS值最低优化0.107,最高优化0.205,有效提高了重建人脸图像的真实感。
(
Because the traditional generation countermeasure network model exhibits excessive distortion when reconstructing the face image, it is difficult to effectively improve the realism of the face image.Under the condition of reducing the distortion, an improved super-resolution reconstruction method of face images based on SRGAN model is proposed.To increase the correlation between the reconstructed and surrounding pixels, the Dual Attention(DA) mechanism module is embedded in the generator and discriminator of the SRGAN model to obtain more accurate feature dependency in the spatial and channel domains.An adaptive activation function, termed ACON, is used to replace the original activation function in an SRGAN network, and different activation forms are designed for each neuron by dynamically learning the parameters of the ACON activation function, to improve the ability of network feature expression.The reconstruction experiment is performed on the CelebA using the improved SRGAN face image reconstruction algorithm.The results show that compared with the original algorithm, the PSNR value increases by 0.675 dB, the SSIM value increases by 0.016, and the LPIPS value is optimized by 0.036, effectively reducing the distortion of key parts, such as the eyes, in the reconstructed face image.Compared with other mainstream algorithms that do not generate countermeasure networks, the LPIPS value is optimized by 0.107 at the lowest and 0.205 at the highest, effectively improving the realism of the reconstructed face image.
|
| [11] |
|
| [12] |
|
| [13] |
|
| [14] |
This paper presents a new class of the "frequency domain"-based signal/image enhancement algorithms including magnitude reduction, log-magnitude reduction, iterative magnitude and a log-reduction zonal magnitude technique. These algorithms are described and applied for detection and visualization of objects within an image. The new technique is based on the so-called sequency ordered orthogonal transforms, which include the well-known Fourier, Hartley, cosine, and Hadamard transforms, as well as new enhancement parametric operators. A wide range of image characteristics can be obtained from a single transform, by varying the parameters of the operators. We also introduce a quantifying method to measure signal/image enhancement called EME. This helps choose the best parameters and transform for each enhancement. A number of experimental results are presented to illustrate the performance of the proposed algorithms.
|
| [15] |
Quality evaluation of underwater images is a key goal of underwater video image retrieval and intelligent processing. To date, no metric has been proposed for underwater color image quality evaluation (UCIQE). The special absorption and scattering characteristics of the water medium do not allow direct application of natural color image quality metrics especially to different underwater environments. In this paper, subjective testing for underwater image quality has been organized. The statistical distribution of the underwater image pixels in the CIELab color space related to subjective evaluation indicates the sharpness and colorful factors correlate well with subjective image quality perception. Based on these, a new UCIQE metric, which is a linear combination of chroma, saturation, and contrast, is proposed to quantify the non-uniform color cast, blurring, and low-contrast that characterize underwater engineering and monitoring images. Experiments are conducted to illustrate the performance of the proposed UCIQE metric and its capability to measure the underwater image enhancement results. They show that the proposed metric has comparable performance to the leading natural color image quality metrics and the underwater grayscale image quality metrics available in the literature, and can predict with higher accuracy the relative amount of degradation with similar image content in underwater environments. Importantly, UCIQE is a simple and fast solution for real-time underwater video processing. The effectiveness of the presented measure is also demonstrated by subjective evaluation. The results show better correlation between the UCIQE and the subjective mean opinion score.
|
| [16] |
|
/
| 〈 |
|
〉 |