<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Mater.</journal-id>
<journal-title>Frontiers in Materials</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Mater.</abbrev-journal-title>
<issn pub-type="epub">2296-8016</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1212909</article-id>
<article-id pub-id-type="doi">10.3389/fmats.2023.1212909</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Materials</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Damage localization method using ultrasonic lamb waves and Wav2Vec2.0 neural network</article-title>
<alt-title alt-title-type="left-running-head">Qian et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fmats.2023.1212909">10.3389/fmats.2023.1212909</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Qian</surname>
<given-names>Lubin</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2294454/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Liu</surname>
<given-names>Sihao</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Fan</surname>
<given-names>Guopeng</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Liu</surname>
<given-names>Xinlong</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhang</surname>
<given-names>Hui</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Mei</surname>
<given-names>Yaohua</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Xing</surname>
<given-names>Yuhui</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Zhiqiang</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>School of Urban Rail Transportation</institution>, <institution>Shanghai University of Engineering Science</institution>, <addr-line>Shanghai</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Shanghai Urban Railway Operation Co., Ltd.</institution>, <addr-line>Shanghai</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Wuhan Ruimin Testing Technology Co., Ltd.</institution>, <addr-line>Wuhan</addr-line>, <addr-line>Henan</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Luoyang Sunrui Rubber and Plastic Technology Co., Ltd.</institution>, <addr-line>Luoyang</addr-line>, <addr-line>Henan</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1267729/overview">Fuyin Ma</ext-link>, Xi&#x2019;an Jiaotong University, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2302130/overview">Yong Li</ext-link>, Nanyang Technological University, Singapore</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1442910/overview">Hui Zhang</ext-link>, Southeast University, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2302326/overview">Maoxun Sun</ext-link>, University of Shanghai for Science and Technology, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2354343/overview">Ben Li</ext-link>, Nanyang Institute of Technology, China</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Lubin Qian, <email>qianlb@sues.edu.cn</email>; Guopeng Fan, <email>phdfanry@sues.edu.cn</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>09</day>
<month>08</month>
<year>2023</year>
</pub-date>
<pub-date pub-type="collection">
<year>2023</year>
</pub-date>
<volume>10</volume>
<elocation-id>1212909</elocation-id>
<history>
<date date-type="received">
<day>27</day>
<month>04</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>10</day>
<month>07</month>
<year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2023 Qian, Liu, Fan, Liu, Zhang, Mei, Xing and Wang.</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Qian, Liu, Fan, Liu, Zhang, Mei, Xing and Wang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>In this paper, a Wav2Vec2.0 neural network based on an attention mechanism is proposed to locate defects in array ultrasonic testing signals. This method does not require knowledge of the <italic>a priori</italic> condition of the sample sound velocity or the feature extraction of ultrasonic scattering signals. First, an array piezoelectric ultrasonic testing system is used to detect a signal through hole defects at different positions in the plate structure. Then, three different neural networks&#x2014;1D-CNN, Muti-Transformer, and Wav2Vec2.0&#x2014;are used to locate the defects in the collected ultrasonic testing data. The performance of the network is verified with the data set collected through finite element simulation and the experimental system, and the identification accuracy and the calculation efficiency of different networks are compared and analyzed. To provide a solution for the poor balance of the experimental data set and the weak noise resistance of the simulation data set, a data set expansion method based on time domain transformation technology is proposed. The research results show that, the positioning accuracy of the Wav2Vec2.0 neural network proposed in this article is 98.46%, and the positioning accuracy is superior to Muti Transformer and ID-CNN.</p>
</abstract>
<kwd-group>
<kwd>lamb wave</kwd>
<kwd>Wav2Vec2.0 neural network</kwd>
<kwd>ultrasonic testing</kwd>
<kwd>structural health monitoring</kwd>
<kwd>defect identification</kwd>
</kwd-group>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content>
</contract-sponsor>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Metamaterials</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>With the sustained and rapid development of the national economy, industries such as aerospace, petrochemical, and transportation have made significant progress. Thin plates, as a special structure, are widely used in various industries (<xref ref-type="bibr" rid="B11">Lu et al., 2016</xref>). As an active monitoring technology, structural health monitoring (SHM) based on Lamb waves can be used to realize the rapid and large-scale nondestructive testing of plates (<xref ref-type="bibr" rid="B9">Li et al., 2019</xref>). SHM often uses piezoelectric ceramics as an exciter to excite Lamb waves on the surface of a monitored structure and uses piezoelectric ceramics as sensors to detect and receive defect response signals (<xref ref-type="bibr" rid="B23">Stepinski and Ambrozinski, 2013</xref>). In recent years, array imaging methods based on ultrasound guided waves have been widely applied in the field of Structural Health Monitoring (SHM)/Non-Destructive Testing (NDT), and various array imaging algorithms have been developed; <xref ref-type="bibr" rid="B13">Ma et al. (2020)</xref>, <xref ref-type="bibr" rid="B14">(2022)</xref> analyzed the time-reversal (TR) mirrors in acoustic focusing and imaging. However, existing array imaging methods are difficult to effectively extract weak scattering information of damage from array sensing data, and cannot accurately characterize the location, degree, and type of damage.</p>
<p>With the development of artificial intelligence, deep learning provides a new tool in piezoelectric ultrasonic guided wave sensing signal characterization, feature extraction, and anomaly detection (<xref ref-type="bibr" rid="B20">Rizvi and Abbas, 2021</xref>; <xref ref-type="bibr" rid="B17">Rautela et al., 2021</xref>; <xref ref-type="bibr" rid="B22">Sikdar and Pal, 2021</xref>). Since the 1970s, researchers have been using neural networks to identify defects in ultrasonic testing data, and this research mainly focuses on two aspects. On the one hand, the method of feature extraction and shallow neural network is used to classify the detection data. On the other hand, deep learning methods are directly used, and the input data is mainly manually extracted feature values. In recent years, deep learning algorithms have gradually developed to automatically extract features. <xref ref-type="bibr" rid="B21">Shi et al. (2016)</xref> used a wavelet transform to extract the feature information from an ultrasonic A-scan signal and classified and identified the defect through a deep neural network (DNN), with an increased accuracy of 21.66%. <xref ref-type="bibr" rid="B20">Shao et al. (2020)</xref> proposed a defect-type recognition method based on a VGG16 convolution neural network to directly identify the original signal of array ultrasonic testing and proved that a two-dimensional convolution neural network had higher recognition accuracy. <xref ref-type="bibr" rid="B15">Meng et al., (2017)</xref> proposed the use of a convolutional neural network (CNN) to classify and train the wavelet features of different layers through hierarchical wavelet transform to achieve the defect classification of carbon fiber materials, with an accuracy rate of 98.15%. <xref ref-type="bibr" rid="B6">He et al. (2014)</xref> used the electro-mechanical impedance (EMI) technology of a PZT unit to measure impedance spectrum and used a backpropagation (BP) neural network to quantify the health status of a truss structure. The results showed that the developed neural network model can not only determine the position of a loose rod but can also quantify the looseness level (<xref ref-type="bibr" rid="B6">He et al., 2014</xref>). <xref ref-type="bibr" rid="B16">Munir et al., (2020)</xref> used a CNN model to classify ultrasonic signals after denoising through variational autoencoders (VAEs). <xref ref-type="bibr" rid="B4">Drai et al., (2002)</xref> extracted the eigenvalues of the echo signals of welding defects through the time domain, frequency domain and discrete wavelet, and used K-nearest neighbor (KNN) clustering and artificial neural networks (ANNs) for classification and recognition. <xref ref-type="bibr" rid="B7">Khumaidi et al., (2017)</xref> proposed the use of ultrasonic inspection images of welding surface defects as input and used a CNN to classify different types of welding defects, with an accuracy rate of 95.83%. <xref ref-type="bibr" rid="B2">Caputo et al., (2021)</xref> and others used finite elements and experimental methods to generate ultrasonic guided wave scattering signals at different positions in an aluminum plate, and they obtained the complex mapping relationship between guided wave scattering signals and damage locations through an artificial neural network. This achieved the accurate location of damage in the aluminum plate and achieved good results in the location of damage in a carbon fiber-reinforced polymer composite plate. <xref ref-type="bibr" rid="B5">Ewald et al., (2021)</xref> and others simulated ultrasonic Lamb wave signals in aviation composite materials collected by multiple piezoelectric sensors using the finite element method and combined the signals collected by multiple sensors into a group of mixed time-frequency maps through a normalized short-time Fourier transform. A convolutional neural network was used to extract the abstract features of the mixed time-frequency image, and the mapping relationship between the abstract features and the damage location was established to achieve accurate damage identification. <xref ref-type="bibr" rid="B24">Su et al. (2019)</xref> collected the ultrasonic Lamb wave scattering signals of different types of defects in composite materials through experiments, obtained the amplitude frequency characteristics of the scattering signal using a fast Fourier transform, and established a complex mapping relationship between the amplitude frequency characteristics of the scattering signal and the damage location through the depth neural network to achieve accurate defect identification.</p>
<p>In summary, in recent years, intelligent structural damage assessment methods based on deep learning algorithms have been able to achieve damage location determination (<xref ref-type="bibr" rid="B3">Chen and Liu, 2018</xref>; <xref ref-type="bibr" rid="B8">Li, 2019</xref>), but their feature extractor structure only focuses on local features and ignores the global features of the signal, failing to fully utilize the correlation information between channels of multi-channel piezoelectric array signals. In response to the above issues, this article applies the Wav2Vec2.0 neural network to the raw ultrasound signal data collected by array probes. The array data is used as the input of the Wav2Vec2.0 network, and the Transformer and product quantization structural components are used to extract general features of the ultrasound detection data. The extracted general features are applied to downstream defect localization tasks, achieving an improvement in recognition accuracy on larger datasets.</p>
</sec>
<sec id="s2">
<title>2 Damage location method based on Wav2Vec2.0 neural network</title>
<sec id="s2-1">
<title>2.1 Wav2Vec2.0 neural network</title>
<p>Wav2vec 2.0 is discussed in a paper published by the Facebook AI Lab in 2020 (<xref ref-type="bibr" rid="B1">Alexei et al., 2020</xref>). Through unsupervised training, this model can enable the network to map the original voice samples to a feature space more representative of data features. The use of the calculated feature vector to replace traditional features such as Mel-scaleFrequency Cepstral Coefficients (MFCC) can improve subsequent tasks such as speech recognition or voice event detection. The network structure of the Wav2Vec2.0 pre-training model is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Wav2Vec2.0 network structure.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g001.tif"/>
</fig>
<p>The Wav2Vec2.0 pre-training model includes a convolutional neural network for coding and a context network based on Transformer (<xref ref-type="bibr" rid="B19">Sadhu et al., 2021</xref>). The convolution neural network maps the original input audio signal to the hidden space to obtain the hidden variable Z. The hidden variable Z is mapped to the quantized hidden variable Q through the quantization module and then some positions of the random mask are input to the Transformers network to obtain the context feature vector C. Wav2Vec2.0 uses the comparative learning method to calculate the loss for the masked positions to carry out self-supervised pre-training. For the context feature vector <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mi>c</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> generated by each masked position of the model, the positive example is the quantization hidden variable <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:msub>
<mml:mi>q</mml:mi>
<mml:mi>t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> generated by the quantization module at the same position, and the negative example is the quantization hidden variable <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> generated by the quantization module at other masked positions in the current sentence. The formula of the loss function of comparative learning is as follows:</p>
<p>The formula of the loss function of comparative learning is as follows:<disp-formula id="e1">
<mml:math id="m4">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">L</mml:mi>
<mml:mi mathvariant="normal">m</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">c</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">Q</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">s</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
<mml:mi mathvariant="normal">m</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">c</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">q</mml:mi>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:mi mathvariant="normal">k</mml:mi>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where sim uses cosine similarity.</p>
<p>In addition to the loss function of comparative learning <inline-formula id="inf4">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, Wav2vec 2.0 uses the diversity loss function.<disp-formula id="e2">
<mml:math id="m6">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">L</mml:mi>
<mml:mi mathvariant="normal">d</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi mathvariant="normal">G</mml:mi>
<mml:mi mathvariant="normal">V</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">g</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi mathvariant="normal">G</mml:mi>
</mml:msubsup>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="normal">H</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="normal">p</mml:mi>
<mml:mo>&#xaf;</mml:mo>
</mml:mover>
<mml:mi mathvariant="normal">g</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>
</p>
<p>By maximizing the entropy value of the probability distribution of the selecting code in each codebook group in the quantization module, the model can select each vector in the codebook more evenly to avoid the problem of pattern collapse. The final loss function is L &#x3d; <inline-formula id="inf5">
<mml:math id="m7">
<mml:mrow>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x2b; <inline-formula id="inf6">
<mml:math id="m8">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:msub>
<mml:mi>L</mml:mi>
<mml:mi>d</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>After building the model of wav2vec 2.0, this paper uses wav2vec 2.0 as a feature extractor, applies Transformer and quantization module components to extract the general features of ultrasonic testing data, and then adds a layer of randomly initialized linear layer in the network structure as a Linear classifier to achieve accurate damage identification.</p>
</sec>
<sec id="s2-2">
<title>2.2 Damage location method based on Wav2Vec2.0</title>
<p>The flowchart of a macroscopic damage location system for a thin plate structure proposed in this paper is shown in <xref ref-type="fig" rid="F2">Figure 2</xref>. The deep network feature extraction method combining convolution and the attention mechanism used to determine the macroscopic damage location of the thin plate structure through the Wave2Vec2 model. The specific process is as follows:</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Flowchart of macroscopic damage location system of the thin plate structure.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g002.tif"/>
</fig>
<p>
<statement content-type="step" id="Step_1">
<label>Step 1</label>
<p>The Lamb wave response signal is collected.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_2">
<label>Step2</label>
<p>Signal filtering removes the noise signals and normalizes them.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_3">
<label>Step 3</label>
<p>The Wave2Vec2 model is trained, and the trained model is obtained.</p>
</statement>
</p>
<p>
<statement content-type="step" id="Step_4">
<label>Step4</label>
<p>The test data are processed, including filtering and normalization; Steps 5 and 6: The test data are put back into the training model, and the damage results are obtained and analyzed.</p>
</statement>
</p>
</sec>
</sec>
<sec id="s3">
<title>3 Experiment setup</title>
<sec id="s3-1">
<title>3.1 Datasets</title>
<p>In this study, due to the lack of a publicly available database on ultrasonic non-destructive testing data and the difficulty in obtaining a large amount of measured data in experiments, numerical simulations and experiments were conducted to collect data on aluminum plates with single circular through-hole damage.</p>
<sec id="s3-1-1">
<title>3.1.1 Collection of simulation datasets</title>
<p>In order to understand the interaction between Lamb waves and damage, a numerical analysis model for ultrasonic guided wave excitation/sensing in aluminum plate structures with typical damage was established based on commercial finite element software PZflex. The sound field of ultrasonic guided waves was simulated, as shown in <xref ref-type="fig" rid="F3">Figure 3</xref>. The simulated test sample is an isotropic aluminum plate, and its model parameters are shown in <xref ref-type="table" rid="T1">Table 1</xref>. In order to excite a single S0 mode ultrasonic guided wave, this article uses a customized 5-cycle sine wave with Gaussian Hanning window in MATLAB as the excitation signal, and the excitation frequency is selected as 200&#xa0;kHz (<xref ref-type="bibr" rid="B28">Zhu et al., 2023</xref>). By analyzing the excitation signal using MATLAB, the normalized time-domain and frequency signals are shown in <xref ref-type="fig" rid="F4">Figure 4</xref>.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Finite element model and process simulation of ultrasonic signal propagation in the model. <bold>(A)</bold> The finite element model; <bold>(B)</bold> Ultrasonic signal propagation.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g003.tif"/>
</fig>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Finite element Model parameters.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Model size (mm<sup>3</sup>)</th>
<th align="left">Density (kg/m<sup>3</sup>)</th>
<th align="left">Poisson&#x2019;s ratio</th>
<th align="left">Young modulus (GPa)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">600 &#xd7; 600 &#xd7; 2</td>
<td align="left">2690</td>
<td align="left">0.35</td>
<td align="left">69</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Ultrasonic excitation signal: <bold>(A)</bold> Time domain signal; <bold>(B)</bold> Spectrum.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g004.tif"/>
</fig>
<p>This article divides the aluminum plate into 64 areas, each with a size of 5&#xa0;cm &#xd7; 5&#xa0;cm, as shown in <xref ref-type="fig" rid="F5">Figure 5</xref>. Simulate symmetrical open cracks at different locations by drilling through holes with a diameter of 1&#xa0;cm in different areas (<xref ref-type="bibr" rid="B27">Zhang et al., 2019</xref>). Similarly, the aluminum plate can be divided into 441 areas, each with a size of 2&#xa0;cm &#xd7; 2&#xa0;cm and collect ultrasonic sensor signals through simulation. The sensor positions are shown in <xref ref-type="table" rid="T2">Table 2</xref>.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Schematic diagram of sensor and damage location.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g005.tif"/>
</fig>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Sensor positions.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Sensor number</th>
<th colspan="2" align="center">Positions (Unit: mm)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">1</td>
<td colspan="2" align="center">(145,75)</td>
</tr>
<tr>
<td align="center">2</td>
<td colspan="2" align="center">(190,75)</td>
</tr>
<tr>
<td align="center">3</td>
<td colspan="2" align="center">(235,75)</td>
</tr>
<tr>
<td align="center">4</td>
<td colspan="2" align="center">(280,75)</td>
</tr>
<tr>
<td align="center">5</td>
<td colspan="2" align="center">(325,75)</td>
</tr>
<tr>
<td align="center">6</td>
<td colspan="2" align="center">(370,75)</td>
</tr>
<tr>
<td align="center">7</td>
<td colspan="2" align="center">(415,75)</td>
</tr>
<tr>
<td align="center">8</td>
<td colspan="2" align="center">(460,75)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-1-2">
<title>3.1.2 Experimental data collection</title>
<p>The size of the aluminum plate is 600&#xa0;mm &#xd7; 600&#xa0;mm &#xd7; 2&#xa0;mm. This article pasted 8 piezoelectric elements on an aluminum plate to form a linear array (<xref ref-type="bibr" rid="B10">Liu et al., 2016</xref>). The 8 piezoelectric elements in the linear array were PZT-5 piezoelectric sheets produced by Wuxi Haiying Company, with a diameter of 10&#xa0;mm and a thickness of 0.3&#xa0;mm. The center distance between adjacent two piezoelectric elements was 45&#xa0;mm. The excitation signal with a center frequency of 200&#xa0;kHz is generated by the JPR-600C ultrasonic generator. The excitation signal is loaded onto the designated channel piezoelectric plate through the MUX-16 multi-channel conversion switch, and the ultrasonic guided wave data propagated in the aluminum plate is received through the piezoelectric plate. The sampling frequency of the system is 2&#xa0;MHz.The specific hardware models are shown in <xref ref-type="table" rid="T3">Table 3</xref>. The experimental device and the photos of the ultrasonic array data acquisition system are shown in <xref ref-type="fig" rid="F6">Figure 6</xref>. The sensor and damage location of the experimental system are consistent with the simulation system and will not be repeated here.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Main hardware parameters.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Serial number</th>
<th colspan="2" align="center">Name</th>
<th colspan="2" align="center">Model</th>
<th align="center">Number</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">1</td>
<td colspan="2" align="center">PZT-5</td>
<td colspan="2" align="center">200&#xa0;kHz</td>
<td align="center">8</td>
</tr>
<tr>
<td align="center">2</td>
<td colspan="2" align="center">BNC Cable</td>
<td colspan="2" align="center">General Cable</td>
<td align="center">8</td>
</tr>
<tr>
<td align="center">3</td>
<td colspan="2" align="center">Ultrasonic signal generator</td>
<td colspan="2" align="center">JPR-600C</td>
<td align="center">1</td>
</tr>
<tr>
<td align="center">4</td>
<td colspan="2" align="center">Channel conversion switch</td>
<td colspan="2" align="center">MUX-16</td>
<td align="center">1</td>
</tr>
<tr>
<td align="center">5</td>
<td colspan="2" align="center">Industrial notebook</td>
<td colspan="2" align="center">CF-54G</td>
<td align="center">1</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Experimental device of ultrasonic array data acquisition system.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g006.tif"/>
</fig>
</sec>
</sec>
<sec id="s3-2">
<title>3.2 Preprocessing and database settings</title>
<p>In the experiment, due to the internal noise and environmental interference of the system, noise is inevitably introduced. The noise type is white noise. &#x201c;Db5&#x201d; wavelets and four-scale decompositions are used to eliminate noise (<xref ref-type="bibr" rid="B26">Tiwari et al., 2017</xref>). <xref ref-type="fig" rid="F7">Figure 7</xref> shows the results of wavelet decomposition denoising for the received ultrasonic guided wave signal. During the excitation signal triggering process, each piezoelectric element in the piezoelectric array takes turns as a driver. When one of the piezoelectric elements acts as a driver, the other piezoelectric elements act as sensors. <xref ref-type="fig" rid="F8">Figure 8</xref> shows signals received by all sensors when the first sensor is the excitation source.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Schematic diagram of wavelet reconstruction signals: <bold>(A)</bold> Original signals; <bold>(B)</bold> Signals after wavelet reconstruction.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g007.tif"/>
</fig>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Signals received by all sensors when the first sensor is the excitation source: <bold>(A)</bold> Original signals; <bold>(B)</bold> Scattering signals.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g008.tif"/>
</fig>
<p>To eliminate the false signal generated by the boundary reflection wave, the filtered signal needs to be windowed (<xref ref-type="bibr" rid="B12">Lu and LiSong, 2016</xref>). The signal after adding the boundary reflection coefficient to the signal is expressed as:<disp-formula id="e3">
<mml:math id="m9">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:msub>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">b</mml:mi>
</mml:mrow>
<mml:mi mathvariant="normal">w</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">w</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:msub>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">b</mml:mi>
</mml:mrow>
<mml:mi mathvariant="normal">j</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mtext>&#x2003;</mml:mtext>
<mml:mi mathvariant="normal">o</mml:mi>
<mml:mi mathvariant="normal">r</mml:mi>
<mml:mtext>&#x2003;</mml:mtext>
<mml:msubsup>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:msub>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:msub>
<mml:mi mathvariant="normal">b</mml:mi>
</mml:mrow>
<mml:mi mathvariant="normal">w</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">w</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:msub>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mi mathvariant="normal">j</mml:mi>
</mml:msubsup>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi mathvariant="normal">t</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>It can be seen from the above formula that when the boundary reflection coefficient is added to the signal received by the sensor, all signals before the first boundary reflection wave reaches the <italic>i</italic>th sensor are completely retained, and the amplitudes of all signals after the first boundary reflection wave reaches the <italic>i</italic>th sensor are attenuated in exponential mode. <xref ref-type="fig" rid="F9">Figure 9</xref> shows that the boundary-reflected wave signal attenuates the original signal.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Effect of boundary reflection coefficient on the original signals: <bold>(A)</bold> No reflection coefficient is added; <bold>(B)</bold> The reflection coefficient is added.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g009.tif"/>
</fig>
<p>The good training of neural networks requires a large amount of data as support, and the array ultrasound guided wave signal data required in this article can be obtained through numerical simulation or experiments. The signal obtained from numerical simulation has high accuracy, but it takes a long time; The efficiency of experimental methods is relatively higher, but they require many experimental samples, have high costs, and are prone to environmental noise interference, resulting in relatively poor results.</p>
<p>There are two ways to enhance a data set. The first way is to collect the ultrasonic nondestructive testing data by changing the size of the finite element divided by the program in the finite element simulation process for the finite element simulation data set, as shown in <xref ref-type="fig" rid="F10">Figure 10</xref>. A certain amount of 3&#xa0;dB and &#x2212;3&#xa0;dB Gaussian white noise is added to the collected simulation data set, as shown in <xref ref-type="fig" rid="F11">Figure 11</xref>. The second method involves translating the signal up and down for the experimental data set and doubling the number of samples, as shown in <xref ref-type="fig" rid="F12">Figure 12</xref>.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Ultrasonic NDT data in simulation model: <bold>(A)</bold> [0: 1,500]; <bold>(B)</bold> [900: 1,500].</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g010.tif"/>
</fig>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>Simulation data plus noise.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g011.tif"/>
</fig>
<fig id="F12" position="float">
<label>FIGURE 12</label>
<caption>
<p>Expansion of experimental data set: <bold>(A)</bold> [0: 1,000]; <bold>(B)</bold> [80: 360].</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g012.tif"/>
</fig>
<p>To summarize, 192 and 1,323 groups of data are collected for simulation defects by changing the size of the finite element and expanding to 384 and 2,646 groups through the above method, totaling 576 and 3,969 groups of data. For the experimental data, six experiments are carried out at each defect location. Overall, 384 groups of experimental data are collected from 64 groups of defect locations and are amplified to 768 groups by the amplification method, a total of 896 groups of experimental data. The number of samples after amplification is shown in <xref ref-type="table" rid="T4">Table 4</xref>. The dataset names and sample sizes used in this experiment are shown in <xref ref-type="table" rid="T5">Table 5</xref>. The dataset used in each experimental scheme is allocated to the neural network for experimentation, with 70% of the dataset being the training set, 20% being the testing set, and 10% being the validation set.</p>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Number of samples after amplification.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Data type</th>
<th colspan="2" align="center">Simulation</th>
<th align="center">Experiment</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">Defect category</td>
<td align="center">64</td>
<td align="center">441</td>
<td align="center">64</td>
</tr>
<tr>
<td align="center">Original collection</td>
<td align="center">192</td>
<td align="center">1,323</td>
<td align="center">384</td>
</tr>
<tr>
<td align="center">Original collection</td>
<td align="center">384</td>
<td align="center">2,646</td>
<td align="center">768</td>
</tr>
<tr>
<td align="center">Total</td>
<td align="center">576</td>
<td align="center">3,969</td>
<td align="center">896</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Dataset and sample size.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Data set</th>
<th align="center">Defect category</th>
<th align="center">Number</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">S-Defect64-orgin</td>
<td align="center">Original Simulation 64 types of defects</td>
<td align="center">192</td>
</tr>
<tr>
<td align="center">S-Defect441-orgin</td>
<td align="center">Original Simulation 441 types of defects</td>
<td align="center">1,323</td>
</tr>
<tr>
<td align="center">E-Defect64-orgin</td>
<td align="center">Original Experiment 64 types of defects</td>
<td align="center">384</td>
</tr>
<tr>
<td align="center">S-Defect64-enhance</td>
<td align="center">Enhanced Simulation 64 types of defects</td>
<td align="center">576</td>
</tr>
<tr>
<td align="center">S-Defect441-enhance</td>
<td align="center">Enhanced Simulation 441 types of defects</td>
<td align="center">3,969</td>
</tr>
<tr>
<td align="center">E-Defect64-enhance</td>
<td align="center">Enhanced Experiment 441 types of defects</td>
<td align="center">896</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3-3">
<title>3.3 Models and schemes for comparison with this article</title>
<p>In order to compare and analyze the performance of the proposed model, this article compares the training effects with the following two models. The experimental environment used in this study is the Ubuntu 18.04 system, which uses a GPU NVIDIA GeForce RTX 3090. Its graphics card memory is 24&#xa0;GB, and the CPU model is Intel (R) Xeon (R) Platinum 8157 CPU @ 2.30&#xa0;GHz. The programming language is Python 3.8, and the in-depth learning model is built using Python 1.10.</p>
<sec id="s3-3-1">
<title>3.3.1 1D-CNN</title>
<p>The 1D-CNN model used for comparing experimental results is consistent with the convolutional neural network used for encoding in the Wave2vec2.0 model proposed in this paper. In the experiment, the 1D-CNN model is set with a seven-layer CNN, the step size is (5,2,2,2,2,2,2,2), the convolution kernel width is (10,3,3,3,3,2,2), the parameters of the feedforward neural network are set to 512, and the weight attenuation is 0.0001. The model uses the Adam optimizer, as shown in <xref ref-type="fig" rid="F13">Figure 13</xref>.</p>
<fig id="F13" position="float">
<label>FIGURE 13</label>
<caption>
<p>1D-CNN network structure and parameters.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g013.tif"/>
</fig>
</sec>
<sec id="s3-3-2">
<title>3.3.2 Muti-Transformer</title>
<p>Following the approach of reference (<xref ref-type="bibr" rid="B25">Sun et al., 2021</xref>), the context network based on Transformer in the Wave2Vec2.0 neural network proposed in this paper is transformed into a Muti-Transformer model that adapts to multi-channel signals. In the experiment, the number of attention modules in the Muti-Transformer is set to three layers, the number of attention heads in each layer is eight, the dropout is set to 0.3, the parameter of the feedforward neural network is set to 512, and the weight attenuation is 0.0001. The model uses the Adam optimizer. The size of pooled cores is 32 &#xd7; 32 and 8 &#xd7; 8. The size of the convolution core is 64 &#xd7; 64, as shown in <xref ref-type="fig" rid="F14">Figure 14</xref>.</p>
<fig id="F14" position="float">
<label>FIGURE 14</label>
<caption>
<p>Muti-Transformer network structure and parameters.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g014.tif"/>
</fig>
</sec>
</sec>
</sec>
<sec id="s4">
<title>4 Results and analysis</title>
<sec id="s4-1">
<title>4.1 Wav2Vec2.0 neural network training and result analysis</title>
<p>Considering the experimental requirements of the defect guided wave positioning model based on the Wav2vec2.0 neural network, the open-source deep learning framework Pytorch is selected as the main framework when conducting experiments related to s deep learning. The specific experimental environment and computer configuration are shown in <xref ref-type="table" rid="T6">Table 6</xref> below:</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Configuration of computer deep learning environment used in the experiment.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">CPU</th>
<th align="center">Intel(R) Xeon(R)Platinum 8260C CPU @ 2.30&#xa0;GHz</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">GPU</td>
<td align="center">NVIDIA GeForce RTX 3090</td>
</tr>
<tr>
<td align="center">Memory capacity</td>
<td align="center">86&#xa0;GB</td>
</tr>
<tr>
<td align="center">Graphics memory capacity</td>
<td align="center">24&#xa0;GB</td>
</tr>
<tr>
<td align="center">Operating system</td>
<td align="center">Ubuntu 18.04</td>
</tr>
<tr>
<td align="center">CUDA version</td>
<td align="center">11.3</td>
</tr>
<tr>
<td align="center">CuDNN version</td>
<td align="center">8</td>
</tr>
<tr>
<td align="center">Python version</td>
<td align="center">3.8</td>
</tr>
<tr>
<td align="center">Pytorch version</td>
<td align="center">1.10</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In this paper, relative entropy is chosen as the optimized loss function. Relative entropy (KL divergence), also known as cross entropy, is used to represent the difference between two probability distributions. When two random distributions are the same, their relative entropy is zero. When the difference between two random distributions increases, their relative entropy also increases.<disp-formula id="e4">
<mml:math id="m10">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2225;</mml:mo>
<mml:mi>q</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
</p>
<p>Expanding relative entropy yields Eq. <xref ref-type="disp-formula" rid="e5">5</xref>:<disp-formula id="e5">
<mml:math id="m11">
<mml:mrow>
<mml:mtable columnalign="left">
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>&#x2225;</mml:mo>
<mml:mi>q</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="|">
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mfenced open="[" close="]" separators="|">
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>n</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mi>log</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>p</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>q</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>
</p>
<p>In order to evaluate the model, this article uses the E-Defect64-enhance dataset provided in <xref ref-type="sec" rid="s3-2">Section 3.2</xref> as deep learning experimental data and calculates the accuracy and loss values of the model. From <xref ref-type="fig" rid="F15">Figures 15A, B</xref>, it can be seen that in the training set, when epochs reach the 600th time, the accuracy rate is the highest, and the accuracy rate of the training samples reaches 100%.</p>
<fig id="F15" position="float">
<label>FIGURE 15</label>
<caption>
<p>Model iteration results. <bold>(A)</bold> Training accuracy curve; <bold>(B)</bold> Training loss curve; <bold>(C)</bold> Test accuracy curve; <bold>(D)</bold> Test loss curve.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g015.tif"/>
</fig>
<p>For the test set, place the test set data into the trained model, and the model testing iteration results are shown in <xref ref-type="fig" rid="F15">Figures 15C, D</xref> The results obtained from damage localization and quantitative identification are shown in <xref ref-type="fig" rid="F16">Figure 16</xref> When epochs reach the end, the accuracy is the highest and the error is the smallest. At this point, the accuracy rate of the training sample reaches 98.46%. From <xref ref-type="fig" rid="F16">Figure 16</xref>, it can be seen that among the 64 types of test samples, only one damage location was identified incorrectly, while the other samples were correctly tested with accuracy rate of 98.46%. The misidentified samples identified the defect at 47 as 40.</p>
<fig id="F16" position="float">
<label>FIGURE 16</label>
<caption>
<p>Results from damage location and quantitative identification.</p>
</caption>
<graphic xlink:href="fmats-10-1212909-g016.tif"/>
</fig>
</sec>
<sec id="s4-2">
<title>4.2 Analysis of deep learning network models</title>
<sec id="s4-2-1">
<title>4.2.1 Network model noise resistance</title>
<p>To verify the noise resistance of the neural network model, using the S-Defect64-orgin and E-Defect64-orgin datasets respectively, different network models were used for training, with 1,200 iterations. The positioning accuracy is shown in <xref ref-type="table" rid="T7">Table 7</xref>. Among them, the positioning accuracy of 1D-CNN, Muti Transformer, and Wav2vec2.0 neural network models in the S-Defect64-orgin dataset is 100%. The positioning accuracy in the E-Defect64-orgin dataset is 87.5%, 92.78%, and 96.87%, respectively. It can be seen that the positioning accuracy of the model in the simulation dataset has significantly improved compared to the experimental dataset, because the noise of the ultrasonic guided wave signal collected in the experiment is significantly higher than that in the simulation dataset. <xref ref-type="table" rid="T7">Table 7</xref> shows that the high-precision detection network proposed in this paper has strong noise resistance and excellent robustness for datasets with different signal-to-noise ratios.</p>
<table-wrap id="T7" position="float">
<label>TABLE 7</label>
<caption>
<p>Effect of experimental data and simulation data on recognition accuracy.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Data set</th>
<th align="center">ID-CNN (%)</th>
<th align="center">Muti-transformer (%)</th>
<th align="center">Wav2Vec2.0 (%)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">S-Defect64-orgin</td>
<td align="center">100</td>
<td align="center">100</td>
<td align="center">100</td>
</tr>
<tr>
<td align="center">E-Defect64-orgin</td>
<td align="center">87.50</td>
<td align="center">92.38</td>
<td align="center">96.87</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4-2-2">
<title>4.2.2 Accuracy of network model positioning results</title>
<p>To verify the accuracy of the neural network model in identifying defect locations, different network models were trained using the S-Defect64-orgin and S-Defect441-orgin datasets, with 1,200 iterations. The scale of defects corresponding to different datasets varies. The defect localization scale of the S-Defect64-orgin dataset is 5&#xa0;cm, while the defect localization scale of the S-Defect441-orgin dataset is 2&#xa0;cm. The positioning accuracy of different neural network models is shown in <xref ref-type="table" rid="T8">Table 8</xref>. Among them, the positioning accuracy of 1D-CNN, Muti Transformer, and Wav2vec2.0 neural network models in the S-Defect64-orgin dataset is 100%. The positioning accuracy in the S-Defect441-orgin dataset is 89.84%, 96.09%, and 98.46%, respectively. It can be seen that the positioning accuracy of the three models in the S-Defect64-orgin dataset is significantly improved compared to the S-Defect441-org dataset. This is because the signal category in the S-Defect441-org dataset is significantly higher than that in the S-Defect64-orgin dataset. The three models can fit the function well at 1,200 iterations, but the fitting ability of the model in the S-Defect441-org dataset did not reach its optimal level. <xref ref-type="table" rid="T8">Table 8</xref> also proves that the Wav2vec2.0 neural network has higher performance in identifying defect locations with higher accuracy.</p>
<table-wrap id="T8" position="float">
<label>TABLE 8</label>
<caption>
<p>Impact of dataset size on recognition accuracy.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Data set</th>
<th align="center">ID_CNN (%)</th>
<th align="center">Muti-transformer (%)</th>
<th align="center">Wav2Vec2.0 (%)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">S_Defect64_orgin</td>
<td align="center">100</td>
<td align="center">100</td>
<td align="center">100</td>
</tr>
<tr>
<td align="center">S_Defect441_orgin</td>
<td align="center">89.84</td>
<td align="center">96.09</td>
<td align="center">98.46</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4-2-3">
<title>4.2.3 The generalization ability of neural networks</title>
<p>To verify the generalization ability of neural networks for different datasets, using the S-Defect64-orgin, E-Defect64-orgin datasets, and the enhanced S-Defect64-enhancer, E-Defect64-enhancer datasets, different network models were used for training, with 1,200 iterations. The positioning accuracy results are shown in <xref ref-type="table" rid="T9">Table 9</xref>. The positioning accuracy of the 1D-CNN, Muti Transformer, and Wav2Vec2.0 models in the S-Defect64-enhance dataset is 100%. The positioning accuracy of the 1D-CNN, Muti Transformer, and Wav2Vec2.0 models in the E-Defect64-enhanc dataset is 90.62%, 95.31%, and 98.46%, respectively. It can be seen that the data augmentation method has significantly improved the positioning accuracy of various network structures. Similarly, this also indicates that the wave2vec2.0 neural network has better generalization ability.</p>
<table-wrap id="T9" position="float">
<label>TABLE 9</label>
<caption>
<p>Effect of data enhancement methods on recognition accuracy.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Data set</th>
<th align="center">ID_CNN (%)</th>
<th align="center">Muti-transformer (%)</th>
<th align="center">Wav2Vec2.0 (%)</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="center">S_Defect64_orgin</td>
<td align="center">93.75</td>
<td align="center">100</td>
<td align="center">100</td>
</tr>
<tr>
<td align="center">S_Defect64_enhance</td>
<td align="center">100</td>
<td align="center">100</td>
<td align="center">100</td>
</tr>
<tr>
<td align="center">E_Defect64_orgin</td>
<td align="center">87.50</td>
<td align="center">92.38</td>
<td align="center">96.87</td>
</tr>
<tr>
<td align="center">E_Defect64_enhance</td>
<td align="center">90.62</td>
<td align="center">95.31</td>
<td align="center">98.46</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
</sec>
<sec sec-type="conclusion" id="s5">
<title>5 Conclusion</title>
<p>In this paper, a method based on the Wav2Vec2.0 neural network is proposed to directly locate defects in array ultrasonic testing signals. This method does not require knowledge of the <italic>a priori</italic> condition of the sample sound velocity or the feature extraction of ultrasonic scattering signals. The study draws the following conclusion:<list list-type="simple">
<list-item>
<p>(1) The positioning accuracy of the Wav2Vec2.0 neural network proposed in this article is 98.46%, and the positioning accuracy is superior to Muti-Transformer and ID-CNN.</p>
</list-item>
<list-item>
<p>(2) A metal plate structure piezoelectric array ultrasonic guided wave data acquisition scheme based on numerical simulation and on-site experiments was proposed, and the obtained piezoelectric array ultrasonic guided wave data processing method was used to produce a piezoelectric array ultrasonic guided wave dataset.</p>
</list-item>
<list-item>
<p>(3) The analysis of the impact on the positioning accuracy of deep learning models shows that:</p>
<list list-type="simple">
<list-item>
<p>1) The Wav2vec2.0 neural network has stronger fitting ability than other networks in experimental datasets, indicating that the method proposed in this paper has stronger robustness</p>
</list-item>
<list-item>
<p>2) The Wav2vec2.0 neural network has stronger fitting ability in larger datasets, which means it has higher recognition performance for smaller scale defects.</p>
</list-item>
<list-item>
<p>3) The use of data augmentation methods can improve model performance. The Wav2vec2.0 neural network has better recognition performance in enhanced datasets and stronger generalization ability.</p>
</list-item>
</list>
</list-item>
</list>
</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The raw data supporting the conclusion of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="s7">
<title>Author contributions</title>
<p>Data curation, XL, YM, YX, and ZW; Investigation, HZ and YX; Methodology, SL and GF; Project administration, HZ and ZW; Resources, LQ and HZ; Software, SL; Supervision, GF, XL, and ZW; Validation, LQ and YX; Visualization, XL and YM; Writing&#x2013;original draft, SL and YM; Writing&#x2013;review and editing, LQ and GF. All authors contributed to the article and approved the submitted version.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>This project was funded by the National Natural Science Foundation of China (No. 12104290).</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of interest</title>
<p>SL was employed by Shanghai Urban Railway Operation Co., Ltd. XL was employed by Wuhan Ruimin Testing Technology Co., Ltd. ZW was employed by Luoyang Sunrui Rubber and Plastic Technology Co., Ltd.</p>
<p>The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Alexei</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Henry</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Abdelrahman</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Michael</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020</year>). <source>Wav2vec 2.0: A framework for self-supervised learning of speech representations facebook ai research</source>. <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2006.11477">https://arxiv.org/abs/2006.11477</ext-link>.</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Caputo</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>De Luca</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Perfetto</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lamanna</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Damage detection in flat panels by guided waves based artificial neural network trained through finite element method</article-title>. <source>Materials</source> <volume>14</volume> (<issue>24</issue>), <fpage>7602</fpage>. <pub-id pub-id-type="doi">10.3390/ma14247602</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2018</year>). <source>Application and challenges of signal processing techniques for Lamb waves structural integrity evaluation: Part B-defects imaging and recognition techniques</source>. <publisher-loc>London, UK</publisher-loc>: <publisher-name>Intech open</publisher-name>.</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Drai</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Khelil</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Benchaala</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>Time frequency and wavelet transform applied to selected problems in ultrasonics NDE</article-title>. <source>NDT E Int.</source> <volume>35</volume> (<issue>8</issue>), <fpage>567</fpage>&#x2013;<lpage>572</lpage>. <pub-id pub-id-type="doi">10.1016/s0963-8695(02)00041-5</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ewald</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>VenkatAsokkumar</surname>
<given-names>R. S. A.</given-names>
</name>
<name>
<surname>Benedictus</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Boller</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Groves</surname>
<given-names>R. M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Perception modelling by invariant representation of deep learning for automated structural diagnostic in aircraft maintenance: A study case using DeepSHM</article-title>. <source>Mech. Syst. Signal Process.</source> <volume>165</volume>, <fpage>108153</fpage>. <pub-id pub-id-type="doi">10.1016/j.ymssp.2021.108153</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Damage localization and quantification of truss structure based on electromechanical impedance technique and neural network</article-title>. <source>Shock Vib.</source> <volume>2014</volume>, <fpage>727404</fpage>. <pub-id pub-id-type="doi">10.1155/2014/727404</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Khumaidi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Yuniarno</surname>
<given-names>E. M.</given-names>
</name>
<name>
<surname>Purnomo</surname>
<given-names>M. H.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Welding defect classification based on convolution neural network(CNN) and Gaussian kernel</article-title>. <source>IEEE</source>, <fpage>261</fpage>&#x2013;<lpage>265</lpage>.</citation>
</ref>
<ref id="B8">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>Self-supervised representation learning from videos for facial action unit detection[C]</article-title>,&#x201d; in <source>IEEE conference on computer vision and pattern recognition(CVPR) 2019, oral</source> (<publisher-name>IEEE</publisher-name>).</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Quantitative detection of lamination defect in thin-walled metallic pipe by using circumferential Lamb waves based on wavenumber analysis method</article-title>. <source>NDT E Int. Indep. Nondestruct. Test. Eval.</source> <volume>102</volume>, <fpage>56</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1016/j.ndteint.2018.11.005</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>SunSong</surname>
<given-names>K. G.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Damage localization in aluminum plate with compact rectangular phased piezoelectric transducer array</article-title>. <source>Mech. Syst. Signal Process.</source> <volume>70</volume>, <fpage>625</fpage>&#x2013;<lpage>636</lpage>. <pub-id pub-id-type="doi">10.1016/j.ymssp.2015.09.022</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>A multi-delay-and-sum imaging algorithm for damage detection using piezoceramic transducers</article-title>. <source>J. Intelligent Material Syst. Struct.</source> <volume>28</volume> (<issue>9</issue>), <fpage>1150</fpage>&#x2013;<lpage>1159</lpage>. <pub-id pub-id-type="doi">10.1177/1045389x16666184</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>LiSong</surname>
<given-names>Y. G.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>A delay-and-Boolean-ADD imaging algorithm for damage detection with a small number of piezoceramic transducers</article-title>. <source>Smart Mater. Struct.</source> <volume>25</volume> (<issue>9</issue>), <fpage>095030</fpage>. <pub-id pub-id-type="doi">10.1088/0964-1726/25/9/095030</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ma</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Realizing broadband sub-wavelength focusing and a high intensity enhancement with a space-time synergetic modulated acoustic prison</article-title>. <source>J. Mater. Chem. C</source> <volume>8</volume> (<issue>28</issue>), <fpage>9511</fpage>&#x2013;<lpage>9519</lpage>. <pub-id pub-id-type="doi">10.1039/d0tc01984d</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ma</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Acoustic focusing and imaging via phononic crystal and acoustic metamaterials</article-title>. <source>J. Appl. Phys.</source> <volume>131</volume>, <fpage>011103</fpage>. <pub-id pub-id-type="doi">10.1063/5.0074503</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Meng</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chua</surname>
<given-names>Y. J.</given-names>
</name>
<name>
<surname>Wouterson</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Ong</surname>
<given-names>C. P. K.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Ultrasonic signal classification and imaging system for composite materials via deep convolutional neural networks</article-title>. <source>Neurocomputing</source> <volume>257</volume>, <fpage>128</fpage>&#x2013;<lpage>135</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2016.11.066</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Munir</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>H. J.</given-names>
</name>
<name>
<surname>Song</surname>
<given-names>S. J.</given-names>
</name>
<name>
<surname>Kang</surname>
<given-names>S. S.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Performance enhancement of convolutional neural network for ultrasonic flaw classification by adopting autoencoder</article-title>. <source>NDT E Int.</source> <volume>111</volume>, <fpage>102218</fpage>. <pub-id pub-id-type="doi">10.1016/j.ndteint.2020.102218</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rautela</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Senthilnath</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Moll</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Gopalakrishnan</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Combined two-level damage identification strategy using ultrasonic guided waves and physical knowledge assisted machine learning</article-title>. <source>Ultrasonics</source> <volume>115</volume>, <fpage>106451</fpage>. <pub-id pub-id-type="doi">10.1016/j.ultras.2021.106451</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rizvi</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Abbas</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Lamb wave damage severity estimation using ensemble-based machine learning method with separate model network</article-title>. <source>Smart Mater. Struct.</source> <volume>30</volume> (<issue>11</issue>), <fpage>115016</fpage>. <pub-id pub-id-type="doi">10.1088/1361-665x/ac2e1a</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sadhu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>C. W.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Wav2vec-C: A self-supervised model for speech representation learning[C]</article-title>. <source>Interspeech Conf</source>.</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shao</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zheng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Deep convolutional neural networks for thyroid tumor grading using ultrasound B-mode images</article-title>. <source>J. Acoust. Soc. Am.</source> <volume>148</volume> (<issue>3</issue>), <fpage>1529</fpage>&#x2013;<lpage>1535</lpage>. <pub-id pub-id-type="doi">10.1121/10.0001924</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shi</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Shi</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Analysis on defect classification by deep neural networks and wavelet packet transform [J]</article-title>. <source>Acta Acust.</source> <volume>41</volume> (<issue>4</issue>), <fpage>499</fpage>&#x2013;<lpage>506</lpage>.</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sikdar</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Bag of visual words based machine learning framework for disbond characterisation in composite sandwich structures using guided waves</article-title>. <source>Smart Mater. Struct.</source> <volume>30</volume> (<issue>7</issue>), <fpage>075016</fpage>. <pub-id pub-id-type="doi">10.1088/1361-665x/ac01a8</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stepinski</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ambrozinski</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Uhl</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Designing 2D arrays for SHM of planar structures: A review [C]. Conference on nondestructive characterization for composite materials</article-title>. <source>Aerosp. Eng. Civ. Infrastructure, Homel. Secur.</source> <volume>8694</volume>, <fpage>86941R</fpage>. <pub-id pub-id-type="doi">10.1117/12.2010000</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Su</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>F.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Improved damage localization and quantification of CFRP using lamb waves and convolution neural network</article-title>. <source>IEEE Sensors J.</source> <volume>19</volume> (<issue>14</issue>), <fpage>5784</fpage>&#x2013;<lpage>5791</lpage>. <pub-id pub-id-type="doi">10.1109/jsen.2019.2908838</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sun</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Xie</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>EEG classification with transformer-based models[C]</article-title>,&#x201d; in <source>2021 IEEE 3rd global conference on life Sciences and technologies (LifeTech)</source> (<publisher-name>IEEE</publisher-name>).</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tiwari</surname>
<given-names>K. A.</given-names>
</name>
<name>
<surname>Raisutis</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Samaitis</surname>
<given-names>V.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Hybrid signal processing technique to improve the defect estimation in ultrasonic non-destructive testing of composite structures</article-title>. <source>Sensors</source> <volume>17</volume> (<issue>12</issue>), <fpage>2858</fpage>. <pub-id pub-id-type="doi">10.3390/s17122858</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Xia</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Adaptive sparse reconstruction of damage localization via Lamb waves for structure health monitoring</article-title>. <source>Computing</source> <volume>101</volume> (<issue>6</issue>), <fpage>679</fpage>&#x2013;<lpage>692</lpage>. <pub-id pub-id-type="doi">10.1007/s00607-018-00694-0</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Xiang</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Fan</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Super-resolution ultrasonic Lamb wave imaging based on sign coherence factor and total focusing method</article-title>. <source>Mech. Syst. Signal Process.</source> <volume>190</volume>, <fpage>110121</fpage>. <pub-id pub-id-type="doi">10.1016/j.ymssp.2023.110121</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>