Open Access Article
This Open Access Article is licensed under a
Creative Commons Attribution 3.0 Unported Licence

Pesticide detection combining the Wasserstein generative adversarial network and the residual neural network based on terahertz spectroscopy

Ruizhao Yanga, Yun Lic, Binyi Qin*ab, Di Zhaoa, Yongjin Gana and Jincun Zhenga
aSchool of Physics and Telecommunication Engineering, Yulin Normal University, Yulin, China. E-mail: qby207@163.com
bGuangxi Colleges and Universities Key Laboratory of Complex System Optimization and Big Data Processing, Yulin Normal University, Yulin, China
cCollege of Chemistry and Food Science, Yulin Normal University, Yulin, China

Received 15th September 2021 , Accepted 20th December 2021

First published on 11th January 2022


Abstract

Feature extraction is a key factor to detect pesticides using terahertz spectroscopy. Compared to traditional methods, deep learning is able to obtain better insights into complex data features at high levels of abstraction. However, reports about the application of deep learning in THz spectroscopy are rare. The main limitation of deep learning to analyse terahertz spectroscopy is insufficient learning samples. In this study, we proposed a WGAN-ResNet method, which combines two deep learning networks, the Wasserstein generative adversarial network (WGAN) and the residual neural network (ResNet), to detect carbendazim based on terahertz spectroscopy. The Wasserstein generative adversarial network and pretraining model technology were employed to solve the problem of insufficient learning samples for training the ResNet. The Wasserstein generative adversarial network was used for generating more new learning samples. At the same time, pretraining model technology was applied to reduce the training parameters, in order to avoid residual neural network overfitting. The results demonstrate that our proposed method achieves a 91.4% accuracy rate, which is better than those of support vector machine, k-nearest neighbor, naïve Bayes model and ensemble learning. In summary, our proposed method demonstrates the potential application of deep learning in pesticide residue detection, expanding the application of THz spectroscopy.


1 Introduction

The problem of pesticide residues has always been a top priority for the public. The excessive use of pesticides in crop planting leads to the frequent occurrence of pesticide residues in crop products, which threatens human health. Accurate and rapid pesticide residue detection is of great significance for food quality and safety control. High-performance liquid chromatography,1,2 gas chromatography,3 capillary electrophoresis4 and immunoassay techniques5 are common methods for the detection of pesticide residues. However, these methods are destructive, harmful solvents are required and they are time-consuming.6–8 Therefore, a rapid and effective analytical method is urgently needed for the determination of pesticide residues.

Terahertz (THz) spectroscopy is considered to be a promising detection method due to its low-energy, high resolution and penetrability.9–11 Because it is sensitive to the vibrational modes, THz spectra contain abundant useful information on the vibrational modes of the target. In recent years, some researchers have combined THz fingerprints and chemometric techniques to detect foreign bodies,12,13 toxic and harmful compounds,10,14,15 pesticides,16–18 antibiotics,19,20 microorganisms21–23 and adulteration.24 The above studies show that data feature extraction is a key factor affecting the detection results. Compared to traditional methods, deep learning is able to obtain better insights into complex data features at high levels of abstraction. The residual neural network (ResNet), proposed by He,25 is a deep learning network, which solves the degradation issues of traditional deep convolutional networks. It has been applied in the real-time quality assessment of pediatric MRI images,26 the clinical diagnosis of COVID-19 patients,27 the identification of cashmere and sheep wool fibers,28 rotating machinery fault diagnosis29 and so on. However, there are few reports about the application of deep learning in THz spectroscopy. The main reason is that learning samples of THz spectra are too few to meet deep learning requirements. It is known that the learning results are worse when a deep learning model is short of learning samples. Measuring more THz spectra data is one way to solve the problem of insufficient learning samples. But it is not a good approach, because it demands a higher cost and more time.

The generative adversarial network (GAN) is a sample generation model, which was proposed by Goodfellow30 in 2014. It is able to generate new samples with the same distribution as real samples to expand the size of labeled samples.31 The GAN contains a generator and a discriminator. The GAN learns the distribution of real samples during the game between a generator and a discriminator. In the process of GAN training, the generator tries its best to fit the distribution of real samples and generate new samples, while the discriminator tries its best to distinguish real samples from new samples. In recent years, GAN has been used for generating new conversation data,32 new samples for the minority class of various imbalanced datasets33 and new high-quality images.34 As there is a shortage of training data, it is difficult to train a learning model from scratch. Fine-tuning a deep learning model, which has been pretrained using a large set of labeled natural images, is a promising method to solve the shortage of learning samples. It has been applied successfully to various computer vision tasks such as food image recognition,35 mammogram image classification36 and multi-label legal document classification.37

Carbendazim is a type of broad-spectrum benzimidazole fungicide, which has been commonly employed to control plant diseases in cereals and fruits. Rice is an important food crop for human beings, which a wide area has been cultivated for. A large amount of carbendazim is used in the prevention of rice blast and rice sheath blight fungus. Studies have shown that high doses of carbendazim can damage testicles, which causes infertility.38,39 In this study, we proposed the WGAN-ResNet method, which combines two deep learning networks, the Wasserstein generative adversarial network (WGAN) and the residual neural network (ResNet), to detect carbendazim based on THz spectroscopy. The WGAN was employed to generate new learning samples, which solves the problem of learning results being worse caused by insufficient learning samples. The ResNet was applied to quantify different concentrations of carbendazim samples. At the same time, pretraining model technology was employed to reduce the training parameters in the ResNet. The results demonstrate that our proposed method shows the potential application of deep learning in pesticide residue detection, expanding the application of THz spectroscopy.

2 Experimental and theoretical methods

2.1 Experimental system and procedure

The composition of the experimental system is shown in Fig. 1. A femtosecond laser beam (with a pulse width of about 100 fs and a wavelength centered around 780 nm) is generated by an ultra-fast fiber laser. As the laser beam goes through a cubic beam splitter, it is divided into a pump beam and a probe beam. The THz beam is elicited as the pump beam, which is concentrated on a photoconductive antenna. After the THz beam goes across the sample, it carries the sample information. When the THz beam encounters the probe beam at the ZnTe crystal, the probe beam is modulated by the THz beam. And then, the modulated probe beam goes through a quarter-wave-plate (QWP) and a Wollaston prism (WP). Finally, the modulated probe beam is then detected by a set of balanced photodiodes. To reduce the absorption of the THz signal by atmospheric water vapor, the THz beam path is enclosed in a dry air purged box.
image file: d1ra06905e-f1.tif
Fig. 1 Schematic of the experimental system.

Fig. 2 shows the flow chart for the whole process. It includes the preparation of the sample, data acquisition, generating new samples and the ResNet model. Firstly, samples were made into tablets after drying and sieving. Secondly, the absorption coefficients of the samples were calculated by using the THz time-domain spectra, and then the absorption coefficients were translated to two-dimensional images. Thirdly, new samples were generated by WGAN to increase the training samples. Finally, ResNet was trained and employed to quantify the samples. The details of these procedures are described in the following sections.


image file: d1ra06905e-f2.tif
Fig. 2 Flow chart of the detection of carbendazim using WGAN-ResNet based on THz spectroscopy.

2.2 Sample preparation

Carbendazim powder, with a purity of 98%, was purchased from Adamas and used without further purification. Rice powder was purchased from a local market. First, carbendazim and rice powder were put into a vacuum drying oven. The temperature was set to 323 K, and the drying time was set to 1 hour. Then, the two powders were separately sieved with a 100 mesh sieve. After that, sample pellets, mixed with different concentrations of carbendazim and rice, were pressed into 1 mm thick tablets using a hydraulic press. Finally, 13 concentrations of carbendazim samples (0%, 2%, 4%, 6%, 8%, 10%, 15%, 20%, 25%, 30%, 40%, 50% and 100%) were prepared. The total number of samples was 429, including 33 samples of each concentration.

2.3 Data acquisition

In the experiment, the THz time-domain spectrum of dry air was used as a reference signal. To address random noise, we carried out the measurements of each reference and each sample three times. When the time-domain spectra underwent fast Fourier transformation, the amplitude and phase in the frequency domain were obtained. The samples’ refractive indices n(ω) and absorption coefficients α(ω) were calculated based on the amplitude and phase.40,41
 
image file: d1ra06905e-t1.tif(1)
 
image file: d1ra06905e-t2.tif(2)
where ω is the frequency, c is the speed of light, d is the thickness of the sample, φ(ω) is the phase difference of the sample and reference, and A(ω) is the amplitude ratio of the sample and reference.

The depth of the network is very important for the performance of the model. When the number of network layers is increased, the network can extract more complex feature patterns. But deep networks present a degradation problem: when the network depth increases, the network accuracy becomes saturated or even decreases. The ResNet is able to solve the degradation problem as the network depth increases. The ResNet is good at two-dimensional image recognition tasks.29,42–44 To use our proposed WGAN-ResNet analysis with the THz spectrum, we firstly translated a one-dimensional absorption coefficient to a two-dimensional image as follows,

 
image file: d1ra06905e-t3.tif(3)
where x is the sample absorption coefficient, which is a n-dimensional column vector, and xT is a transpose of x. Thus, A is a n × n size image.

After the above calculation, we obtained 429 images. These images were called actual images. Then, these actual images were put into the WGAN to generate 13 concentrations gradients of new image samples (0%, 2%, 4%, 6%, 8%, 10%, 15%, 20%, 25%, 30%, 40%, 50% and 100%). Each concentration was 3495 new image samples. To distinguish between new images and actual images, we called the new images generated images.

2.4 Generating new samples

To address the shortage of learning samples, we employed the WGAN for data augmentation. WGAN was proposed by Martin Arjovsky. He introduced the Earth-Mover (EM) distance, instead of KL divergence or JS divergence,45 to address the issue of GANs being hard to train.46 The EM distance is also called Wasserstein-1, which is defined by:
 
image file: d1ra06905e-t4.tif(4)
where image file: d1ra06905e-t5.tif is the set of all joint distributions γ(x,y), of which the marginals are image file: d1ra06905e-t6.tif and image file: d1ra06905e-t7.tif, respectively.

Eqn (4) can be translated based on the Kantorovich–Rubinstein duality:

 
image file: d1ra06905e-t8.tif(5)
As ‖fL ≤ 1 is replaced by ‖fLK, eqn (5) can be rewritten as:
 
image file: d1ra06905e-t9.tif(6)
If we have a parameterized family of functions, image file: d1ra06905e-t10.tif, that are all K-Lipschitz for some K, we could consider solving the problem:
 
image file: d1ra06905e-t11.tif(7)
Thus, the generator loss function of WGAN is:
 
image file: d1ra06905e-t12.tif(8)
and the discriminator loss function of WGAN is:
 
image file: d1ra06905e-t13.tif(9)

The architectures of the generator and discriminator in the WGAN are illustrated in Fig. 3. The generated images were produced by the generator, when random noise was input into the generator. And then an image was decided to be either an actual image or a generated image by the discriminator. The generator was trained to generate images which are more similar to actual images, and the discriminator was also trained to discriminate between images more accurately. The generator and the discriminator were adversarial with each other. When the discriminator could not make a decision on whether an image is a generated image or an actual image, the training was finished.


image file: d1ra06905e-f3.tif
Fig. 3 The architectures of the generator and a discriminator in the WGAN. (a) The architecture of the generator. (b) The architecture of the discriminator.

2.5 ResNet model

A deep network can obtain better feature extraction capability than a shallow network. However, a degradation problem will occur when the network depth increases. To deal with the degradation problem, He et al. proposed the ResNet.25 They make these stacked layers fit a residual mapping, taking the place of making these layers fit a desired underlying mapping.

Let the desired underlying mapping be denoted as H(x), where x denotes the input of the first of these layers. The stacked nonlinear layers fit another mapping of F(x)≔H(x) − x. So, the original mapping is recast into F(x) + x. At last, the formulation of F(x) + x can be realized by feedforward neural networks with identity shortcut connections, shown in Fig. 4.


image file: d1ra06905e-f4.tif
Fig. 4 Residual learning: a building block from ref. 25.

In this study, we selected two ResNets (an 18 layer ResNet and a 152 layer ResNet). To satisfy our classification task, we changed the last fully connected layer output of the original ResNet from 1000 to 13. The network architecture of the 18 layer ResNet and 152 layer ResNet are listed in Table 1. The number beside the bracket represents the number of blocks stacked. Down-sampling was performed by conv3_1, conv4_1, and conv5_1 with a stride of 2.

Table 1 Architectures of the two ResNets
Layer name Output size 18 layer ResNet 152 layer ResNet
conv1 112 × 112 7 × 7.64, stride 2
conv2_x 56 × 56 3 × 3 max pool, stride 2
image file: d1ra06905e-t14.tif image file: d1ra06905e-t15.tif
conv3_x 28 × 28 image file: d1ra06905e-t16.tif image file: d1ra06905e-t17.tif
conv4_x 14 × 14 image file: d1ra06905e-t18.tif image file: d1ra06905e-t19.tif
conv5_x 7 × 7 image file: d1ra06905e-t20.tif image file: d1ra06905e-t21.tif
fc 1 × 1 Average pool, 13-d fc, softmax Average pool, 13-d fc, softmax


Pretraining model technology can be used to train a large target network without overfitting when the target dataset is smaller than the base dataset.47 In the experiment, we first trained the ResNet based on the ImageNet dataset, and then we transferred it to our target network to be trained on a target dataset. ImageNet is a large-scale hierarchical image database, which contains 3.2 million cleanly annotated images spread over 5247 categories.48 It has been the most influential dataset in computer vision.25,49–51

3 Results and discussion

3.1 Spectral analysis of the samples

In this paper, the THz absorption coefficients of the samples were obtained using the THz-TDS system. Fig. 5 shows THz absorption coefficients of 13 concentration gradients of the samples in the range of 0.4 THz to 1.4 THz. In Fig. 5, rice is the 0% concentration sample, and carbendazim is the 100% concentration sample. It can be observed that carbendazim exhibits two distinct absorption peaks at 1.15 THz and 1.32 THz. However, as the sample concentration decreases, the absorption peaks of carbendazim become progressively less obvious. Thus, it is hard to accurately distinguish between rice and a low concentration sample by the absorption coefficient. For further study, the samples’ absorption coefficients were translated to two-dimensional images using eqn (3).
image file: d1ra06905e-f5.tif
Fig. 5 The absorption coefficients of 13 concentration gradients of the samples.

3.2 The new images generated by the WGAN

The actual images and generated images are shown in Fig. 6. The actual images (Fig. 6(a)–(c)) are placed in the first row, arranged in the order of 0%, 2% and 100% concentration. The generated images (Fig. 6(d)–(f)) are placed in the second row, arranged in the order of 0%, 2% and 100% concentration. It is clear that the generated images are similar to the actual images.
image file: d1ra06905e-f6.tif
Fig. 6 (a)–(c) The actual images and (d)–(f) generated images (0%, 2% and 100% concentration, respectively).

The structural similarity index (SSIM)52 was employed to quantificationally measure the similarity of the generated and actual images. The more similar the generated and actual images, the closer the SSIM value is to 1. For the 0%, 2% and 100% concentration samples, the SSIM values are 0.92, 0.94 and 0.98, respectively. The results mean that the generated images keep the key features of the actual images well.

3.3 The identification results of the WGAN-ResNet

The identification results of the ResNet and WGAN-ResNet are displayed in Fig. 7. The accuracy rate is defined as follows:
 
image file: d1ra06905e-t22.tif(10)

image file: d1ra06905e-f7.tif
Fig. 7 The identification results of the ResNet and WGAN-ResNet.

For the ResNet, the quantification accuracy rate of the 152 layer ResNet is 2.57% higher than that of the 18 layer. For the WGAN-ResNet, the quantification accuracy rate of the 152 layer model is 5.98% higher than that of the 18 layer. For both the ResNet and WGAN-ResNet, the 152 layer models have better performances than those of the 18 layer, which is consistent with the previous report.25 This indicates that the deeper network has better feature extraction ability.

The 18 layer ResNet has 1.8 GFLOPs, and the 152 layer ResNet has 11.3 GFLOPs.25 This means that the model complexity of the 152 layer ResNet is higher than that of the 18 layer ResNet. Thus, training the ResNet models requires a large number of samples. When the ResNet was trained without the WGAN, the number of training samples was only 429. The number of training samples could not meet the needs of model training. In order to make more samples for training the ResNet models, WGAN was employed. The results show that the quantification accuracy rates of WGAN-ResNet are better than ResNet. For the 18 layer model, the quantification accuracy rate of the WGAN-ResNet is 0.86% higher than that of the ResNet. And for the 152 layer model, the identify accuracy rate of the WGAN-ResNet is 4.27% higher than that of the ResNet. This is because the ResNet model parameters were trained sufficiently using the new images generated by the WGAN. This also indicates that the WGAN is a feasible way to augment data for a shortage of learning samples.

To avoid model overfitting, we introduced pretraining model technology to reduce the ResNet model parameters. The task which trains the ResNet based on the ImageNet dataset is called task A. Task B is the task that trained the ResNet based on our dataset. The network AnB_ResNet_18 is a network architecture based on the 18 layer ResNet, and the model parameters of the first n layers are copied from task A and frozen, while the parameters of the remaining 5 – n layers are randomly initialized and trained based on task B. The network AnB_ResNet_152 is a network architecture based on the 152 layer ResNet, and the model parameters of the first n layers are copied from task A and frozen, while the parameters of the remaining 5 – n layers are randomly initialized and trained based on task B. To train the ResNet networks, we used the Adam method with a learning rate of 1e−4 and a batch size of 128.

The quantification accuracy rates with different numbers of frozen layers are displayed in Fig. 8. When the frozen layer n is 0, it represents that the model parameters in the five layers shown in Table 1 will change with training. When the frozen layer n is 1, the first layer conv1 shown in Table 1 is frozen. And when the frozen layer n is 2, the first two layers conv1 and conv2_x are frozen, and so on. When the frozen layer n is 5, it means that the model parameters in conv1, conv2_x, conv3_x, conv4_x and conv5_x cannot be changed by training. As shown in Fig. 8, the accuracy rate rises as the frozen layers increase at the beginning. However, as the frozen layers increase further, the accuracy rate falls. The features of samples in our dataset are quit different from those in the ImageNet dataset. When the frozen layers increase gradually, the co-adaptation and the feature extraction capability deteriorates. The best accuracy rate is 91.4%, which is obtained by using the 152 layer ResNet with the first layer frozen.


image file: d1ra06905e-f8.tif
Fig. 8 Accuracy rate of the 18 layer and 152 layer WGAN-ResNet with different numbers of frozen layers.

For further analysis, we compared our proposed WGAN-ResNet with a support vector machine (SVM),53 k-nearest neighbor (KNN),54 naïve Bayes model55 and ensemble learning.56 The hyper-parameters of SVM were optimized by a genetic algorithm (GA)57 and particle swarm optimization (PSO).58,59 SVM, KNN, the naïve Bayes model and ensemble learning can be considered to be shallow learning. The sample features used by shallow learning are low-level features (generally edge texture features). Different from SVM, KNN, the naïve Bayes model and ensemble learning, the ResNet is not only able to extract low-level features, but is also able to extract high-level features.25 The high-level features are based on low-level features, which have richer semantic information. The accuracy rates of the above methods are shown in Fig. 9. Our proposed WGAN-ResNet achieved a 91.4% accuracy rate, which is higher than those of the compared methods.


image file: d1ra06905e-f9.tif
Fig. 9 Comparison of the WGAN-ResNet with SVM, KNN, a naïve Bayes model and ensemble learning.

4 Conclusion

In this paper, we proposed a new pesticide residue terahertz spectroscopy detection method, which combines two deep learning networks, the WGAN and ResNet, together. Our method demonstrates the potential application of deep learning in the detection of pesticide residues and expands the application of THz spectroscopy. Different from the previous spectral analysis methods, we translated the one-dimensional absorption coefficients into two-dimensional images. To solve the problem of insufficient learning samples, we employed a WGAN to generate new learning samples and pretraining model technology to reduce the training parameters in the ResNet. By using our proposed WGAN-ResNet, the best accuracy rate is 91.4%, which is better than those of SVM, KNN, the naïve Bayes model and ensemble learning. In summary, our proposed method taps into the potential application of deep learning in pesticide residue detection, expanding the application of THz spectroscopy.

Trace detection and the extraction of more features from complex samples will be our future research directions. To extract more features from complex samples, we will use the information fusion method. This fuses more spectrum parameters (such as the refractive index and dielectric constant) together as the input of the WGAN-ResNet. For trace detection, we will enhance the interaction between the pesticide and the terahertz spectrum using metamaterials.

Author contributions

B. Q. and R. Y. did the terahertz experiments. B. Q and R. Y. provided the funding. B. Q. and Y. L. were the major contributors to writing the manuscript. B. Q. and R. Y. developed the idea and supervised the project. D. Z., Y. G. and J. Z. reviewed and edited the manuscript.

Conflicts of interest

There are no conflicts to declare.

Acknowledgements

This research was funded by the National Natural Science Foundation of China (grant no. 62041111), the Natural Science Foundation of Guangxi Province (grant no. 2019GXNSFBA245076), the Guangxi Key Laboratory of Automatic Detecting Technology and Instruments Foundation (grant no. YQ19208), the Middle-aged and Young Teachers Basic Ability Promotion Project of Guangxi (grant no. 2021KY0585), the Major Cooperative Project between Yulin Municipal Government and Yulin Normal University (grant no. YLSXZD2019015) and the Doctoral Scientific Research Foundation of Yulin Normal University (grant no. G2019K02).

References

  1. B. A. Khan, A. Farid, M. R. Asi, H. Shah and A. K. Badshah, Food Chem., 2009, 114, 286–288 CrossRef CAS.
  2. H. U. Yongping, X. U. Kegong, Y. U. Shuning, H. J. Cui and A. Xin, Anim. Husb. Vet. Med., 2017, 49, 47–51 Search PubMed.
  3. Y. R. Tahboub, M. F. Zaater and Z. A. Al-Talla, J. Chromatogr. A, 2005, 1098, 150–155 CrossRef CAS PubMed.
  4. D. Moreno-Gonzíćlez, F. J. Lara, L. Gíćmiz-Gracia and A. Garcíła-Campa, J. Chromatogr. A, 2014, 1360, 1–8 CrossRef PubMed.
  5. Y. C. Q. H. L. Zhang, Biosens. Bioelectron., 2016, 79, 430–434 CrossRef PubMed.
  6. C. Chíćferpericíćs, Í. Maquieira and R. Puchades, TrAC, Trends Anal. Chem., 2010, 29, 1038–1049 CrossRef.
  7. M. V. Barbieri, C. Postigo, N. Guillem-Argiles, L. S. Monllor-Alcaraz, J. I. Simionato, E. Stella, D. Barcelo and L. Miren, Sci. Total Environ., 2019, 653, 958–967 CrossRef CAS PubMed.
  8. F. Qu, L. Lin, C. Cai, B. Chu and P. Nie, Food Chem., 2021, 334, 127474 CrossRef CAS PubMed.
  9. G. Ok, K. Park, H. J. Kim, H. S. Chun and S. W. Choi, Appl. Opt., 2014, 53, 1406–1412 CrossRef PubMed.
  10. S. Lu, X. Zhang, Z. Zhang, Y. Yang and Y. Xiang, Food Chem., 2016, 494–501 CrossRef CAS PubMed.
  11. B. Qin, Z. Li, Z. Luo, H. Zhang and L. Yun, J. Spectrosc., 2017, 2017(2017–8-9), 1–8 Search PubMed.
  12. Z. Huo, L. I. Zhi, C. Tao and B. Qin, Spectrochim. Acta, Part A, 2017, 184, 335–341 CrossRef PubMed.
  13. C. Joerdens and M. Koch, Opt. Eng., 2008, 47, 281–291 Search PubMed.
  14. W. Liu, P. Zhao, C. Wu, C. Liu, J. Yang and Z. Lei, Food Chem., 2019, 293, 213–219 CrossRef CAS PubMed.
  15. S. H. Baek, H. B. Lim and H. S. Chun, J. Agric. Food Chem., 2014, 62, 5403 CrossRef CAS PubMed.
  16. B. Qin, L. Zhi, Z. Luo, L. Yun and Z. Huo, Opt. Quantum Electron., 2017, 49, 244 CrossRef.
  17. B. Qin, Z. Li, F. Hu, C. Hu, T. Chen, H. Zhang and Y. Zhao, IEEE Trans. Terahertz Sci. Technol., 2018, 8, 149–154 CAS.
  18. P. Nie, F. Qu, L. Lin, Y. He and L. Huang, Int. J. Mol. Sci., 2021, 22, 1–13 Search PubMed.
  19. Y. Long, B. Li and H. Liu, Appl. Opt., 2018, 57, 544 CrossRef CAS PubMed.
  20. J. Qin, L. Xie and Y. Ying, Food Chem., 2017, 224, 262–269 CrossRef CAS PubMed.
  21. S. J. Park, J. T. Hong, S. J. Choi, H. S. Kim, W. K. Park, S. T. Han, J. Y. Park, S. Lee, D. S. Kim and Y. H. Ahn, Sci. Rep., 2014, 4, 4988 CrossRef CAS PubMed.
  22. X. Yang, D. Wei, S. Yan, Y. Liu, S. Yu, M. Zhang, Z. Yang, X. Zhu, Q. Huang and H. L. Cui, J. Biophotonics, 2016, 1050–1058 CrossRef CAS PubMed.
  23. X. Yang, K. Yang, Y. Luo and W. Fu, Appl. Microbiol. Biotechnol., 2016, 100, 5289–5299 CrossRef CAS PubMed.
  24. J. Liu, Opt. Quantum Electron., 2017, 49, 1 CrossRef CAS.
  25. K. He, X. Zhang, S. Ren and J. Sun, 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2016, pp. 770–778 Search PubMed.
  26. S. Liu, K.-H. Thung, W. Lin, P.-T. Yap and D. Shen, IEEE Trans. Image Process., 2020, 29, 7697–7706 Search PubMed.
  27. I. Shiri, A. Akhavanallaf, A. Sanaat, Y. Salimi and H. Zaidi, Eur. Radiol., 2021, 31, 1420–1431 CrossRef CAS PubMed.
  28. J. Luo, K. Lu, Y. Chen and B. Zhang, Micron, 2021, 143, 103023 CrossRef PubMed.
  29. B. Peng, H. Xia, X. Lv, M. Annor-Nyarko and J. Zhang, Appl. Intell., 2021, 6, 1–8 Search PubMed.
  30. I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville and Y. Bengio, Generative Adversarial Nets, MIT Press, 2014 Search PubMed.
  31. A. Creswell, T. White, V. Dumoulin, K. Arulkumaran, B. Sengupta and A. A. Bharath, IEEE Signal Process. Mag., 2018, 35, 53–65 Search PubMed.
  32. J. Li, W. Monroe, T. Shi, S. Jean, A. Ritter and D. Jurafsky, Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2017 Search PubMed.
  33. G. Douzas and F. Bacao, Expert Syst. Appl., 2018, 91, 464–471 CrossRef.
  34. T. Karras, T. Aila, S. Laine and J. Lehtinen, 2017, arXiv e-prints, arXiv:1710.10196.
  35. K. Yanai and Y. Kawano, 2015 IEEE International Conference on Multimedia Expo Workshops, ICMEW, 2015, pp. 1–6 Search PubMed.
  36. K. Clancy, S. Aboutalib, A. Mohamed, J. Sumkin and S. Wu, J. Digit. Imag., 2020, 33, 1257–1265 CrossRef PubMed.
  37. D. Song, A. Vold, K. Madan and F. Schilder, Inf. Syst., 2021, 101718 CrossRef.
  38. T. A. Aire, Anat. Embryol., 2005, 210, 43–49 CrossRef CAS PubMed.
  39. M. Nakai and R. A. Hess, Anat. Rec., 1997, 247, 379 CrossRef CAS PubMed.
  40. L. Duvillaret, F. Garet and J. L. Coutaz, IEEE J. Sel. Top. Quantum Electron., 1996, 2, 739–746 CrossRef CAS.
  41. T. D. Dorney, R. G. Baraniuk and D. M. Mittleman, J. Opt. Soc. Am. A, 2001, 18, 1562–1571 CrossRef CAS PubMed.
  42. H. Yang, L. Chen, Z. Cheng, M. Yang and W. Li, BMC Med., 2021, 19, 80 CrossRef PubMed.
  43. D. K. Kim, B. J. Cho, M. J. Lee and H. K. Ju, Medicine, 2021, 100, e24756 CrossRef PubMed.
  44. S. K. Roy, S. Manna, T. Song and L. Bruzzone, IEEE Trans. Geosci. Rem. Sens., 2020, 1–13 Search PubMed.
  45. M. Arjovsky, S. Chintala and L. Bottou, 2017, arXiv, preprint arXiv:1701.07875.
  46. M. Arjovsky and L. Bottou, 2018, arXiv, preprint arXiv: 1710.10196.
  47. J. Yosinski, J. Clune, Y. Bengio and H. Lipson, Adv. Neural Inf. Process Syst., 2014, 27, 1–14 Search PubMed.
  48. D. Jia, D. Wei, R. Socher, L. J. Li, L. Kai and F. F. Li, Proc of IEEE Computer Vision and Pattern Recognition, 2009, pp. 248–255 Search PubMed.
  49. K. Simonyan and A. Zisserman, 2014, arXiv, preprint arXiv:1409.1556.
  50. J. Redmon, S. Divvala, R. Girshick and A. Farhadi, 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2016, pp. 779–788 Search PubMed.
  51. F. Liu, C. Shen and G. Lin, 2015 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, 2015, pp. 5162–5170 Search PubMed.
  52. Z. Wang, A. C. Bovik, H. R. Sheikh and E. P. Simoncelli, IEEE Trans. Image Process., 2004, 13, 600–612 CrossRef PubMed.
  53. C. Saunders, M. O. Stitson, J. Weston, R. Holloway, L. Bottou, B. Scholkopf and A. Smola, Computer ence, 2002, 1, 1–28 Search PubMed.
  54. T. Denoeux, IEEE Trans. Syst., Man Cybern., 1995, 25, 804–813 CrossRef.
  55. S. Ng, Y. Xing and K. L. Tsui, Appl. Energy, 2014, 118, 114–123 CrossRef.
  56. A. Chandra and Y. Xin, J. Math. Model. Algorithm., 2006, 5, 417–445 CrossRef.
  57. Y. Ge, Y. Lin, Q. He, W. Wang and S. M. Huang, Energy, 2021, 233, 121220 CrossRef.
  58. J. Kennedy and R. Eberhart, Icnn95-international Conference on Neural Networks, 1995 Search PubMed.
  59. X. F. Song, Y. Zhang, D. W. Gong and X. Z. Gao, IEEE Trans. Cybern., 2021, 1–14 Search PubMed.

This journal is © The Royal Society of Chemistry 2022