research papers\(\def\hfill{\hskip 5em}\def\hfil{\hskip 3em}\def\eqno#1{\hfil {#1}}\)

Journal logoJOURNAL OF
SYNCHROTRON
RADIATION
ISSN: 1600-5775

Deep-learning-based ring artifact correction for tomographic reconstruction

crossmark logo

aBeijing Synchrotron Radiation Facility, X-ray Optics and Technology Laboratory, Institute of High Energy Physics, Chinese Academy of Sciences, Yuquan Road, Shijingshan District, Beijing 010000, People's Republic of China, bUniversity of Chinese Academy of Sciences, Yuquan Road, Shijingshan District, Beijing 010000, People's Republic of China, and cCAS Key Laboratory for Biomedical Effects of Nanomedicines and Nanosafety and CAS Center for Excellence in Nanoscience, National Center for Nanoscience and Technology of China, Beijing 100190, People's Republic of China
*Correspondence e-mail: zhangk@ihep.ac.cn, yuanqx@ihep.ac.cn

Edited by A. Momose, Tohoku University, Japan (Received 9 August 2022; accepted 1 February 2023; online 10 March 2023)

X-ray tomography has been widely used in various research fields thanks to its capability of observing 3D structures with high resolution non-destructively. However, due to the nonlinearity and inconsistency of detector pixels, ring artifacts usually appear in tomographic reconstruction, which may compromise image quality and cause nonuniform bias. This study proposes a new ring artifact correction method based on the residual neural network (ResNet) for X-ray tomography. The artifact correction network uses complementary information of each wavelet coefficient and a residual mechanism of the residual block to obtain high-precision artifacts through low operation costs. In addition, a regularization term is used to accurately extract stripe artifacts in sinograms, so that the network can better preserve image details while accurately separating artifacts. When applied to simulation and experimental data, the proposed method shows a good suppression of ring artifacts. To solve the problem of insufficient training data, ResNet is trained through the transfer learning strategy, which brings advantages of robustness, versatility and low computing cost.

1. Introduction

X-ray tomography technology has the advantages of strong penetration ability, high imaging resolution and rich contrast source (Kalender, 2006[Kalender, W. A. (2006). Phys. Med. Biol. 51, R29-R43.]; Kareh et al., 2014[Kareh, K. M., Lee, P. D., Atwood, R. C., Connolley, T. & Gourlay, C. M. (2014). Nat. Commun. 5, 4464.]; Pfeiffer, 2018[Pfeiffer, F. (2018). Nat. Photon. 12, 9-17.]; Sakdinawat & Attwood, 2010[Sakdinawat, A. & Attwood, D. (2010). Nat. Photon. 4, 840-848.]), and is widely used in the fields of medicine, biology, material science and chemistry due to these excellent properties (Fu et al., 2022[Fu, T. Y., Monaco, F., Li, J. Z., Zhang, K., Yuan, Q. X., Cloetens, P., Pianetta, P. & Liu, Y. J. (2022). Adv. Funct. Mater. 32, 2270218.]; Jiang et al., 2020[Jiang, Z. S., Li, J. Z., Yang, Y., Mu, L. Q., Wei, C. X., Yu, X. Q., Pianetta, P., Zhao, K. J., Cloetens, P., Lin, F. & Liu, Y. J. (2020). Nat. Commun. 11, 2310.]; Kareh et al., 2014[Kareh, K. M., Lee, P. D., Atwood, R. C., Connolley, T. & Gourlay, C. M. (2014). Nat. Commun. 5, 4464.]; Lee et al., 2021[Lee, H. R., Liao, L., Xiao, W., Vailionis, A., Ricco, A. J., White, R., Nishi, Y., Chiu, W., Chu, S. & Cui, Y. (2021). Nano Lett. 21, 651-657.]; Li et al., 2022[Li, J. Z., Sharma, N., Jiang, Z. S., Yang, Y., Monaco, F., Xu, Z. R., Hou, D., Ratner, D., Pianetta, P., Cloetens, P., Lin, F., Zhao, K. J. & Liu, Y. J. (2022). Science, 376, 517-521.]). However, because of the nonlinearity and inconsistency of detector pixels, a large number of ring and semi-ring artifacts exist in X-ray tomography, seriously reducing the 3D imaging quality (Paleo & Mirone, 2015[Paleo, P. & Mirone, A. (2015). J. Synchrotron Rad. 22, 1268-1278.]; Croton et al., 2019[Croton, L. C. P., Ruben, G., Morgan, K. S., Paganin, D. M. & Kitchen, M. J. (2019). Opt. Express, 27, 14231-14245.]; Jha et al., 2014[Jha, D., Sørensen, H. O., Dobberschütz, S., Feidenhans'l, R. & Stipp, S. L. S. (2014). Appl. Phys. Lett. 105, 143107.]; Boin & Haibel, 2006[Boin, M. & Haibel, A. (2006). Opt. Express, 14, 12071-12075.]). Therefore, a ring artifact correction method that does not decrease the image resolution is necessary for the reconstruction of reliable high-resolution 3D sample structure. Various methods have been proposed for ring artifact correction. These methods can be divided into two categories. One is based on a specifically designed computed tomography (CT) scanning procedure (Davis & Elliott, 1997[Davis, G. R. & Elliott, J. C. (1997). Nucl. Instrum. Methods Phys. Res. A, 394, 157-162.]; Hubert et al., 2018[Hubert, M., Pacureanu, A., Guilloud, C., Yang, Y., da Silva, J. C., Laurencin, J., Lefebvre-Joud, F. & Cloetens, P. (2018). Appl. Phys. Lett. 112, 203704.]; Pelt & Parkinson, 2018[Pelt, D. M. & Parkinson, D. Y. (2018). Meas. Sci. Technol. 29, 034002.]). For example, continuously changing the relative positions of samples and detectors before each projective acquisition can effectively reduce ring artifacts, but this requires a high-precision positioner and increases acquisition complexity. The other method relies on sinogram stripe removal (Münch et al., 2009[Münch, B., Trtik, P., Marone, F. & Stampanoni, M. (2009). Opt. Express, 17, 8567-8591.]; Vo et al., 2018[Vo, N. T., Atwood, R. C. & Drakopoulos, M. (2018). Opt. Express, 26, 28396-28412.]; Massimi et al., 2018[Massimi, L., Brun, F., Fratini, M., Bukreeva, I. & Cedola, A. (2018). Phys. Med. Biol. 63, 045007.]; Miqueles et al., 2014[Miqueles, E. X., Rinkel, J., O'Dowd, F. & Bermúdez, J. S. V. (2014). J. Synchrotron Rad. 21, 1333-1346.]; Titarenko, 2016[Titarenko, V. (2016). IEEE Signal Process. Lett. 23, 800-804.]; Yan et al., 2016[Yan, L. X., Wu, T., Zhong, S. & Zhang, Q. D. (2016). Phys. Med. Biol. 61, 1278-1292.]). Ring artifacts are generated by stripe artifacts in sinogram images. Therefore, if stripe artifacts can also be removed by image post-processing, then ring artifact correction can be achieved. For example, the Fourier–wavelet (FW) correction method, which is one of the most popular correction algorithms, combines wavelet transform and Fourier filtering to remove stripe artifacts in sinogram images. However, this method needs complex parameter adjustment to adapt to different stripe widths, and compromising between resolution and quality is also inevitable. With the recent rapid development of artificial intelligence technology (Bai et al., 2022[Bai, J. L., Liu, Y. T. & Yang, H. W. (2022). Sensors, 22, 3228.]), the strip noise removal (SNR) network has also been proposed (Guan et al., 2019[Guan, J. T., Lai, R. & Xiong, A. (2019). IEEE Access, 7, 44544-44554.]). Due to the neural network's excellent abilities of feature detection and extraction, this correction method performs ring artifact correction especially well. However, given that the traditional network structure is difficult to achieve a deep network with high accuracy, this method is not quite adapted to strong artifacts.

To overcome the above drawback in existing correction methods, a new ring artifact correction method (RRAC) based on a residual neural network (ResNet) is proposed. This method can use complementary information of wavelet coefficients to remove artifacts while preserving and restoring details of the original image. The main views and contributions of this paper are summarized as follows:

(1) The artifact correction network in this paper is designed based on the residual block, which not only saves operational costs but also improves the accuracy of the network.

(2) The input of the artifact correction network is wavelet coefficients of the sinogram, and its image size is one-quarter of the original image. This input method significantly reduces the operation time and memory consumption. The network output is the artifact in wavelet coefficients. This output mode is not only simpler but also conducive to accurately obtaining the intensity and distribution of artifacts.

(3) To prevent the network from outputting information other than stripe artifacts, this paper also adds a regularization term to the loss function, so that the artifact correction network can more accurately separate the details of the sample from the artifact.

(4) In order to make the RRAC method show high accuracy and strong robustness under limited experimental data, we use the transfer learning strategy to solve the problem of insufficient experimental data.

2. Method

2.1. Deep-learning-based artifact correction

The ring artifact correction method based on the ResNet workflow (Fig. 1[link]) mainly includes two steps: training [Fig. 1[link](a)] and application [Fig. 1[link](b)]. The neural network uses a process analogous to the human brain, which requires training a component with labeled data named the training dataset. The training process is as follows: first, the training set data are decomposed into four coefficients by Haar discrete wavelet transform (HDWT) (Lai & Chang, 2006[Lai, B. L. & Chang, L. W. (2006). 1st Pacific Rim Symposium on Image and Video Technology (PRSIVT 2006), 10-13 December 2006, Hsinchu, Taiwan, pp. 1085-1093.]). Second, the decomposed coefficients are fed into the artifact correction network. The complementary information of different wavelet sub-band coefficients can help the network well preserve and restore the detailed information of an image while eliminating artifacts. Moreover, given that the dimension of the wavelet coefficients is half of the original image, training data pre-processed by HDWT can reduce the required memory space and speed up the training. Finally, the network output is the predicted artifact in a wavelet sub-band coefficient After training, the RRAC method can apply artifact correction to other data without a given ground truth. The application process [Fig. 1[link](b)] is similar to the training process. First, four wavelet coefficients are generated by the wavelet transform of the original sinogram image. Second, according to these coefficients, the predicted artifact can be produced by the network and, when subtracted from the input wavelet coefficients, clean wavelet coefficients can be obtained. Finally, high-quality reconstruction results without ring artifacts can be generated by inverse Haar discrete wavelet transform (IHDWT) and filtered back projection (FBP) reconstruction (Guersoy et al., 2014[Gürsoy, D., De Carlo, F., Xiao, X. & Jacobsen, C. (2014). J. Synchrotron Rad. 21, 1188-1193.]; Pelt et al., 2016[Pelt, D. M., Gürsoy, D., Palenstijn, W. J., Sijbers, J., De Carlo, F. & Batenburg, K. J. (2016). J. Synchrotron Rad. 23, 842-849.]).

[Figure 1]
Figure 1
Flow chart of the ring artifact correction method based on ResNet. (a) Training process. (b) Application process.

2.2. Design of the artifact correction neural network

The core of the RRAC method is the artifact correction network (Fig. 2[link]), which is designed on the basis of ResNet (He et al., 2016[He, K. M., Zhang, X. Y., Ren, S. Q., Sun, J. & IEEE (2016). 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), 27-30 June 2016, Las Vegas, NV, USA, pp. 770-778.]). The artifact correction network comprises 14 convolution layers, including two 3 × 3 convolution layers and six residual blocks. Each residual block includes two 3 × 3 convolution layers and one shortcut connection. In the residual block, input data are corrected by the residuals obtained from the two convolution layers. This structural design of the residual block has the following three advantages. First, it can speed up the network training process and save memory space. Second, the residual block can eliminate the problem of vanishing/exploding gradients and consequently be conducive to the establishment of a high-performance deep network (Balduzzi et al., 2017[Balduzzi, D., Frean, M., Leary, L., Lewis, J. P., Ma, K. W. D. & McWilliams, B. (2017). 34th International Conference on Machine Learning (ICML 2017), 6-11 August 2017, Sydney, Australia, pp. 536-549.]; Sandler et al., 2018[Sandler, M., Howard, A., Zhu, M. L., Zhmoginov, A. & Chen, L. C. (2018). 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2018), 18-23 June 2018, Salt Lake City, UT, USA, pp. 4510-4520.]). Third, the residual is facilitative to image detail preservation. Furthermore, except that the last convolution layer is four channels corresponding to four input wavelet coefficients, the other convolution kernels are 64 channels. Different network layers are connected by the ReLU activation function (Xu et al., 2015[Xu, B., Wang, N., Chen, T. & Li, M. (2015). arXiv:1505.00853.]), which brings nonlinear mapping to the network and enables it to deal with nonlinear problems. Moreover, the reason why the size of the image is not changed in the correction network is that the image details will inevitably be lost in the down-sampling process. The purpose of the correction network is to output stripe artifacts in the wavelet coefficients. Most stripe artifacts are single-pixel or several-pixel stripes. The down-sampling will lose some small artifact, which makes it difficult for the artifact correction network to work on some small artifacts.

[Figure 2]
Figure 2
Network structure of the artifact correction network.

For the artifact correction network, the loss function can evaluate network output and guide the update of network parameters during the training process, so it plays an important guiding role to the RRAC method. The loss function comprises two terms: wavelet loss function LM (Chen et al., 2018[Chen, H. G., He, X. H., Qing, L. B., Xiong, S. H., Nguyen, T. Q. & IEEE (2018). 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2018), 18-23 June 2018, Salt Lake City, UT, USA, pp. 824-833.]; Huang et al., 2017[Huang, H. B., He, R., Sun, Z. N., Tan, T. N. & IEEE (2017). 16th IEEE International Conference on Computer Vision (ICCV 2017), 22-29 October 2017, Venice, Italy, pp. 1698-1706.]) and regular loss function LW.

The wavelet loss function is the mean square error (MSE) (Ledig et al., 2017[Ledig, C., Theis, L., Huszar, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang, Z. H., Shi, W. Z. & IEEE (2017). 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2017), 22-25 July 2017, Honolulu, HI, USA, pp. 105-114.]), which is one of the common loss functions and can evaluate well the error between the network output and the ground truth. The MSE is defined as follows,

[{\rm{MSE}} \ =\ {{{1}}\over{{N}}} \sum\limits_{i\,=\,1}^{N} \,\left(P_i-P_i^{\,\prime}\,\right)^{2}, \eqno(1)]

where N is the total number of pixels. Pi is the ith pixel value of the ground truth, and [P_i^{\,\prime}] is the ith pixel value of the network output. LM can be formulated as follows,

[{{L}}_{\rm{M}} = {\rm{MSE}}_{\rm{WA}} + {\rm{MSE}}_{\rm{WH}} + {\rm{MSE}}_{\rm{WV}} + {\rm{MSE}}_{\rm{WD}}, \eqno(2)]

where MSEWA, MSEWH, MSEWV and MSEWD are the MSE values for the approximation wavelet coefficient (WA), horizontal wavelet coefficient (WH), vertical wavelet coefficient (WV) and diagonal wavelet coefficient (WD), respectively.

The stripe's gray value varies less along the stripe direction than in the perpendicular direction (Chen et al., 2017[Chen, Y., Huang, T. Z., Deng, L. J., Zhao, X. L. & Wang, M. (2017). Neurocomputing, 267, 95-106.]; Liu et al., 2016[Liu, X. X., Lu, X. L., Shen, H. F., Yuan, Q. Q., Jiao, Y. L. & Zhang, L. P. (2016). IEEE Trans. Geosci. Remote Sensing, 54, 3049-3060.]). Therefore, the regular loss function LW adapts smoothness in the fringe direction to estimate the fringe noise and is defined as follows,

[{L}_{\rm{W}} = \big\|\nabla{S}_{\rm{WA}}\big\|^{2} + \big\|\nabla{S}_{\rm{WH}}\big\|^{2}, \eqno(3)]

where ∇ denotes the partial differential operator along the stripe direction. SWA is the stripe component of WA, and SWH is the stripe component of WH.

In this study, the loss function of the network is

[{\rm{Loss}} = {L}_{\rm{M}}+\lambda{L}_{\rm{W}}, \eqno(4)]

where λ is the regularization coefficient, which is used to balance the relationship between the two loss functions and prevent the network from under- or over-fitting.

3. Experiments and discussion

3.1. Evaluation of the RRAC method by synthetic data

The RRAC method is compiled in a Python environment, and the neural network is built on the PyTorch framework (Paszke et al., 2019[Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z. M., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J. J. & Chintala, S. (2019). 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), 8-14 December 2019, Vancouver, Canada.]). All the tests are carried out on a workstation with a CPU of a 2.2 GHz Intel Xeon silver 4114 and a NVIDIA Quadro p6000 graphics processing unit.

The accuracy of the RRAC method is evaluated by synthetic data. Due to the lack of sufficient public CT datasets to train and test the network, the Div2k dataset is selected as the synthetic data, which is composed of many high-resolution 2D images (Agustsson & Timofte, 2017[Agustsson, E. & Timofte, R. (2017). 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2017), 22-25 July 2017, Honolulu, HI, USA, pp. 1122-1131.]). The Div2k dataset is pre-processed as follows to establish training and test sets: first, 824 pre-processed slice images P1 are randomly picked up from the Div2k dataset, converted to grayscale images, and clipped. Second, images P1 (512 × 512) are Radon-transformed to obtain the ground truth sinogram image S1. (A total of 360 projective images were recorded over an angular range of 0° to 179.5°.) Finally, sinogram S1 is added with random stripe artifacts to acquire artifact sinogram image S2. In total, the training set includes 712 randomly selected groups of ground truth S1 and artifact sinogram images S2. The remaining 112 image groups are adopted as a test set to evaluate network accuracy. The training and test sets established as above not only solve the problem of insufficient data but also have better visibility of the artifact than CT data, which facilitate the subsequent evaluation. Moreover, compared with CT data, the better variety of the Div2k dataset brings the network more versatility. The degree of artifacts added manually is evaluated by the peak signal-to-noise ratio (PSNR). A high PSNR value means few artifacts in the image.

The effect of complementary information of wavelet coefficients on artifact correction is evaluated through ablation experiments. The correction network is trained by approximation and horizontal wavelet coefficients, which is called the related artifact correction model, whereas training through all wavelet coefficients is called the complete artifact correction model. The correction results of these two models are shown in Fig. 3[link]. By comparing reconstructed results it can be seen that, although the related correction model can remove the ring artifacts, it has a poor ability to restore and preserve the original details of the image. The result using the complete correction model is almost the same as the ground truth. The above results prove that the complete correction model can effectively preserve and restore detailed information while removing artifacts through complementary information of each wavelet coefficient.

[Figure 3]
Figure 3
Reconstructed slice with ground truth (a), without correction (b), with related correction model (c), with complete correction model (d). (eh) Magnified views of the selected regions.

After training, the network can be applied to the test data to access its performance. The raw sinogram, as illustrated in Fig. 4[link](b), is obtained by adding the ground truth sinogram [Fig. 4[link](a)] with random stripe artifacts. Its stripe correction results using the FW, SNR and RRAC methods are displayed in Figs. 4[link](c), 4[link](d) and 3(e), respectively. Figs. 4[link](f)–4(j) show enlargements of the outlined areas in Figs. 4[link](a)–4(e). Compared with the unprocessed sinogram, an obvious reduction of stripes can be observed from sinograms processed by all three methods. However, the RRAC method has achieved the lowest level of residual stripes. Its sinogram is also fairly identical to that of the ground truth. The SNR method cannot remove some strong artifacts with a certain width, and the FW method is even more inferior. The corresponding slices reconstructed by the FBP algorithm are presented in Figs. 4[link](k)–4(o). Ring artifacts in slices are sensitive to the stripes in sinograms, even the invisible residual stripes after FW correction can produce strong rings in slices. The RRAC method can obtain the minimum and almost invisible ring artifacts and preserve the most image details of the slices.

[Figure 4]
Figure 4
Correction results of different artifact removal methods for Div2k simulation data. Sinogram images (a) with ground truth, (b) without correction (PSNR: 13.51), (c) with the FW method, (d) with the SNR method, and (e) with the RRAC method. (fj) Magnified sinogram images corresponding to the red squares shown in (a)–(e). (ko) Sinogram images corresponding to slice images. (pt) Magnified views of the selected regions shown in (k)–(o).

The evaluation is performed quantitatively to further explore the RRAC method. The PSNR and structural similarity coefficient (SSIM,0∼1) are chosen as evaluation criteria (Wang et al., 2004[Wang, Z., Bovik, A. C., Sheikh, H. R. & Simoncelli, E. P. (2004). IEEE Trans. Image Process. 13, 600-612.]). A high PSNR value means few artifacts in an image and high accuracy of the corresponding correction method, whereas a large SSIM value implies a good capability of preserving and restoring details during the correction process. The PSNR and SSIM values of different correction methods are calculated and shown in Fig. 5[link]. Although the accuracy of the FW method for artifacts of different degrees is relatively stable, this method not only has low accuracy but also corrupts the details during the correction process. Although the SNR method shows better accuracy than the FW method, its accuracy is more vulnerable to artifact strength, so the robustness and versatility of the SNR method are inferior. Compared with previous methods, the RRAC method can obtain a more precise correction and more stable performance for different degrees of artifacts.

[Figure 5]
Figure 5
PSNR/SSIM results of various correction methods.

3.2. Application of the RRAC method by experimental data

The RRAC method has shown its advantages of high precision and performance regarding synthetic data. However, due to the significant difference between synthetic and experimental data, the network trained by simulated data is inapplicable to actual application directly. The network also cannot be well trained through limited experimental datasets. Thus, the transfer learning strategy has been adopted to solve experimental data insufficiency (Tan et al., 2018[Tan, C. Q., Sun, F. C., Kong, T., Zhang, W. C., Yang, C. & Liu, C. F. (2018). 27th International Conference on Artificial Neural Networks (ICANN 2018), 5-7 October 2018, Rhodes, Greece, pp. 270-279.]; Zhang & Gao, 2019[Zhang, L. & Gao, X. (2019). arXiv:1903.04687.]). This strategy is as follows: the network is initialized with the parameters trained by the large synthetic dataset and is then further trained with a small experimental dataset. This training strategy not only achieves excellent results under training data shortage but also greatly reduces the training difficulty. CT data of shale collected on the 4W1A station of Beijing Synchrotron Radiation Facility (Yuan et al., 2012[Yuan, Q., Zhang, K., Hong, Y., Huang, W., Gao, K., Wang, Z., Zhu, P., Gelb, J., Tkachuk, A., Hornberger, B., Feser, M., Yun, W. & Wu, Z. (2012). J. Synchrotron Rad. 19, 1021-1028.]) are adopted to further explore the feasibility of the proposed method as a real-world application. This projection (512 × 512) is acquired from 361 angles over 180° at 8 keV by a micrometre-resolution X-ray microscope with an effective pixel size of 2.5 µm.

Fig. 6[link](a) shows the unprocessed raw sinogram image. Figs. 6[link](b), 6(c) and 6(d) are sinogram images corrected by the FW, SNR and proposed methods, respectively. Figs. 6[link](e)–6(h) show enlargements of the outlined areas in Figs. 6[link](a)–6(d). Reconstructed slice images [Figs. 6[link](i)–6(l)] are obtained from the FBP reconstruction. Figs. 6[link](m)–6(p) show enlargements of the outlined areas in Figs. 6[link](i)–6(l). Residual stripes can be found in the sinogram corrected by the FW method [Fig. 6[link](b)]. Subsequently, the image details of reconstructed slices [Fig. 6[link](j)] are corrupted by ring artifacts. In Fig. 6[link](c), most artifacts are removed using the SNR method, except some artifacts of a certain width, resulting in wide ring artifacts in some slices [Fig. 6[link](k)]. Referring to the RRAC method, almost no artifacts can be observed in the sinograms [Fig. 6[link](d)] and slices [Fig. 6[link](l)]. Furthermore, the RRAC method can preserve the fine structure of the shale when removing the stripes, so the resolution of the reconstructed slice images [Fig. 6[link](l)] is also greatly improved. In summary, the above experiment exhibits that after incorporating the transfer learning strategy the RRAC method also shows the remarkable capability of stripe artifact removal in the application of experimental data and outperforms the traditional method.

[Figure 6]
Figure 6
Correction results of different artifact removal methods for shale data. Sinogram images (a) without correction, (b) with FW correction, (c) with SNR correction, (d) with deep learning-based artifact correction. (eh) Magnified sinogram images corresponding to the red squares shown in (a)–(d). (il) Sinogram images corresponding to slice images. (mp) Magnified views of the selected regions shown in (i)–(l).

4. Conclusion

This study introduces a ring artifact correction method named RRAC, which is based on ResNet. Compared with the SNR method, ResNet is introduced, which is thought to be efficient in solving complex problems with deep networks. HDWT is also incorporated to decompose the sinogram into complementary coefficients before being fed into the network. In the synthetic data experiment, the combination of residual network and HDWT exhibits better capability of removing ring artifacts than the reference methods while fully preserving image details. Given that a large training dataset comprising experimental data is unavailable, the transfer learning strategy enables network training with limited experimental data. To reasonably evaluate its accuracy and performance, the RRAC method is evaluated by synthetic and experimental data. When applied to synthetic data, the RRAC method prevails on accuracy and detail preservation in visual and quantitative comparisons with other methods. Moreover, through the experiment on real CT data, the transfer learning strategy succeeds in maintaining the superiority of the RRAC method over the other methods without the availability of abundant experimental data for training. In summary, our proposed method is effective and adaptable to various data types with minimal training data requirements. Its application helps further improve CT 3D reconstruction quality and facilitate subsequent data analysis.

Footnotes

These authors contributed equally to this work.

Acknowledgements

We acknowledge 4W1A beamline of the Beijing Synchrotron Radiation Facility and BL18B beamline of Shanghai Synchrotron Radiation Facility for the experimental data and facilities provided. T. Fu and Y. Wang contributed equally to this work. K. Zhang and Q. Yuan conceived the study. J. Zhang, S. Wang, W. Huang, Y. L. Wang, C. Yao and C. Zhou contributed to the interpretation of the data. T. Fu, Y. Wang and K. Zhang wrote the manuscript with valuable input from all coauthors.

Funding information

Funding for this research was provided by: National Key Research and Development Program of China (award No. 2022YFA1603600, No. 2021YFA1600800); National Natural Science Foundation of China (award No. U2032107).

References

First citationAgustsson, E. & Timofte, R. (2017). 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2017), 22–25 July 2017, Honolulu, HI, USA, pp. 1122–1131.  Google Scholar
First citationBai, J. L., Liu, Y. T. & Yang, H. W. (2022). Sensors, 22, 3228.  Web of Science CrossRef PubMed Google Scholar
First citationBalduzzi, D., Frean, M., Leary, L., Lewis, J. P., Ma, K. W. D. & McWilliams, B. (2017). 34th International Conference on Machine Learning (ICML 2017), 6–11 August 2017, Sydney, Australia, pp. 536–549.  Google Scholar
First citationBoin, M. & Haibel, A. (2006). Opt. Express, 14, 12071–12075.  Web of Science CrossRef PubMed Google Scholar
First citationChen, H. G., He, X. H., Qing, L. B., Xiong, S. H., Nguyen, T. Q. & IEEE (2018). 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2018), 18–23 June 2018, Salt Lake City, UT, USA, pp. 824–833.  Google Scholar
First citationChen, Y., Huang, T. Z., Deng, L. J., Zhao, X. L. & Wang, M. (2017). Neurocomputing, 267, 95–106.  Web of Science CrossRef Google Scholar
First citationCroton, L. C. P., Ruben, G., Morgan, K. S., Paganin, D. M. & Kitchen, M. J. (2019). Opt. Express, 27, 14231–14245.  Web of Science CrossRef CAS PubMed Google Scholar
First citationDavis, G. R. & Elliott, J. C. (1997). Nucl. Instrum. Methods Phys. Res. A, 394, 157–162.  CrossRef CAS Web of Science Google Scholar
First citationFu, T. Y., Monaco, F., Li, J. Z., Zhang, K., Yuan, Q. X., Cloetens, P., Pianetta, P. & Liu, Y. J. (2022). Adv. Funct. Mater. 32, 2270218.  CrossRef Google Scholar
First citationGuan, J. T., Lai, R. & Xiong, A. (2019). IEEE Access, 7, 44544–44554.  Web of Science CrossRef Google Scholar
First citationGürsoy, D., De Carlo, F., Xiao, X. & Jacobsen, C. (2014). J. Synchrotron Rad. 21, 1188–1193.  Web of Science CrossRef IUCr Journals Google Scholar
First citationHe, K. M., Zhang, X. Y., Ren, S. Q., Sun, J. & IEEE (2016). 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), 27–30 June 2016, Las Vegas, NV, USA, pp. 770–778.  Google Scholar
First citationHuang, H. B., He, R., Sun, Z. N., Tan, T. N. & IEEE (2017). 16th IEEE International Conference on Computer Vision (ICCV 2017), 22–29 October 2017, Venice, Italy, pp. 1698–1706.  Google Scholar
First citationHubert, M., Pacureanu, A., Guilloud, C., Yang, Y., da Silva, J. C., Laurencin, J., Lefebvre-Joud, F. & Cloetens, P. (2018). Appl. Phys. Lett. 112, 203704.  Web of Science CrossRef Google Scholar
First citationJha, D., Sørensen, H. O., Dobberschütz, S., Feidenhans'l, R. & Stipp, S. L. S. (2014). Appl. Phys. Lett. 105, 143107.  Web of Science CrossRef Google Scholar
First citationJiang, Z. S., Li, J. Z., Yang, Y., Mu, L. Q., Wei, C. X., Yu, X. Q., Pianetta, P., Zhao, K. J., Cloetens, P., Lin, F. & Liu, Y. J. (2020). Nat. Commun. 11, 2310.  Web of Science CrossRef PubMed Google Scholar
First citationKalender, W. A. (2006). Phys. Med. Biol. 51, R29–R43.  Web of Science CrossRef PubMed Google Scholar
First citationKareh, K. M., Lee, P. D., Atwood, R. C., Connolley, T. & Gourlay, C. M. (2014). Nat. Commun. 5, 4464.  Web of Science CrossRef PubMed Google Scholar
First citationLai, B. L. & Chang, L. W. (2006). 1st Pacific Rim Symposium on Image and Video Technology (PRSIVT 2006), 10–13 December 2006, Hsinchu, Taiwan, pp. 1085–1093.  Google Scholar
First citationLedig, C., Theis, L., Huszar, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang, Z. H., Shi, W. Z. & IEEE (2017). 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2017), 22–25 July 2017, Honolulu, HI, USA, pp. 105–114.  Google Scholar
First citationLee, H. R., Liao, L., Xiao, W., Vailionis, A., Ricco, A. J., White, R., Nishi, Y., Chiu, W., Chu, S. & Cui, Y. (2021). Nano Lett. 21, 651–657.  Web of Science CrossRef CAS PubMed Google Scholar
First citationLi, J. Z., Sharma, N., Jiang, Z. S., Yang, Y., Monaco, F., Xu, Z. R., Hou, D., Ratner, D., Pianetta, P., Cloetens, P., Lin, F., Zhao, K. J. & Liu, Y. J. (2022). Science, 376, 517–521.  Web of Science CrossRef CAS PubMed Google Scholar
First citationLiu, X. X., Lu, X. L., Shen, H. F., Yuan, Q. Q., Jiao, Y. L. & Zhang, L. P. (2016). IEEE Trans. Geosci. Remote Sensing, 54, 3049–3060.  Web of Science CrossRef Google Scholar
First citationMassimi, L., Brun, F., Fratini, M., Bukreeva, I. & Cedola, A. (2018). Phys. Med. Biol. 63, 045007.  Web of Science CrossRef PubMed Google Scholar
First citationMiqueles, E. X., Rinkel, J., O'Dowd, F. & Bermúdez, J. S. V. (2014). J. Synchrotron Rad. 21, 1333–1346.  Web of Science CrossRef CAS IUCr Journals Google Scholar
First citationMünch, B., Trtik, P., Marone, F. & Stampanoni, M. (2009). Opt. Express, 17, 8567–8591.  Web of Science PubMed Google Scholar
First citationPaleo, P. & Mirone, A. (2015). J. Synchrotron Rad. 22, 1268–1278.  Web of Science CrossRef IUCr Journals Google Scholar
First citationPaszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z. M., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J. J. & Chintala, S. (2019). 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), 8–14 December 2019, Vancouver, Canada.  Google Scholar
First citationPelt, D. M., Gürsoy, D., Palenstijn, W. J., Sijbers, J., De Carlo, F. & Batenburg, K. J. (2016). J. Synchrotron Rad. 23, 842–849.  Web of Science CrossRef CAS IUCr Journals Google Scholar
First citationPelt, D. M. & Parkinson, D. Y. (2018). Meas. Sci. Technol. 29, 034002.  Web of Science CrossRef Google Scholar
First citationPfeiffer, F. (2018). Nat. Photon. 12, 9–17.  Web of Science CrossRef CAS Google Scholar
First citationSakdinawat, A. & Attwood, D. (2010). Nat. Photon. 4, 840–848.  Web of Science CrossRef CAS Google Scholar
First citationSandler, M., Howard, A., Zhu, M. L., Zhmoginov, A. & Chen, L. C. (2018). 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2018), 18–23 June 2018, Salt Lake City, UT, USA, pp. 4510–4520.  Google Scholar
First citationTan, C. Q., Sun, F. C., Kong, T., Zhang, W. C., Yang, C. & Liu, C. F. (2018). 27th International Conference on Artificial Neural Networks (ICANN 2018), 5–7 October 2018, Rhodes, Greece, pp. 270–279.  Google Scholar
First citationTitarenko, V. (2016). IEEE Signal Process. Lett. 23, 800–804.  Web of Science CrossRef Google Scholar
First citationVo, N. T., Atwood, R. C. & Drakopoulos, M. (2018). Opt. Express, 26, 28396–28412.  Web of Science CrossRef PubMed Google Scholar
First citationWang, Z., Bovik, A. C., Sheikh, H. R. & Simoncelli, E. P. (2004). IEEE Trans. Image Process. 13, 600–612.  Web of Science CrossRef PubMed Google Scholar
First citationXu, B., Wang, N., Chen, T. & Li, M. (2015). arXiv:1505.00853.  Google Scholar
First citationYan, L. X., Wu, T., Zhong, S. & Zhang, Q. D. (2016). Phys. Med. Biol. 61, 1278–1292.  Web of Science CrossRef CAS PubMed Google Scholar
First citationYuan, Q., Zhang, K., Hong, Y., Huang, W., Gao, K., Wang, Z., Zhu, P., Gelb, J., Tkachuk, A., Hornberger, B., Feser, M., Yun, W. & Wu, Z. (2012). J. Synchrotron Rad. 19, 1021–1028.  Web of Science CrossRef CAS IUCr Journals Google Scholar
First citationZhang, L. & Gao, X. (2019). arXiv:1903.04687.  Google Scholar

This is an open-access article distributed under the terms of the Creative Commons Attribution (CC-BY) Licence, which permits unrestricted use, distribution, and reproduction in any medium, provided the original authors and source are cited.

Journal logoJOURNAL OF
SYNCHROTRON
RADIATION
ISSN: 1600-5775
Follow J. Synchrotron Rad.
Sign up for e-alerts
Follow J. Synchrotron Rad. on Twitter
Follow us on facebook
Sign up for RSS feeds