Markov Chain Monte Carlo-Based L1/L2 Regularization and Its Applications in Low-Dose CT Denoising

Abstract

In this paper, a low-dose CT denoising method based on L 1 / L 2 regularization method of Markov chain Monte Carlo is studied. Firstly, the mathematical model and regularization method of low-dose CT denoising are summarized, and then the theoretical basis of MCMC method and its application in image denoising are introduced. We evaluated the performance of various regularization strategies by comparing the denoising effects of L 1 , L 2 , and L 1 / L 2 regularization terms in MCMC sampling at Gaussian noise levels. The experimental results show that L 1 / L 2 regularization has the best performance in balancing noise removal and image detail retention, significantly superior to single L 1 and L 2 regularization, which proves its effectiveness for low-dose CT denoising.

Share and Cite:

Yu, S. (2025) Markov Chain Monte Carlo-Based L1/L2 Regularization and Its Applications in Low-Dose CT Denoising. Journal of Applied Mathematics and Physics, 13, 419-428. doi: 10.4236/jamp.2025.132021.

1. Introduction

Low-dose CT imaging is designed to reduce the radiation dose to the patient by reducing the intensity of the X-rays to obtain CT images [1]. However, this dose reduction can significantly increase the noise in the image, affecting the image quality and diagnostic value [2]. Low-dose CT denoising technology is designed to remove noise from noisy CT images to improve image quality. The inverse problem of image denoising is generally regarded as an ill-posed problem [3], which is characterized by the non-uniqueness of the solution, instability and extreme sensitivity to the input data. For image denoising, small changes in noise can cause instability in the recovery results. In order to solve the problem of noise reduction, regularization method is introduced to impose appropriate constraints in the process of solving, so as to improve the stability and reliability of the solution. Regularization avoids over-fitting noise by incorporating prior knowledge (such as signal smoothness, sparsity, or other statistical properties) and enhances the model’s ability to recover the real image.

However, the traditional regularization method is prone to the accumulation of errors in the process of iterative solution, which leads to the problem of fuzzy boundary in the image after denoising.

Therefore, this paper proposes a low-dose CT denoising method based on L 1 / L 2 regularization method of Markov chain Monte Carlo. By constructing a posterior probability distribution of low-dose CT image data with noise, regularization prior is incorporated to control the discomfort. Then, the posterior probability distribution is randomly sampled to ensure the image details while effectively denoising. This method effectively balances the effect of noise reduction and detail preservation in theory and shows significant advantages in practical application. Firstly, the mathematical model and regularization method of image denoising are introduced, and the implementation process of MCMC algorithm is described in detail, including initialization and sampling, and how to optimize the performance of the algorithm by adaptive step size.

Through numerical simulation experiment and real CT image experiment, we compare the denoising effect under different regularization strategies, using peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) as evaluation indexes. By comparing the experimental results of L 1 , L 2 and L 1 / L 2 regularization based on Markov chain Monte Carlo, it is observed that L 1 / L 2 regularization has the best denoising effect on low-dose CT, especially in preserving image structure and detail.

2. Regularization Model of Low-Dose CT Image Denoising

In the process of low-dose CT denoising, the main factor to be considered is that while removing noise, the integrity of the image itself must be ensured. u R n is the original image, and the noisy image is b R m . The image denoising problem can generally be expressed as the inverse problem [4] of the following problems:

b=Iu+η (1)

where η is noise and I is the identity matrix. Due to the existence of noise, the inverse problem shows discomfort, mainly reflected in the non-uniqueness and instability of the solution. Regularization is a common method to solve ill-determined problems, which can improve the quality and stability of the solution. The main idea is to add regularization term Reg( u ) to the prior distribution, and after adding regularization term, the denoised CT image is obtained by solving the following minimization problem:

u * =arg min u { 1 2 Aub 2 2 +αReg( u ) } (2)

where Aub 2 2 is the fidelity term, Reg( u ) is the regularization term, and α is the regularization coefficient, used to balance the fidelity term and the regularization term.

L 1 regularization [5] and L 2 regularization on a given gradient:

u 1 = x u 1 + y u 1

u 2 = x u 2 2 + y u 2 2

L 1 / L 2 regularization term:

u 1 u 2

where is the gradient operator, let D x u, D y u R n be the first-order forward difference operators of the horizontal and vertical directions of the image respectively, then D=[ D x u, D y u ]= .

3. Solution of Low-Dose CT Denoising Problem and Sampling Method

3.1. Estimation of the Posterior Density Function

The prior density function corresponding to the regularization term is

π pr ( u ) e αReg( u ) ,

suppose that the η is a Gaussian random vector with a mean of 0, and the covariance matrix Γ which is positive definite, i.e.

η~N( 0,Γ ) .

The likelihood density function is

π( b|u )exp{ 1 2 ( Aub ) T Γ 1 ( Aub ) }. (3)

The posterior probability density of u obtained by Bayes’ formula is

π( u|b )exp{ 1 2 ( Aub ) T Γ 1 ( Aub )αReg( u ) }. (4)

Assuming that the noise covariance matrix is Γ= β 2 I , the standard form of the posterior probability density function of u is

π( u|b )exp{ 1 2 β 2 Aub l 2 2 αReg( u ) }, (5)

CM method is used to estimate the posterior density function, and the calculation formula of CM method is

u CM = n uπ ( u|b )du. (6)

3.2. MCMC Sampling Method

The MCMC algorithm is used to sample the posterior density function [6], and sample sequence u ( 1 ) , u ( 2 ) ,, u ( M ) is obtained. Suppose that the number of samples in the pre-burning period of the probe posterior probability density function is m 0 , when the number of samples sampled is large enough, the remaining number after removing the samples in the pre-burning period is M m 0 , and the above integral can be approximated as the average of M m 0 samples, that is

u CM = n uπ( u|b )du 1 M m 0 m= m 0 +1 M u ( m ) . (7)

The MCMC sampling algorithm is used to sample the posterior density function, and a proposed distribution Q( u |u )=N( u, σ 2 I ) is obtained, where u is the current state, u is the proposed new state, σ 2 is the variance, I is the identity matrix, new state u =u+δ×ε , ε~N( 0,I ) , the proposed distribution is symmetric, i.e. Q( u |u )=Q( u| u ) , since the jump probabilities from u to u and from u to u are the same. The adaptive step size adjustment is used to make the acceptance rate close to 0.234 [7]. Every 10,000 sampling times, the current acceptance rate is checked and the step size is adjusted accordingly. The rule of step size adjustment can be expressed as

stepsize new ={ stepsize×1.1, if acceptance rate>0.234 stepsize÷1.1, otherwise (8)

in practical applications, 1.1 is a common adjustment factor that provides a reasonable gradual adjustment amplitude (about 10% step change), without causing too drastic a change in sampling step size, and can quickly adapt to the characteristics of the state space, in the experiment, the adjustment factor 1.1 keeps the sampling acceptance rate in the range of 20% - 30%, which is close to the theoretical optimal value of 23.4%.

The acceptance rate is calculated as

α=min( 1, Q( u |b )Q( u| u ) Q( u|b )Q( u |u ) )=min( 1, Q( u |b ) Q( u|b ) ). (9)

Since the sampling method usually deals with the probability density in logarithmic form, in order to simplify the calculation, the posterior probability is logarithmic and set as

Q( u|b )=logπ( b|u )+logπ( u )+C, (10)

where C is a normalized constant and is independent of u , so it can be ignored.

The adaptive step size can be dynamically adjusted according to the acceptance rate in the sampling process. When the acceptance rate is higher than the target value, increasing the step size can accelerate the speed of exploring the state space, and thus reach the equilibrium state faster. On the contrary, if the acceptance rate is too low, reducing the step size can improve the acceptance rate, ensure that the algorithm effectively explores the state space, and enable the algorithm to adjust to the scale suitable for the current target distribution more quickly, so as to accelerate the smooth distribution.

We summarize the MCMC:

Step 1: Set the initial value u ( 1 ) =b=[ b 1 , b 2 ,, b n ] , m=1 , the sample of the combustion period is m 0 , the total number of samples is M, and the initial step size is δ .

Step 2: Update candidate samples u= u ( m ) +δε , ε~N( 0,I ) .

Step 3: The logarithmic posterior probability distribution after the regularization term is

Q( u|b )= 1 2 β 2 Aub 2 Reg( u )

Calculate the receiving probability α as

α=min( 1, Q( u|b ) Q( u ( m ) |b ) )=min( 1,exp( Q( u|b )Q( u ( m ) |b ) ) ) .

Step 4: If αt , t~U( 0,1 ) , then u ( m+1 ) =u , otherwise reject candidate sample order u ( m+1 ) = u ( m ) .

Step 5: When m = M, sampling stops. Otherwise, continue sampling and make m = m + 1. Repeat the second and third steps.

Step 6: After every 10,000 samples, adjust Step 1 according to the current acceptance rate δ

δ={ δ×1.1, α>0.234 δ÷1.1, otherwise

go to Step 2 update δ .

Step 7: Based on the sampling results, the denoised image estimated by conditional mean (CM) is calculated

u CM 1 M m 0 m= m 0 +1 M u ( m ) .

In order to compare the denoising effects of L 1 , L 2 regularization terms and L 1 / L 2 regularization terms, and compare the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) [8] of the denoised images, the higher the PSNR value is, the smaller the error between the denoised image and the original image, and the better the denoised effect is. The closer the SSIM value is to 1, the higher the similarity between the denoised image and the original image. The calculation of PSNR and SSIM is given below:

PSNR=10 log 10 ( MAX x 2 MSE ) (11)

where MSE= 1 mn i=1 m j=1 n ( x( i,j )y( i,j ) ) 2 , where x is the original image, y is the denoised image, m and n are the size of the image, and MAX I is the maximum possible pixel value of the image.

SSIM= ( 2 μ x μ y + c 1 )( 2 σ xy + c 2 ) ( μ x 2 + μ y 2 + c 1 )( σ x 2 + σ y 2 + c 2 ) (12)

where x represents the original image, y represents the denoised image, μ x and μ y are the mean values of the images x and y, σ x 2 and σ y 2 are the variances of the images x and y, σ xy is the covariance of the images x and y, c 1 = ( k 1 L ) 2 , c 2 = ( k 2 L ) 2 is a small constant to avoid denominators of zero, where L is the pixel value of the dynamic range, and k 1 =0.01 and k 2 =0.03 are the default values.

4. Experiment Results

4.1. Low-Dose Noise Image

In the study of low-dose CT image denoising algorithm, in order to objectively evaluate the effectiveness of the denoising algorithm, people often choose the recognized Sheep-Logan model as the research object.

Noise was added to the projection data of Sheep-Logan head model to mimic low-dose CT images and the noise approximately follows the non-stationary Gaussian distribution with the mean of 0. The noise variance formula [9] is as follows:

σ i 2 = f i ×exp( μ i φ ) (13)

where μ i and σ i 2 are the projected data mean and noise variance obtained on the i-th detector, f i and φ are two configuration parameters, set here f i =0.3 , φ=30 .

Inverse Radon transform is used to invert the projected data after noise, and the noise image is obtained. Figure 1 shows the comparison between the original image and the noise image.

Figure 1. The left image is the original image without noise, and the right image is the noisy image.

4.2. Comparison of the Denoising Effect of L 1 , L 2 and L 1 / L 2 Regularization Terms

Next, the regularization terms Reg( u )= u 1 , Reg( u )= u 2 and Reg( u )= u 1 u 2 are respectively used to carry out MCMC sampling on the projected data containing Gaussian noise, set the sampling to 400,000 times, β=0.05 , α= 10 5 and obtain three denoised images, as shown in Figure 2.

Figure 2. The first one is the denoised image using L 1 regularization, the second one is the denoised image using L 2 regularization, and the third one is the denoised image using L 1 / L 2 regularization.

Table 1. Comparison of PSNR and SSIM for noisy images and denoised images using L 1 , L 2 , and L 1 / L 2 regularization.

PSNR

SSIM

Noisy

23.98

0.7124

L 1 regularization

24.02

0.8886

L 2 regularization

22.54

0.8059

L 1 / L 2 regularization

25.39

0.9130

Figure 3. Local comparison results, select the local area in the red box in Figure 2 (50 - 79 columns, 80 - 109 rows, a total of 30 × 30 pixels) for analysis.

The PSNR and SSIM of the three images are shown in Table 1, it can be seen that L 1 / L 2 regularization has the highest PSNR and SSIM values, while L 2 regularization has the lowest PSNR and SSIM values. It can be seen that L 1 / L 2 regularization has the best effect on image denoising and detail retention. In comparison, L 1 regularization can also effectively improve image quality. However, it is inferior to L 1 / L 2 regularization in terms of denoising and detail retention, while L 2 regularization performs the worst in terms of denoising and detail retention, and from a local zooming in image, as shown in Figure 3, L 1 / L 2 is also superior to L 1 and L 2 regularization in terms of detail retention and denoising.

4.3. Explore the Denoising Effect of Real Low-Dose CT Image

In order to further prove the effectiveness of this algorithm in practical application, this experiment selected full dose CT and quarter dose CT images from data set files on the Kaggle platform. Kaggle is a globally renowned data science and machine learning competition platform that provides rich datasets for users to use. In order to study the effectiveness of the algorithm, this paper selected full dose CT and quarter dose chest CT images. The section thickness was 1mm, and the quarter dose CT was low-dose CT. As shown in Figure 4, it can be seen that there was obvious noise in the quarter dose CT images.

Figure 4. The left image is a full dose CT image, and the right image is a quarter dose image.

The algorithm in this paper was used to de-noise low-dose CT, the sampling times were still set to 400,000 times, and the parameters were set to α=0.02 and β= 10 3 to obtain three different kinds of CT images after regularization, as shown in Figure 5.

Figure 5. The first one is the denoised image using L 1 regularization, the second one is the denoised image using L 2 regularization, and the third one is the denoised image using L 1 / L 2 regularization.

According to the experimental results, this is shown in Figure 5 and Table 2, the PSNR of CT images after L 1 , L 2 and L 1 / L 2 regularization has been improved, among which L 1 / L 2 regularization has been improved the most, and SSIM has also been improved in addition to L 2 regularization decrease, which is also the most obvious improvement in L 1 / L 2 regularization. In addition, by observing the local amplification area, as shown in Figure 6, L 1 / L 2 regularized and denoised CT images also have the best performance in noise removal and detail retention.

Table 2. Comparison of PSNR and SSIM for noisy images and denoised images using L 1 , L 2 , and L 1 / L 2 regularization.

PSNR

SSIM

Quarter dose

28.21

0.9088

L 1 regularization

29.75

0.9152

L 2 regularization

28.33

0.8907

L 1 / L 2 regularization

30.30

0.9238

Figure 6. Local comparison results, select the local area in the red box in Figure 5 for analysis.

Through the real clinical low-dose CT image denoising analysis, the effectiveness and advantages of the L 1 / L 2 regularization denoising effect based on MCMC sampling algorithm are further proved.

5. Conclusions

In this paper, the low-dose CT denoising algorithm is analyzed and verified, and the regularization model based on MCMC sampling is introduced in this field. By employing MCMC sampling to process noisy image data, the denoising effect of three regularization strategies— L 1 regularization, L 2 regularization, and L 1 / L 2 regularization—was evaluated.

The results demonstrated that L 1 / L 2 regularization outperformed both L 1 and L 2 regularizations in multiple evaluation metrics. Particularly, L 1 / L 2 regularization excelled in balancing noise removal with the preservation of image details, offering superior denoising results. This highlights the significance of using regularization in denoising models to effectively balance image fidelity and noise suppression. In addition, the adaptive step size in MCMC sampling further improves the stability and convergence speed of the algorithm, making the denoising process more efficient and practical. The experimental results show that the proposed method is not only feasible in theory, but also can achieve the expected denoising performance in practical application.

In summary, this study demonstrated the effectiveness and superiority of combining regularization models with MCMC sampling algorithms for image denoising. Future research could focus on optimizing computational efficiency and exploring its applications in broader image-processing tasks.

Conflicts of Interest

The author declares no conflicts of interest regarding the publication of this paper.

References

[1] Mayo, J.R., Hartman, T.E., Lee, K.S., Primack, S.L., Vedal, S. and Müller, N.L. (1995) CT of the Chest: Minimal Tube Current Required for Good Image Quality with the Least Radiation Dose. American Journal of Roentgenology, 164, 603-607.
https://doi.org/10.2214/ajr.164.3.7863879
[2] Haque, A., Wang, A.S. and Imran, A. (2022) Noise2Quality: Non-Reference, Pixel-Wise Assessment of Low Dose CT Image Quality. Medical Imaging 2022: Image Perception, Observer Performance, and Technology Assessment, San Diego, 20-24 February 2022, 120351C.
https://doi.org/10.1117/12.2611254
[3] Hadamard, J. (1923) Lectures on Cauchy’s Problem in Linear Partial Differential Equations. Yale University Press.
[4] Bertero, M. and Boccacci, P. (1998) Introduction to Inverse Problems in Imaging. IOP Publishing Ltd.
https://doi.org/10.1887/0750304359
[5] Liu, J., Huang, J. and Zhang, S. (2012) Limited-Angle CT Reconstruction via the Minimization. Inverse Problems and Imaging, 6, 357-375.
[6] Metropolis, N., Rosenbluth, A.W., Rosenbluth, M.N., Teller, A.H. and Teller, E. (1953) Equation of State Calculations by Fast Computing Machines. The Journal of Chemical Physics, 21, 1087-1092.
https://doi.org/10.1063/1.1699114
[7] Roberts, G.O., Gelman, A. and Gilks, W.R. (1997) Optimal Scaling for Various Metropolis-Hastings Algorithms. Statistical Science, 7, 473-483.
[8] Wang, Z., Bovik, A.C., Sheikh, H.R. and Simoncelli, E.P. (2004) Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Transactions on Image Processing, 13, 600-612.
https://doi.org/10.1109/tip.2003.819861
[9] Li, T., Li, X., Wang, J., et al. (2004) Nonlinear Sinogram Smoothing for Low-Dose X-Ray CT. IEEE Transactions on Nuclear Science, 51, 2505-2513.
https://doi.org/10.1109/tns.2004.834824

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.