Machine Learning for Tomographic Imaging. Professor Ge Wang. Читать онлайн. Newlib. NEWLIB.NET

Автор: Professor Ge Wang
Издательство: Ingram
Серия:
Жанр произведения: Медицина
Год издания: 0
isbn: 9780750322164
Скачать книгу
basis set: a strategy employed by V1? Vis. Res. 37 3311–25

      IOP Publishing

      Machine Learning for Tomographic Imaging

      Ge Wang, Yi Zhang, Xiaojing Ye and Xuanqin Mou

      Chapter 2

      Tomographic reconstruction based on a learned dictionary

      2.1 Prior information guided reconstruction

      In the first chapter, we have discussed the acquisition and representation of prior knowledge on images from two perspectives, which are natural image statistics and neurophysiological HVS functions. In this chapter, we will discuss the computational methods to solve the inverse problem iteratively, aided by prior information.

      Previously, we have presented the approach for solving inverse problems in the Bayesian inference framework, which is equivalent to minimizing the objective function via Lagrangian optimization:

      xˆ=arg minxˆ∈x{ϕ(x,y)+λψ(x)},(2.1)

      where ϕ(x,y) is the data fidelity term to encourage that an estimated CT image x be consistent with observed projection data y; ϕ(x,y) is also a logarithmic function that reflects the noise statistics. As a result, the fidelity term can be characterized by some norm to model the type of statistical noise. For instance, if the noise obeys a Gaussian distribution, we use the L2-norm such as in the form of 12∥Ax−y∥22. In the case of Poisson noise, an informational measure is proper in the form of ∫(Ax−ylnAy). If the inverse problem is subject to impulsive noise, such as salt-and-pepper noise, the L1-norm can be used, which is expressed as ∥Ax−y∥1. On the other hand, the regularization term ψ(x) promotes solutions with some desirable properties. As explained in the first chapter, this term reflects the characteristic of natural images, which can be obtained by removing the redundancy of images. As mentioned before, by using either principal component analysis (PCA) whitening basis functions or zero-phase component analysis (ZCA) whitening basis functions prior to independent component analysis (ICA), the first- and second-order redundancies of images can be removed and the whitened features are uncorrelated. Moreover, a number of excellent basis functions can be learned using a sparse coding technique to capture statistically independent features. To go a step further along this direction, with the use of a multi-layer neural network, we can extract and represent structural information or semantics, which could promote the effectivity and efficiency of the solution to inverse problems. This neural network perspective will be focused on in this book.

      It is underlined that Bayesian inference is a classic approach to solve inverse problems. In the Bayesian framework, an image prior is introduced to constrain the solution space for suppression of measurement noise and image artifacts. This strategy generally requires an iterative algorithm. Indeed, there are many optimization methods to minimize a regularized objective function, the optimal result is almost always iteratively obtained for a balance between the data fidelity and the regularization term. In other words, the fidelity term is not necessarily equal to zero. It is so because the observation y contains both the ideal signal we want to obtain and noise/error that cannot be avoided in practice. Then, a regularizer or prior knowledge can be used to guide the search path for an optimal solution xˆ as a trade-off between imperfect measurement and desirable image properties.

      In the following, let us intuitively explain regularized image reconstruction. Without loss of generality, let us consider the objective function with the fidelity term in the L2-norm:

      It is assumed that the observation y is related to the unknowns of interest x through a degradation operator A and, at the same time, y is corrupted by non-ideal factors in the data acquisition process, such as Gaussian noise, and is modeled as

      Indeed, this imaging model is totally different from conventional image processing tasks which are from images to images or from images to features. The quintessence of image processing is how to discriminate the signal from its noisy or some error contaminated background. It is often not an easy task in the spatial domain because an image appearing as a collection of pixel values does not provide the signal and the noise/error separately. Fortunately, we can transform the image to a feature space in which the image signal and noise/measurement error can be much more easily discriminated. The workflow involved consists of the following three steps. First, an original image is transformed from the spatial domain into a feature space, where a specific aspect of physical properties of the image can be well presented. It should be noted that the transform is invertible and conservative, which means that transformational loss is zero. The Fourier transform and wavelet transform are good examples. With such a lossless transform, structures, errors, and noise in the image are all preserved but exist in a different form. In the second step, according to statistical rules in the transformed feature space, noise and errors can be suppressed by modifying features so that they satisfy the underlying statistical laws; for example, by means of either soft thresholding in the wavelet domain or frequency filtering in the Fourier domain. The former makes refinements in reference to the prior information in the wavelet domain that the structural components should have a sparse distribution of significant coefficients, while the image noise has a broad and weak spectrum of amplitudes in the wavelet domain. The latter is based on the fact that the frequency components of the image concentrate to a low frequency band while the noise spreads over the whole Fourier space. Finally, in the third step the output is obtained by the corresponding inverse transform of the modified features.

      Different to the transform method, the key ingredient of the regularization method is to estimate an underlying image by leveraging prior knowledge on desirable features/structures of images, while eliminating the noise and error. The use of a regularizer constrains the image model. In this way, it is convenient and effective to promote favorable properties of images such that the learned model will represent the image characteristics we want, such as sparseness, low rank, smoothness, and so on. Mathematically, the regularizer can be expressed as a norm that measures the image x in a way that is optimal for the inverse problem of interest.

      It is worth mentioning that both the above strategies utilize natural image statistics but in different ways. To be specific, the Bayesian framework uses prior knowledge as the constraint in the objective function, while the transform approach uses the statistical distributions of signals and noise/errors in the feature space, which can be also regarded as prior knowledge.

      We have discussed the regularization term in the previous chapter. In the following, we will give an intuitive example to show the impact of the regularizer on the efficiency of the solution to the inverse problem. More specifically, the L1-norm-based regularized solution will be compared with that based on the L2-norm.

      We will focus on the regularization issue in the context of the machine learning-based inverse problem solution in this book. Mathematically, the L2-norm-based term is called Tikhonov regularization, or ridge regression, which is a commonly used method. The L1-norm-based term is called Lasso regression, which is essentially a sparsity constraint. In the previous chapter, we have shown that the sparse constraint favors an efficient information representation. Next, let us elaborate the effects of the two regularization terms, respectively.

      Based on equations (2.2)