Cardiac CT (CCT) is of vital importance in heart disease diagnosis but is conventionally limited by its complex workflow that requires dedicated phase and bolus monitoring devices [e.g., electrocardiogram (ECG) gating]. Our previous work has demonstrated the possibility of replacing ECG devices with deep learning (DL)-based monitoring of continuously acquired pulsed mode projections (PMPs, i.e., only a few sparsely sampled projections per gantry rotation). In this work, we report the development of a new projection domain DL-based cardiac phase estimation method that uses ensemble learning [i.e., training multiple convolution neural network (CNN) in parallel] to estimate and reduce DL uncertainty. The estimated DL uncertainty information was then used drive an analytical regularizer in a principled time-dependent manner (i.e., stronger regularization when DL uncertainty is higher). Combined with our previous work on PMP-based bolus curve estimation, the proposed method could potentially be used to achieve autonomous cardiac scanning in a robust (i.e., reduced uncertainty) manner without ECG and bolus timing devices.
Cardiac CT plays an important role in diagnosing heart diseases but is conventionally limited by its complex workflow that requires dedicated phase and bolus tracking [e.g., electrocardiogram (ECG) gating]. This work reports initial progress towards robust and autonomous cardiac CT exams through deep learning (DL) analysis of pulsed-mode projections (PMPs). To this end, cardiac phase and its uncertainty were simultaneously estimated using a novel projection domain cardiac phase estimation network (PhaseNet), which utilizes a sliding-window multi-channel feature extraction approach and a long short-term memory (LSTM) block to extract temporal correlation between time-distributed PMPs. Monte-Carlo dropout layers were utilized to predict the uncertainty of deep learning-based cardiac phase prediction. The performance of the proposed phase estimation pipeline was evaluated using accurate physics-based emulated data.
PhaseNet demonstrated improved phase estimation accuracy compared to more standard methods in terms of RMSE (~43% improvement vs. a standard CNN-LSTM; ~17% improvement vs. a multi-channel residual network [ResNet]), achieving accurate phase estimation with <8% RMSE in cardiac phase (phase ranges from 0-100%). These findings suggest that the cardiac phase can be accurately estimated with the proposed projection domain approach. Combined with our previous work on PMP-based bolus curve estimation, the proposed method could potentially be used to achieve autonomous cardiac CT scanning without ECG device or expert-in-the-loop bolus timing.
Cardiac CT exams are some of the most complex CT exams due to the need to carefully time the scan to capture the heart during a quiescent cardiac phase and when the intravenous contrast bolus is at its peak concentration in the left and/or right heart. We are interested in developing a robust and autonomous cardiac CT exam, using deep learning approaches to extract contrast and cardiac phase timing directly from projections. In this paper, we present a new approach to estimate contrast bolus timing directly from a sparse set of CT projections. We present a deep learning approach to estimate contrast agent concentration in left and right sides of the heart directly from a set of projections. We use a virtual imaging framework to generate training and test data, derived from real patient datasets. We finally combine this with a simple analytical approach to decide on the start of the cardiac CT exam.
Cardiac CT exams are some of the most complex CT exams due to need to carefully time the scan to capture the heart during the quiescent cardiac phase and when the contrast bolus is at its peak concentration. We are interested in developing a robust and autonomous cardiac CT protocol, using deep learning approaches to extract contrast timing and cardiac phase timing directly from pulsed projections. In this paper, we present a new approach to generate large amounts of clinically realistic virtual data for training deep learning networks. We propose a five-dimensional cardiac model generated from 4D cardiac coronary CT angiography (CTA) data for synthetic contrast bolus dynamics and patient ECG profiles. We apply deep learning to segment seven heart compartments and simulate intravenous contrast propagation through each compartment to insert contrast bolus. Additional augmentation techniques by randomizing a bolus curve, patient ECG profile, acquisition timing, and patient motion are applied to increase the amount of data that can be generated. We demonstrate good performance of the deep learning segmentation network, examples of simulated bolus curves using a realistic protocol, and good correspondence between virtually generated projections and real projections from patient scans.
The Zoom-In Partial Scans (ZIPS) method is a recently introduced high-resolution CT technique that utilizes the high geometric magnification in off-center regions in the CT scanner’s field-of-view to boost the intrinsic spatial resolution of existing clinical multi-slice CT. ZIPS performs two off-center partial high-resolution scans of a region of interest (ROI), then an image reconstruction algorithm merges the partial scan data to produce a final high-resolution reconstructed image. In this study, we illustrate the feasibility of ZIPS image reconstruction with simultaneous estimation of inter-scan rigid ROI motion between the two partial scans. A total-variation and an entropy-based image alignment loss function was introduced for registration of the ROI between the two partial scans. Optional denoising filters were also introduced to stabilize the image alignment loss function. The feasibility of the ZIPS reconstruction framework is evaluated in a Catsim simulation environment. Results show that the proposed algorithmic compensation of ROI motion produced images visually indistinguishable from the images reconstructed with the ground truth positions of the ROI. The residual errors of the estimated ROI positions were no greater than 0.12 mm in x- and y- translation and 0.26 degree in rotation. Up to twofold improvement in the modulation transfer function (MTF) was achieved by ZIPS CT relative to a conventional centered scan.
In this study, we implement and compare model based iterative reconstruction (MBIR) with dictionary learning (DL) over MBIR with pairwise pixel-difference regularization, in the context of transportation security. DL is a technique of sparse signal representation using an over complete dictionary which has provided promising results in image processing applications including denoising,1 as well as medical CT reconstruction.2 It has been previously reported that DL produces promising results in terms of noise reduction and preservation of structural details, especially for low dose and few-view CT acquisitions.2
A distinguishing feature of transportation security CT is that scanned baggage may contain items with a wide range of material densities. While medical CT typically scans soft tissues, blood with and without contrast agents, and bones, luggage typically contains more high density materials (i.e. metals and glass), which can produce severe distortions such as metal streaking artifacts. Important factors of security CT are the emphasis on image quality such as resolution, contrast, noise level, and CT number accuracy for target detection. While MBIR has shown exemplary performance in the trade-off of noise reduction and resolution preservation, we demonstrate that DL may further improve this trade-off. In this study, we used the KSVD-based DL3 combined with the MBIR cost-minimization framework and compared results to Filtered Back Projection (FBP) and MBIR with pairwise pixel-difference regularization. We performed a parameter analysis to show the image quality impact of each parameter. We also investigated few-view CT acquisitions where DL can show an additional advantage relative to pairwise pixel difference regularization.
A computationally efficient 2.5D dictionary learning (DL) algorithm is proposed and implemented in the model- based iterative reconstruction (MBIR) framework for low-dose CT reconstruction. MBIR is based on the minimization of a cost function containing data-fitting and regularization terms to control the trade-off between data-fidelity and image noise. Due to the strong denoising performance of DL, it has previously been considered as a regularizer in MBIR, and both 2D and 3D DL implementations are possible. Compared to the 2D case, 3D DL keeps more spatial information and generates images with better quality although it requires more computation. We propose a novel 2.5D DL scheme, which leverages the computational advantage of 2D-DL, while attempting to maintain reconstruction quality similar to 3D-DL. We demonstrate the effectiveness of this new 2.5D DL scheme for MBIR in low-dose CT.
By applying the 2D DL method in three different orthogonal planes and calculating the sparse coefficients accordingly, much of the 3D spatial information can be preserved without incurring the computational penalty of the 3D DL method. For performance evaluation, we use baggage phantoms with different number of projection views. In order to quantitatively compare the performance of different algorithms, we use PSNR, SSIM and region based standard deviation to measure the noise level, and use the edge response to calculate the resolution. Experimental results with full view datasets show that the different DL based algorithms have similar performance and 2.5D DL has the best resolution. Results with sparse view datasets show that 2.5D DL outperforms both 2D and 3D DL in terms of noise reduction. We also compare the computational costs, and 2.5D DL shows strong advantage over 3D DL in both full-view and sparse-view cases.
KEYWORDS: Data modeling, Digital breast tomosynthesis, Breast, Model-based design, Sensors, Optical spheres, 3D modeling, Medical imaging, 3D image processing, Breast imaging
Model-based iterative reconstruction (MBIR) is implemented to process full clinical data sets of dedicated breast tomosynthesis (DBT) in a low dose condition and achieves less spreading of anatomical structure between slices. MBIR is a statistical based reconstruction which can control the trade-off between data fitting and image regularization. In this study, regularization is formulated with anisotropic prior weighting that independently controls the image regularization between in-plane and out-of-plane voxel neighbors. Studies at complete and partial convergence show that the appropriate formulation of data-fit and regularization terms along with anisotropic prior weighting leads to a solution with improved localization of objects within a more narrow range of slices. This result is compared with the solutions using simultaneous iterative reconstruction technique (SIRT), which is one of the state of art reconstruction in DBT. MBIR yields higher contrast-to-noise for medium and large size microcalcifications and diagnostic structures in volumetric breast images and supports opportunity for dose reduction for 3D breast imaging.
KEYWORDS: Signal attenuation, Digital breast tomosynthesis, Sensors, Reconstruction algorithms, Optical spheres, Model-based design, Breast, Data modeling, Computed tomography, Tissues
Model-based iterative reconstruction (MBIR) is an emerging technique for several imaging modalities and appli-
cations including medical CT, security CT, PET, and microscopy. Its success derives from an ability to preserve
image resolution and perceived diagnostic quality under impressively reduced signal level. MBIR typically uses a
cost optimization framework that models system geometry, photon statistics, and prior knowledge of the recon-
structed volume. The challenge of tomosynthetic geometries is that the inverse problem becomes more ill-posed
due to the limited angles, meaning the volumetric image solution is not uniquely determined by the incom-
pletely sampled projection data. Furthermore, low signal level conditions introduce additional challenges due to
noise. A fundamental strength of MBIR for limited-views and limited-angle is that it provides a framework for
constraining the solution consistent with prior knowledge of expected image characteristics. In this study, we
analyze through simulation the capability of MBIR with respect to prior modeling components for limited-views,
limited-angle digital breast tomosynthesis (DBT) under low dose conditions. A comparison to ground truth
phantoms shows that MBIR with regularization achieves a higher level of fidelity and lower level of blurring
and streaking artifacts compared to other state of the art iterative reconstructions, especially for high contrast
objects. The benefit of contrast preservation along with less artifacts may lead to detectability improvement of
microcalcification for more accurate cancer diagnosis.
KEYWORDS: Nonlinear filtering, Image filtering, Digital filtering, Image processing, Image quality, Linear filtering, Denoising, Gaussian filters, Statistical modeling, Signal to noise ratio
Non-linear image processing and reconstruction algorithms that reduced noise while preserving edge detail are currently being evaluated in medical imaging research literature. We have implemented a robust statistics analysis of four widely utilized methods. This work demonstrates consistent trends in filter impact by which such non-linear algorithms can be evaluated. We calculate observer model test statistics and propose metrics based on measured non-Gaussian distributions that can serve as image quality measures analogous to SDNR and detectability. The filter algorithms that vary significantly in their approach to noise reduction include median (MD), bilateral (BL), anisotropic diffusion (AD) and total-variance regularization (TV). It is shown that the detectability of objects limited by Poisson noise is not significantly improved after filtration. There is no benefit to the fraction of correct responses in repeated n-alternate forced choice experiments, for n=2-25. Nonetheless, multi-pixel objects with contrast above the detectability threshold appear visually to benefit from non-linear processing algorithms. In such cases, calculations on highly repeated trials show increased separation of the object-level histogram from the background-level distribution. Increased conspicuity is objectively characterized by robust statistical measures of distribution separation.
Computed Tomography (CT) is widely used for transportation security to screen baggage for potential threats.
For example, many airports use X-ray CT to scan the checked baggage of airline passengers. The resulting
reconstructions are then used for both automated and human detection of threats. Recently, there has been
growing interest in the use of model-based reconstruction techniques for application in CT security systems.
Model-based reconstruction offers a number of potential advantages over more traditional direct reconstruction
such as filtered backprojection (FBP). Perhaps one of the greatest advantages is the potential to reduce reconstruction
artifacts when non-traditional scan geometries are used. For example, FBP tends to produce very
severe streaking artifacts when applied to limited view data, which can adversely affect subsequent processing
such as segmentation and detection.
In this paper, we investigate the use of model-based reconstruction in conjunction with limited-view scanning
architectures, and we illustrate the value of these methods using transportation security examples. The advantage
of limited view architectures is that it has the potential to reduce the cost and complexity of a scanning system,
but its disadvantage is that limited-view data can result in structured artifacts in reconstructed images. Our
method of reconstruction depends on the formulation of both a forward projection model for the system, and a
prior model that accounts for the contents and densities of typical baggage. In order to evaluate our new method,
we use realistic models of baggage with randomly inserted simple simulated objects. Using this approach, we
show that model-based reconstruction can substantially reduce artifacts and improve important metrics of image
quality such as the accuracy of the estimated CT numbers.
Mixed Raster Content (MRC) is a standard for efficient document compression which can dramatically improve
the compression/quality tradeoff as compared to traditional lossy image compression algorithms. The key to
MRC's performance is the separation of the document into foreground and background layers, represented as
a binary mask. Typically, the foreground layer contains text colors, the background layer contains images and
graphics, and the binary mask layer represents fine detail of text fonts.
The resulting quality and compression ratio of a MRC document encoder is highly dependent on the segmentation
algorithm used to compute the binary mask. In this paper, we propose a novel segmentation method based
on the MRC standards (ITU-T T.44). The algorithm consists of two components: Cost Optimized Segmentation
(COS) and Connected Component Classification (CCC). The COS algorithm is a blockwise segmentation algorithm
formulated in a global cost optimization framework, while CCC is based on feature vector classification of
connected components. In the experimental results, we show that the new algorithm achieves the same accuracy
of text detection but with lower false detection of non-text features, as compared to state-of-the-art commercial
MRC products. This results in high quality MRC encoded documents with fewer non-text artifacts, and lower
bit rate.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.