RMSim: controlled respiratory motion simulation on static patient scans.
Donghoon LeeEllen YorkeMasoud ZarepishehSaad NadeemYu-Chi HuPublished in: Physics in medicine and biology (2023)
Objective. This work aims to generate realistic anatomical deformations from static patient scans. Specifically, we present a method to generate these deformations/augmentations via deep learning driven respiratory motion simulation that provides the ground truth for validating deformable image registration (DIR) algorithms and driving more accurate deep learning based DIR. Approach. We present a novel 3D Seq2Seq deep learning respiratory motion simulator (RMSim) that learns from 4D-CT images and predicts future breathing phases given a static CT image. The predicted respiratory patterns, represented by time-varying displacement vector fields (DVFs) at different breathing phases, are modulated through auxiliary inputs of 1D breathing traces so that a larger amplitude in the trace results in more significant predicted deformation. Stacked 3D-ConvLSTMs are used to capture the spatial-temporal respiration patterns. Training loss includes a smoothness loss in the DVF and mean-squared error between the predicted and ground truth phase images. A spatial transformer deforms the static CT with the predicted DVF to generate the predicted phase image. 10-phase 4D-CTs of 140 internal patients were used to train and test RMSim. The trained RMSim was then used to augment a public DIR challenge dataset for training VoxelMorph to show the effectiveness of RMSim-generated deformation augmentation. Main results. We validated our RMSim output with both private and public benchmark datasets (healthy and cancer patients). The structure similarity index measure (SSIM) for predicted breathing phases and ground truth 4D CT images was 0.92 ± 0.04, demonstrating RMSim's potential to generate realistic respiratory motion. Moreover, the landmark registration error in a public DIR dataset was improved from 8.12 ± 5.78 mm to 6.58mm ± 6.38 mm using RMSim-augmented training data. Significance. The proposed approach can be used for validating DIR algorithms as well as for patient-specific augmentations to improve deep learning DIR algorithms. The code, pretrained models, and augmented DIR validation datasets will be released athttps://github.com/nadeemlab/SeqX2Y.
Keyphrases
- deep learning
- virtual reality
- dual energy
- computed tomography
- convolutional neural network
- contrast enhanced
- artificial intelligence
- image quality
- machine learning
- healthcare
- rna seq
- high speed
- positron emission tomography
- respiratory tract
- end stage renal disease
- single cell
- big data
- case report
- ejection fraction
- magnetic resonance imaging
- genome wide
- newly diagnosed
- gene expression
- heavy metals
- peritoneal dialysis
- health insurance
- electronic health record
- chronic kidney disease
- current status
- dna methylation
- resistance training
- patient reported outcomes
- magnetic resonance
- mass spectrometry
- optical coherence tomography
- high intensity