Synthesizing Complex-Valued Multicoil MRI Data from Magnitude-Only Images

Author:

Deveshwar Nikhil123ORCID,Rajagopal Abhejit2,Sahin Sule12ORCID,Shimron Efrat3,Larson Peder E. Z.12ORCID

Affiliation:

1. UC Berkeley-UCSF Graduate Program in Bioengineering, Berkeley, CA 94701, USA

2. Department of Radiology and Biomedical Imaging, University of California, San Francisco, CA 94016, USA

3. Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, CA 94701, USA

Abstract

Despite the proliferation of deep learning techniques for accelerated MRI acquisition and enhanced image reconstruction, the construction of large and diverse MRI datasets continues to pose a barrier to effective clinical translation of these technologies. One major challenge is in collecting the MRI raw data (required for image reconstruction) from clinical scanning, as only magnitude images are typically saved and used for clinical assessment and diagnosis. The image phase and multi-channel RF coil information are not retained when magnitude-only images are saved in clinical imaging archives. Additionally, preprocessing used for data in clinical imaging can lead to biased results. While several groups have begun concerted efforts to collect large amounts of MRI raw data, current databases are limited in the diversity of anatomy, pathology, annotations, and acquisition types they contain. To address this, we present a method for synthesizing realistic MR data from magnitude-only data, allowing for the use of diverse data from clinical imaging archives in advanced MRI reconstruction development. Our method uses a conditional GAN-based framework to generate synthetic phase images from input magnitude images. We then applied ESPIRiT to derive RF coil sensitivity maps from fully sampled real data to generate multi-coil data. The synthetic data generation method was evaluated by comparing image reconstruction results from training Variational Networks either with real data or synthetic data. We demonstrate that the Variational Network trained on synthetic MRI data from our method, consisting of GAN-derived synthetic phase and multi-coil information, outperformed Variational Networks trained on data with synthetic phase generated using current state-of-the-art methods. Additionally, we demonstrate that the Variational Networks trained with synthetic k-space data from our method perform comparably to image reconstruction networks trained on undersampled real k-space data.

Funder

National Institutes of Health

University of California, San Francisco, AI Imaging and Cancer Award

Publisher

MDPI AG

Subject

Bioengineering

Reference49 articles.

1. Advancing machine learning for MR image reconstruction with an open competition: Overview of the 2019 fastMRI challenge;Knoll;Magn. Reson. Med.,2020

2. Mridata.org: An open archive for sharing MRI raw data;Ong;Proc. Intl. Soc. Mag. Reson. Med.,2018

3. (2021, March 22). Calgary Campinas Public Dataset. Available online: https://sites.google.com/view/calgary-campinas-dataset/mr-reconstruction-challenge?authuser=0.

4. Desai, A.D., Schmidt, A.M., Rubin, E.B., Sandino, C.M., Black, M.S., Mazzoli, V., Stevens, K.J., Boutin, R., Re, C., and Gold, G.E. (2021). Proceedings of the Thirty-Fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2), Stanford University.

5. Lim, Y., Toutios, A., Bliesener, Y., Tian, Y., Lingala, S.G., Vaz, C., Sorensen, T., Oh, M., Harper, S., and Chen, W. (2021). A multispeaker dataset of raw and reconstructed speech production real-time MRI video and 3D volumetric images. arXiv.

Cited by 2 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3