Skip to content
/ M4Raw Public

A multi-contrast multi-repetition multi-channel MRI k-space dataset for low-field MRI research

License

Notifications You must be signed in to change notification settings

mylyu/M4Raw

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

50 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

M4Raw

M4Raw: A multi-contrast, multi-repetition, multi-channel MRI k-space dataset for low-field MRI research

The complete dataset V1.6 has been uploaded to Zenodo with test data of 25 subjects.

https://doi.org/10.5281/zenodo.8056074

DOI

Or you can download from BaiduYun

link:https://pan.baidu.com/s/1CqNFMDnrNuXe_qWSyiSA2w?pwd=4mjr

passcode:4mjr

Publication

If you find our work useful in your research or publication, please cite our work:

[1] Lyu, M., Mei, L., Huang, S. et al. M4Raw: A multi-contrast, multi-repetition, multi-channel MRI k-space dataset for low-field MRI research. Sci Data 10, 264 (2023). https://doi.org/10.1038/s41597-023-02181-4

@article{lyu_m4raw_2023,
 title = {{M4Raw}: {A} multi-contrast, multi-repetition, multi-channel {MRI} k-space dataset for low-field {MRI} research},
 volume = {10},
 issn = {2052-4463},
 url = {https://doi.org/10.1038/s41597-023-02181-4},
 doi = {10.1038/s41597-023-02181-4},
 abstract = {Recently, low-field magnetic resonance imaging (MRI) has gained renewed interest to promote MRI accessibility and affordability worldwide. The presented M4Raw dataset aims to facilitate methodology development and reproducible research in this field. The dataset comprises multi-channel brain k-space data collected from 183 healthy volunteers using a 0.3 Tesla whole-body MRI system, and includes T1-weighted, T2-weighted, and fluid attenuated inversion recovery (FLAIR) images with in-plane resolution of {\textasciitilde}1.2 mm and through-plane resolution of 5 mm. Importantly, each contrast contains multiple repetitions, which can be used individually or to form multi-repetition averaged images. After excluding motion-corrupted data, the partitioned training and validation subsets contain 1024 and 240 volumes, respectively. To demonstrate the potential utility of this dataset, we trained deep learning models for image denoising and parallel imaging tasks and compared their performance with traditional reconstruction methods. This M4Raw dataset will be valuable for the development of advanced data-driven methods specifically for low-field MRI. It can also serve as a benchmark dataset for general MRI reconstruction algorithms.},
 number = {1},
 journal = {Scientific Data},
 author = {Lyu, Mengye and Mei, Lifeng and Huang, Shoujin and Liu, Sixing and Li, Yi and Yang, Kexin and Liu, Yilong and Dong, Yu and Dong, Linzheng and Wu, Ed X.},
 month = may,
 year = {2023},
 pages = {264},
}

Abstract

Recently, low-field magnetic resonance imaging (MRI) has gained renewed interest to promote MRI accessibility and affordability worldwide. The presented M4Raw dataset aims to facilitate methodology development and reproducible research in this field. The dataset comprises multi-channel brain k-space data collected from 183 healthy volunteers using a 0.3 Tesla whole-body MRI system, and includes T1-weighted, T2-weighted, and fluid attenuated inversion recovery (FLAIR) images with in-plane resolution of ~1.2 mm and through-plane resolution of 5 mm. Importantly, each contrast contains multiple repetitions, which can be used individually or to form multi-repetition averaged images. After excluding motion-corrupted data, the partitioned training and validation subsets contain 1024 and 240 volumes, respectively. To demonstrate the potential utility of this dataset, we trained deep learning models for image denoising and parallel imaging tasks and compared their performance with traditional reconstruction methods. This M4Raw dataset will be valuable for the development of advanced data-driven methods specifically for low-field MRI. It can also serve as a benchmark dataset for general MRI reconstruction algorithms.

Quick Start

  1. Download the sample data of one subject (three T2w scans, three T1w scans, and two FLAIR scans) from this link. https://drive.google.com/drive/folders/1RQT9oqoBd0xsevhyKSCxKiLuoEb2erhB?usp=share_link

  2. pip install fastmri, which we will use for data visualization in the next step.

  3. run the jupter notebook "M4Raw_tutorial.ipynb".


TAKE A QUICK LOOK

News

2023/08/02: The M4Raw V1.6 has been released to include test data of 25 subjects https://zenodo.org/record/7998442.

2023/06/02: The M4Raw V1.5 has been released to include GRE data.

2023/02/06: The M4Raw pretrained model weights have been uploaded to github (see Release).

2023/01/11: The complete M4Raw dataset V1.1 has been released to zenodo.org.


** Note: M4Raw is a collaborative research project by Shenzhen Technology University (SZTU), the University of Hong Kong (HKU), Jinan University (JNU), and Shenzhen Samii Medical Center (SSMC).