We found a match
Your institution may have access to this item. Find your institution then sign in to continue.
- Title
Split‐slice training and hyperparameter tuning of RAKI networks for simultaneous multi‐slice reconstruction.
- Authors
Nencka, Andrew S.; Arpinar, Volkan E.; Bhave, Sampada; Yang, Baolian; Banerjee, Suchandrima; McCrea, Michael; Mickevicius, Nikolai J.; Muftuler, L. Tugan; Koch, Kevin M.
- Abstract
Purpose: Simultaneous multi‐slice acquisitions are essential for modern neuroimaging research, enabling high temporal resolution functional and high‐resolution q‐space sampling diffusion acquisitions. Recently, deep learning reconstruction techniques have been introduced for unaliasing these accelerated acquisitions, and robust artificial‐neural‐networks for k‐space interpolation (RAKI) have shown promising capabilities. This study systematically examines the impacts of hyperparameter selections for RAKI networks, and introduces a novel technique for training data generation which is analogous to the split‐slice formalism used in slice‐GRAPPA. Methods: RAKI networks were developed with variable hyperparameters and with and without split‐slice training data generation. Each network was trained and applied to five different datasets including acquisitions harmonized with Human Connectome Project lifespan protocol. Unaliasing performance was assessed through L1 errors computed between unaliased and calibration frequency‐space data. Results: Split‐slice training significantly improved network performance in nearly all hyperparameter configurations. Best unaliasing results were achieved with three layer RAKI networks using at least 64 convolutional filters with receptive fields of 7 voxels, 128 single‐voxel filters in the penultimate RAKI layer, batch normalization, and no training dropout with the split‐slice augmented training dataset. Networks trained without the split‐slice technique showed symptoms of network over‐fitting. Conclusions: Split‐slice training for simultaneous multi‐slice RAKI networks positively impacts network performance. Hyperparameter tuning of such reconstruction networks can lead to further improvements in unaliasing performance.
- Subjects
NETWORK performance; DEEP learning; SYMPTOMS; IMAGE reconstruction; INTERPOLATION
- Publication
Magnetic Resonance in Medicine, 2021, Vol 85, Issue 6, p3272
- ISSN
0740-3194
- Publication type
Article
- DOI
10.1002/mrm.28634