TY - JOUR
T1 - Models Genesis
AU - Zhou, Zongwei
AU - Sodha, Vatsal
AU - Pang, Jiaxuan
AU - Gotway, Michael B.
AU - Liang, Jianming
N1 - Funding Information:
This research has been supported partially by ASU and Mayo Clinic through a Seed Grant and an Innovation Grant, and partially by the National Institutes of Health (NIH) under Award Number R01HL128785. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH. This work has utilized the GPUs provided partially by the ASU Research Computing and partially by the Extreme Science and Engineering Discovery Environment (XSEDE) funded by the National Science Foundation (NSF) under grant number ACI-1548562. We thank Z. Guo for implementing Rubik's Cube (Zhuang et al. 2019) and the 3D version of Jigsaw (Noroozi and Favaro, 2016) and DeepCluster (Caron et al. 2018); F. Haghighi and M. R. Hosseinzadeh Taher for implementing the 3D version of in-painting (Pathak et al. 2016), patch-shuffling (Chen et al. 2019a), and working with Z. Guo in evaluating the performance of MedicalNet (Chen et al. 2019b); M. M. Rahman Siddiquee for examining NiftyNet (Gibson et al. 2018b) with our Models Genesis; P. Zhang for comparing two additional random initialization methods with our Models Genesis; S. Bajpai for comparing three loss functions of the proxy task; N. Tajbakhsh for revising our conference paper; R. Feng for valuable discussions; and S. Tatapudi for helping improve the writing of this paper. The content of this paper is covered by patents pending.
Funding Information:
This research has been supported partially by ASU and Mayo Clinic through a Seed Grant and an Innovation Grant, and partially by the National Institutes of Health (NIH) under Award Number R01HL128785. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH. This work has utilized the GPUs provided partially by the ASU Research Computing and partially by the Extreme Science and Engineering Discovery Environment (XSEDE) funded by the National Science Foundation (NSF) under grant number ACI-1548562 . We thank Z. Guo for implementing Rubik’s Cube ( Zhuang et al., 2019 ) and the 3D version of Jigsaw ( Noroozi and Favaro, 2016 ) and DeepCluster ( Caron et al., 2018 ); F. Haghighi and M. R. Hosseinzadeh Taher for implementing the 3D version of in-painting ( Pathak et al., 2016 ), patch-shuffling ( Chen et al., 2019a ), and working with Z. Guo in evaluating the performance of MedicalNet ( Chen et al., 2019b ); M. M. Rahman Siddiquee for examining NiftyNet ( Gibson et al., 2018b ) with our Models Genesis; P. Zhang for comparing two additional random initialization methods with our Models Genesis; S. Bajpai for comparing three loss functions of the proxy task; N. Tajbakhsh for revising our conference paper; R. Feng for valuable discussions; and S. Tatapudi for helping improve the writing of this paper. The content of this paper is covered by patents pending.
Publisher Copyright:
© 2020 Elsevier B.V.
PY - 2021/1
Y1 - 2021/1
N2 - Transfer learning from natural image to medical image has been established as one of the most practical paradigms in deep learning for medical image analysis. To fit this paradigm, however, 3D imaging tasks in the most prominent imaging modalities (e.g., CT and MRI) have to be reformulated and solved in 2D, losing rich 3D anatomical information, thereby inevitably compromising its performance. To overcome this limitation, we have built a set of models, called Generic Autodidactic Models, nicknamed Models Genesis, because they are created ex nihilo (with no manual labeling), self-taught (learnt by self-supervision), and generic (served as source models for generating application-specific target models). Our extensive experiments demonstrate that our Models Genesis significantly outperform learning from scratch and existing pre-trained 3D models in all five target 3D applications covering both segmentation and classification. More importantly, learning a model from scratch simply in 3D may not necessarily yield performance better than transfer learning from ImageNet in 2D, but our Models Genesis consistently top any 2D/2.5D approaches including fine-tuning the models pre-trained from ImageNet as well as fine-tuning the 2D versions of our Models Genesis, confirming the importance of 3D anatomical information and significance of Models Genesis for 3D medical imaging. This performance is attributed to our unified self-supervised learning framework, built on a simple yet powerful observation: the sophisticated and recurrent anatomy in medical images can serve as strong yet free supervision signals for deep models to learn common anatomical representation automatically via self-supervision. As open science, all codes and pre-trained Models Genesis are available at https://github.com/MrGiovanni/ModelsGenesis.
AB - Transfer learning from natural image to medical image has been established as one of the most practical paradigms in deep learning for medical image analysis. To fit this paradigm, however, 3D imaging tasks in the most prominent imaging modalities (e.g., CT and MRI) have to be reformulated and solved in 2D, losing rich 3D anatomical information, thereby inevitably compromising its performance. To overcome this limitation, we have built a set of models, called Generic Autodidactic Models, nicknamed Models Genesis, because they are created ex nihilo (with no manual labeling), self-taught (learnt by self-supervision), and generic (served as source models for generating application-specific target models). Our extensive experiments demonstrate that our Models Genesis significantly outperform learning from scratch and existing pre-trained 3D models in all five target 3D applications covering both segmentation and classification. More importantly, learning a model from scratch simply in 3D may not necessarily yield performance better than transfer learning from ImageNet in 2D, but our Models Genesis consistently top any 2D/2.5D approaches including fine-tuning the models pre-trained from ImageNet as well as fine-tuning the 2D versions of our Models Genesis, confirming the importance of 3D anatomical information and significance of Models Genesis for 3D medical imaging. This performance is attributed to our unified self-supervised learning framework, built on a simple yet powerful observation: the sophisticated and recurrent anatomy in medical images can serve as strong yet free supervision signals for deep models to learn common anatomical representation automatically via self-supervision. As open science, all codes and pre-trained Models Genesis are available at https://github.com/MrGiovanni/ModelsGenesis.
KW - 3D Deep learning
KW - Representation learning
KW - Self-supervised learning
KW - Transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85095915555&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85095915555&partnerID=8YFLogxK
U2 - 10.1016/j.media.2020.101840
DO - 10.1016/j.media.2020.101840
M3 - Article
C2 - 33188996
AN - SCOPUS:85095915555
SN - 1361-8415
VL - 67
JO - Medical image analysis
JF - Medical image analysis
M1 - 101840
ER -