Models Genesis in MICCAI-2019

Hello everyone. My name is Zongwei Zhou. Our paper provides pre-trained 3D models, which learn representation directly from a large number of unlabeled images to generate powerful target models through transfer learning. So, we nickname these generic models as Models Genesis.

Nowadays, ImageNet-based transfer learning is a common practice in medical image analysis. In contrast, our Models Genesis are different from ImageNet models in three ways:

First, we pre-train models directly on medical images, while ImageNet models are pre-trained on natural images. We believe that transfer learning from medical images to medical images should be more powerful than from natural images to medical images.

Second, Models Genesis are trained in 3D directly, while ImageNet models have to be trained in 2D. The most prominent medical modalities are formed in 3D, such as CT and MRI. To fit the ImageNet-based transfer learning paradigm, we have to solve a 3D problem in 2D, which definitely loses 3D spatial information and inevitably compromises the performance. We believe that 3D imaging tasks should be solved in 3D.

Most importantly, ImageNet demands a huge amount of annotation efforts, but we pre-trained Models Genesis by self-supervised learning without using any expert annotation.

Here shows the diagram of our self-supervised learning framework. We design it as a simple image restoration task. Given an image, we first deform it and then feed into a model, let the model restore the original image. We adopt the V-Net structure. The ground truth is the original image, and the input is the deformed image. To deform-an image, we propose four different approaches.

First, non-linear transformation. In CT scans, the pixel intensity of air, organ, or bone, has to be in a certain range of Hounsfield Unit, which means, CT scan itself naturally comes with the pixel-wise annotation. Therefore, we apply a non-linear translating function to the CT images. By restoring the original intensity values, the model needs to learn from the organ appearance, including shape and intensity distribution.

Second, local pixel shuffling. We randomly shuffle the pixel position within small regions and then let the model learn to recover the original image. By doing so, the model must learn the organ texture and local boundaries.

We also have image out-painting and in-painting, where some small regions are hidden from the model by random numbers. In out-painting, to restore the original image, the model must learn from organ spatial layout and global geometry by extrapolation; and in in-painting, the model must learn the local continuities of organs by interpolation.

We combined these four deformations together to let the model learn from multiple perspectives. Our ablation study shows that the combined approach performs more robust compared with each individual. Also, our self-learning framework is scalable because it is easy to incorporate any other meaningful image deformations.

After pre-training, the encoder could be used for target classification tasks, and the encoder-decoder together could be used for target segmentation tasks. We have evaluated Models Genesis on seven different medical applications, including classification and segmentation on CT, MRI, Ultrasound, and X-ray images, across diseases and organs.

I will present the major conclusions here. First thing first, 3D models are critical to utilize spatial information offered by 3D medical images. However, training 3D models directly is not easy, because they usually have more parameters to optimize. Two out of three target applications show that they perform equivalent or even worse than 2D ImageNet models. But our pre-trained Models Genesis significantly outperform 3D models trained from scratch. Here, the red bar is our Models Genesis, and the grey one is learning 3D models from scratch. Furthermore, 3D Models Genesis consistently outperform any 2D approaches, including state-of-the-art ImageNet models and our Models Genesis in 2D. We here introduce these degraded 2D Models Genesis to have an apple-to-apple comparison with ImageNet models. As seen, Models Genesis 2D offer performances equivalent to ImageNet models. This result is unprecedented because no self-supervised methods have thus far performed as well as ImageNet-based transfer learning.

Through all seven medical applications, we envision that Models Genesis may serve as a primary resource of 3D transfer learning for medical imaging. In this paper, we pre-trained Models Genesis only from LUNA16 and NIH Chest X-ray without using the labels tagged with these datasets. We plan to pre-train Models Genesis on other body regions and other modalities such as MRI, and eventually pre-train Models Genesis from all the available medical images on the Internet. We make the development of Models Genesis open science and invite researchers around the world to contribute to this effort. We hope that our collective efforts will lead to the Holy Grail of Models Genesis, effective across diseases, organs, and modalities.

For more information, please join us tomorrow for the poster session.


Talk in Mila

Hello everyone. My name is Zongwei Zhou. Our paper provides pre-trained 3D models, which learn representation directly from a large number of unlabeled images to generate powerful target models through transfer learning. So, we nickname these generic models as Models Genesis.

Nowadays, ImageNet-based transfer learning is a common practice in medical image analysis. In contrast, our Models Genesis are different from ImageNet models in three ways:

First, we pre-train models directly on medical images, while ImageNet models are pre-trained on natural images. We believe that transfer learning from medical images to medical images should be more powerful than from natural images to medical images.

Second, Models Genesis are trained in 3D directly, while ImageNet models have to be trained in 2D. The most prominent medical modalities are formed in 3D, such as CT and MRI. To fit the ImageNet-based transfer learning paradigm, we have to solve a 3D problem in 2D, which definitely loses 3D spatial information and inevitably compromises the performance. We believe that 3D imaging tasks should be solved in 3D.

Most importantly, ImageNet demands a huge amount of annotation efforts, but we pre-trained Models Genesis by self-supervised learning without using any expert annotation. And for the very first time, we are going to show you that our Models Genesis, even with zero expert annotation, outperform ImageNet-based transfer learning across diseases and organs.

Here shows the diagram of our self-supervised learning framework. We design it as a simple image restoration task. Given an image, we first deform it and then feed into a model, let the model restore the original image. We adopt the V-Net structure. The ground truth is the original image, and the input is the deformed image. To deform-an image, we propose four different approaches.

First, non-linear transformation. The intensity values in the CT scans have their own practical meanings. This is different from natural images. For example, in natural images, a flower can be any color, but in CT scans, the pixel intensity of air, organ, or bone, has to be in a certain Hounsfield Unit range. In other words, the intensity values in CT can be considered as a pixel-wise annotation.
Therefore, we apply a non-linear translating function to the CT images. By restoring the original intensity values, the model needs to learn from the organ appearance, including shape and intensity distribution.

Second, local pixel shuffling. We randomly shuffle the pixel position within small regions and then let the model learn to recover the original image. By doing so, the model must learn the organ texture and local boundaries.

We also have image out-painting and in-painting, where some small regions are hidden from the model by random numbers. In out-painting, to restore the original image, the model must learn from organ spatial layout and global geometry by extrapolation; and in in-painting, the model must learn the local continuities of organs by interpolation.

We combined these four deformations together to let the model learn from multiple perspectives. As shown in the figures, compared with each individual approach, the combination does not always offer the best performance,
but in case it is not the best, it performs as well as the best, statistically.
The combined approach shows more robust across all five target tasks. Also, our self-learning framework is scalable because it is easy to incorporate any other meaningful image deformations.

Here comes the question, what is a meaningful image deformation to our framework? You may think, okay, these guys just did some sort of fancy data augmentation to the image and ask the model to restore the original one.

...

After pre-training, the encoder could be used for target classification tasks, and the encoder-decoder together could be used for target segmentation tasks. We have evaluated Models Genesis on seven different medical applications, including classification and segmentation on CT, MRI, Ultrasound, and X-ray images, across diseases and organs.

I will present the major conclusions here. First thing first, 3D models are critical to utilize spatial information offered by 3D medical images. However, training 3D models directly is not easy, because they usually have more parameters to optimize. Two out of three target applications show that they perform equivalent or even worse than 2D ImageNet models. But our pre-trained Models Genesis significantly outperform 3D models trained from scratch. Here, the red bar is our Models Genesis, and the grey one is learning 3D models from scratch. Furthermore, 3D Models Genesis consistently outperform any 2D approaches, including state-of-the-art ImageNet models and our Models Genesis in 2D. We here introduce these degraded 2D Models Genesis to have an apple-to-apple comparison with ImageNet models. As seen, Models Genesis 2D offer performances equivalent to ImageNet models. This result is unprecedented because no self-supervised methods have thus far performed as well as ImageNet-based transfer learning.

Do you really need to build a large scale Medical ImageNet?
...
Therefore, considering the domain gap between medical imaging and natural imaging, we conclude that a large scale systematic labeled Medical ImageNet is necessary. Our Models Genesis are not designed to replace such a large, strongly annotated dataset for medical image analysis like ImageNet for computer vision, but rather helping create one.

Through all seven medical applications, we envision that Models Genesis may serve as a primary resource of 3D transfer learning for medical imaging. In this paper, we pre-trained Models Genesis only from LUNA16 and NIH Chest X-ray without using the labels tagged with these datasets. We plan to pre-train Models Genesis on other body regions and other modalities such as MRI, and eventually pre-train Models Genesis from all the available medical images on the Internet. We make the development of Models Genesis open science and invite researchers around the world to contribute to this effort. We hope that our collective efforts will lead to the Holy Grail of Models Genesis, effective across diseases, organs, and modalities.

I would like to thank all the co-authors: Vatsal, Mahfuzur, Ruibin, Nima, Dr. Gotway, and Dr. Liang.

最后编辑于
©著作权归作者所有,转载或内容合作请联系作者
  • 序言:七十年代末,一起剥皮案震惊了整个滨河市,随后出现的几起案子,更是在滨河造成了极大的恐慌,老刑警刘岩,带你破解...
    沈念sama阅读 203,098评论 5 476
  • 序言:滨河连续发生了三起死亡事件,死亡现场离奇诡异,居然都是意外死亡,警方通过查阅死者的电脑和手机,发现死者居然都...
    沈念sama阅读 85,213评论 2 380
  • 文/潘晓璐 我一进店门,熙熙楼的掌柜王于贵愁眉苦脸地迎上来,“玉大人,你说我怎么就摊上这事。” “怎么了?”我有些...
    开封第一讲书人阅读 149,960评论 0 336
  • 文/不坏的土叔 我叫张陵,是天一观的道长。 经常有香客问我,道长,这世上最难降的妖魔是什么? 我笑而不...
    开封第一讲书人阅读 54,519评论 1 273
  • 正文 为了忘掉前任,我火速办了婚礼,结果婚礼上,老公的妹妹穿的比我还像新娘。我一直安慰自己,他们只是感情好,可当我...
    茶点故事阅读 63,512评论 5 364
  • 文/花漫 我一把揭开白布。 她就那样静静地躺着,像睡着了一般。 火红的嫁衣衬着肌肤如雪。 梳的纹丝不乱的头发上,一...
    开封第一讲书人阅读 48,533评论 1 281
  • 那天,我揣着相机与录音,去河边找鬼。 笑死,一个胖子当着我的面吹牛,可吹牛的内容都是我干的。 我是一名探鬼主播,决...
    沈念sama阅读 37,914评论 3 395
  • 文/苍兰香墨 我猛地睁开眼,长吁一口气:“原来是场噩梦啊……” “哼!你这毒妇竟也来了?” 一声冷哼从身侧响起,我...
    开封第一讲书人阅读 36,574评论 0 256
  • 序言:老挝万荣一对情侣失踪,失踪者是张志新(化名)和其女友刘颖,没想到半个月后,有当地人在树林里发现了一具尸体,经...
    沈念sama阅读 40,804评论 1 296
  • 正文 独居荒郊野岭守林人离奇死亡,尸身上长有42处带血的脓包…… 初始之章·张勋 以下内容为张勋视角 年9月15日...
    茶点故事阅读 35,563评论 2 319
  • 正文 我和宋清朗相恋三年,在试婚纱的时候发现自己被绿了。 大学时的朋友给我发了我未婚夫和他白月光在一起吃饭的照片。...
    茶点故事阅读 37,644评论 1 329
  • 序言:一个原本活蹦乱跳的男人离奇死亡,死状恐怖,灵堂内的尸体忽然破棺而出,到底是诈尸还是另有隐情,我是刑警宁泽,带...
    沈念sama阅读 33,350评论 4 318
  • 正文 年R本政府宣布,位于F岛的核电站,受9级特大地震影响,放射性物质发生泄漏。R本人自食恶果不足惜,却给世界环境...
    茶点故事阅读 38,933评论 3 307
  • 文/蒙蒙 一、第九天 我趴在偏房一处隐蔽的房顶上张望。 院中可真热闹,春花似锦、人声如沸。这庄子的主人今日做“春日...
    开封第一讲书人阅读 29,908评论 0 19
  • 文/苍兰香墨 我抬头看了看天上的太阳。三九已至,却和暖如春,着一层夹袄步出监牢的瞬间,已是汗流浃背。 一阵脚步声响...
    开封第一讲书人阅读 31,146评论 1 259
  • 我被黑心中介骗来泰国打工, 没想到刚下飞机就差点儿被人妖公主榨干…… 1. 我叫王不留,地道东北人。 一个月前我还...
    沈念sama阅读 42,847评论 2 349
  • 正文 我出身青楼,却偏偏与公主长得像,于是被迫代替她去往敌国和亲。 传闻我的和亲对象是个残疾皇子,可洞房花烛夜当晚...
    茶点故事阅读 42,361评论 2 342

推荐阅读更多精彩内容