pith. machine review for the scientific record. sign in

arxiv: 2604.27277 · v1 · submitted 2026-04-30 · 💻 cs.LG · cs.AI· cs.CV

Recognition: unknown

BrainDINO: A Brain MRI Foundation Model for Generalizable Clinical Representation Learning

Authors on Pith no claims yet

Pith reviewed 2026-05-07 08:42 UTC · model grok-4.3

classification 💻 cs.LG cs.AIcs.CV
keywords brain MRIself-supervised learningfoundation modelrepresentation learningtransfer learningneuroimagingclinical tasksgeneralization
0
0 comments X

The pith

A self-supervised model trained on millions of brain MRI slices yields a unified representation that supports diverse clinical tasks with a frozen encoder and lightweight heads.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper demonstrates that a self-distilled model trained on roughly 6.6 million unlabeled axial brain MRI slices drawn from 20 datasets spanning varied populations, diseases, and scanners can produce one representation usable for many different endpoints. Tasks include tumor segmentation, classification of neurodegenerative and neurodevelopmental conditions, brain age estimation, post-stroke temporal prediction, molecular status prediction, sequence classification, and survival modeling. A reader would care because conventional methods usually demand task-specific models and large labeled datasets, which are costly and often unavailable, while this approach works with minimal additional training and shows particular strength when labels are scarce. The learned features turn out to be organized by anatomy and sensitive to pathology even though no task labels guided the initial training.

Core claim

BrainDINO is trained via self-distillation on approximately 6.6 million unlabeled axial slices from 20 heterogeneous datasets. With its encoder frozen and only lightweight task heads attached, the model matches or exceeds natural-image and MRI-specific self-supervised baselines on tumor segmentation, condition classification, age estimation, post-stroke prediction, molecular status prediction, sequence classification, and survival modeling, with the largest gains under low-label regimes. Representation analysis shows the features are anatomically organized and pathology-sensitive despite the complete absence of task supervision during pretraining. These results establish that large-scale, s

What carries the argument

BrainDINO, the self-distilled foundation model that learns generalizable features from unlabeled brain MRI slices through slice-wise self-supervision, allowing transfer by freezing the encoder and training only small task heads.

If this is right

  • The same pretrained encoder can be reused for new tasks by training only small heads, reducing the need for large labeled datasets per application.
  • Advantages are greatest when labeled examples are limited, enabling practical use in data-scarce clinical environments.
  • Anatomically organized and pathology-sensitive features emerge without task labels, potentially aiding clinical interpretation.
  • Slice-wise processing suffices, removing the requirement for volumetric pretraining or full-network fine-tuning on new tasks.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same slice-wise self-supervised strategy could be tested on other imaging modalities or body regions to create comparable foundation models.
  • Hospitals might adapt the model to new clinical questions more rapidly because only lightweight heads need retraining.
  • The unified representation might be combined with non-imaging data such as patient records to improve outcome prediction in future studies.

Load-bearing premise

The 20 selected datasets and their clinical endpoints are representative enough of all brain MRI populations, diseases, and acquisition settings to support broad claims of generalizability.

What would settle it

Performance on a new brain MRI dataset from an unseen scanner vendor or patient population falls substantially below that of task-specific supervised models trained on the same target data.

read the original abstract

Brain MRI underpins a wide range of neuroscientific and clinical applications, yet most learning-based methods remain task-specific and require substantial labeled data. Here we show that a single self-supervised representation can generalize across heterogeneous brain MRI endpoints. We trained BrainDINO, a self-distilled foundation model, on approximately 6.6 million unlabeled axial slices from 20 datasets encompassing broad variation in population, disease, and acquisition setting. Using a frozen encoder with lightweight task heads, BrainDINO supported transfer across tumor segmentation, neurodegenerative and neurodevelopmental conditions classification, brain age estimation, post-stroke temporal prediction, molecular status prediction, MRI sequence classification, and survival modeling. Across tasks and supervision regimes, BrainDINO consistently equaled or exceeded natural-image and MRI-specific self-supervised baselines, with particularly strong advantages under label scarcity. Representation analyses further showed anatomically organized and pathology-sensitive feature structure in the absence of task-specific supervision. Our findings indicate that large-scale slice-wise self-supervised learning can yield a unified brain MRI representation that supports diverse neuroimaging tasks without volumetric pretraining or full-network fine-tuning, establishing a scalable foundation for robust and data-efficient brain imaging analysis.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript introduces BrainDINO, a DINO-style self-supervised foundation model pretrained on ~6.6 million unlabeled axial brain MRI slices drawn from 20 datasets. It claims that a frozen encoder plus lightweight task heads yields a unified representation that transfers to diverse clinical endpoints (tumor segmentation, neurodegenerative/neurodevelopmental classification, brain-age regression, post-stroke temporal prediction, molecular-status prediction, sequence classification, and survival modeling), equaling or exceeding natural-image and MRI-specific SSL baselines especially under label scarcity, while producing anatomically organized and pathology-sensitive features without task-specific supervision or volumetric pretraining.

Significance. If the empirical claims hold, the work would be significant for medical-image foundation-model research by showing that large-scale 2D slice-wise self-supervision on heterogeneous data can produce a versatile brain-MRI representation without 3D volumetric pretraining or full-network fine-tuning. The breadth of downstream tasks and the focus on low-label regimes are clear strengths; the representation analyses further support the utility of the learned features.

major comments (3)
  1. [Pretraining data description] Pretraining-data section: the assertion of 'broad variation in population, disease, and acquisition setting' across the 20 datasets is not accompanied by quantitative coverage metrics (scanner vendors, field-strength distributions, slice-thickness histograms, orientation statistics, or demographic strata). This directly underpins the central generalizability claim and must be addressed with explicit tables or figures.
  2. [Downstream tasks and evaluation] Downstream evaluation: no explicit held-out OOD acquisition protocols (different vendors, non-axial orientations, or unseen field strengths) are tested; downstream tasks appear drawn from distributions overlapping the pretraining pool. This weakens the claim that the representation supports 'arbitrary' heterogeneous brain-MRI endpoints.
  3. [Results and baselines] Results presentation: the abstract and main-text claims of 'consistent outperformance' and 'particularly strong advantages under label scarcity' are not supported by reported quantitative metrics, statistical tests, error bars, or baseline-implementation details sufficient for verification. This is load-bearing for the empirical contribution.
minor comments (2)
  1. Figure captions should explicitly state the metrics, number of runs, and statistical tests shown in each panel.
  2. [Transfer learning protocol] The exact architecture and hyper-parameters of the 'lightweight task heads' used for transfer should be detailed in a table or appendix for reproducibility.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for their constructive comments, which have helped us improve the clarity and rigor of our manuscript. We address each major comment point by point below, providing the strongest honest responses possible based on the current work. Revisions have been made where the comments identify clear gaps in presentation or evidence.

read point-by-point responses
  1. Referee: Pretraining-data section: the assertion of 'broad variation in population, disease, and acquisition setting' across the 20 datasets is not accompanied by quantitative coverage metrics (scanner vendors, field-strength distributions, slice-thickness histograms, orientation statistics, or demographic strata). This directly underpins the central generalizability claim and must be addressed with explicit tables or figures.

    Authors: We agree that quantitative metrics are necessary to substantiate the diversity claim. In the revised manuscript we have added Table 1, which reports scanner vendor distributions, field strength percentages, slice thickness histograms, orientation statistics, and available demographic strata (age, sex) aggregated across all 20 pretraining datasets. We have also included a supplementary figure with per-dataset breakdowns. These additions directly support the heterogeneity assertion without altering any experimental results. revision: yes

  2. Referee: Downstream evaluation: no explicit held-out OOD acquisition protocols (different vendors, non-axial orientations, or unseen field strengths) are tested; downstream tasks appear drawn from distributions overlapping the pretraining pool. This weakens the claim that the representation supports 'arbitrary' heterogeneous brain-MRI endpoints.

    Authors: We acknowledge that the downstream tasks largely use axial acquisitions that overlap with the pretraining distribution in scanner and orientation characteristics. The 20 pretraining datasets already span multiple vendors, field strengths, and patient populations, and several downstream tasks introduce unseen pathologies and demographics. In revision we have added an explicit limitations paragraph discussing the scope of current OOD testing and have included a small-scale supplementary experiment on non-axial slices from one external dataset. Full arbitrary-heterogeneity validation would require additional held-out data collection beyond the scope of this study. revision: partial

  3. Referee: Results presentation: the abstract and main-text claims of 'consistent outperformance' and 'particularly strong advantages under label scarcity' are not supported by reported quantitative metrics, statistical tests, error bars, or baseline-implementation details sufficient for verification. This is load-bearing for the empirical contribution.

    Authors: We have revised the results section to include comprehensive tables reporting mean performance, standard deviations across five random seeds, and paired statistical tests (Wilcoxon signed-rank with Bonferroni correction) for all tasks and label regimes. Error bars have been added to every figure. Baseline implementation details (architectures, hyperparameters, training schedules) are now fully specified in the methods and supplementary material, enabling direct reproduction. These changes make the quantitative support for our claims explicit and verifiable. revision: yes

Circularity Check

0 steps flagged

No circularity: purely empirical self-supervised pretraining and transfer evaluation

full rationale

The paper describes standard self-supervised pretraining of a DINO-style model on ~6.6M unlabeled axial brain MRI slices drawn from 20 datasets, followed by frozen-encoder transfer to a suite of downstream supervised tasks (segmentation, classification, survival, etc.) using lightweight heads. No equations, parameter-fitting steps, or self-referential definitions appear in the abstract or described methodology that would make reported performance metrics equivalent to the pretraining inputs by construction. Downstream results are measured on labeled evaluation sets that are distinct from the unlabeled pretraining corpus; no fitted hyperparameters from downstream tasks are fed back into the pretraining objective, and no uniqueness theorems or ansatzes are invoked via self-citation to force the architecture or loss. The central claim of cross-task generalization is therefore an empirical observation rather than a tautology, placing the work in the normal non-circular category for large-scale representation-learning papers.

Axiom & Free-Parameter Ledger

1 free parameters · 2 axioms · 0 invented entities

Abstract-only review prevents extraction of exact hyperparameters or architectural choices; the central claim rests on the domain assumption that large-scale unlabeled slice-wise self-supervision captures clinically relevant features transferable to supervised tasks.

free parameters (1)
  • unspecified pretraining hyperparameters
    Standard self-supervised training involves many choices (learning rate, augmentation strength, distillation temperature) that are not reported in the abstract.
axioms (2)
  • domain assumption Self-supervised learning on unlabeled brain MRI slices produces features that are useful for downstream supervised clinical tasks without task-specific pretraining.
    This is the load-bearing premise that allows the frozen encoder plus lightweight heads to succeed across the listed endpoints.
  • domain assumption The 20 datasets together represent sufficient variation in population, disease, and acquisition to support claims of generalizability.
    Invoked when the abstract states the model 'supported transfer across' the listed tasks from heterogeneous sources.

pith-pipeline@v0.9.0 · 5536 in / 1565 out tokens · 68207 ms · 2026-05-07T08:42:33.520839+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

76 extracted references · 15 canonical work pages · 7 internal anchors

  1. [1]

    H.et al.The multimodal brain tumor image segmentation benchmark (brats).IEEE transactions on medical imaging34, 1993–2024 (2014)

    Menze, B. H.et al.The multimodal brain tumor image segmentation benchmark (brats).IEEE transactions on medical imaging34, 1993–2024 (2014)

  2. [2]

    Bakas, S.et al.Advancing the cancer genome atlas glioma mri collections with expert segmentation labels and radiomic features.Scientific data4, 170117 (2017)

  3. [3]

    Isensee, F.et al.nnu-net: Self-adapting framework for u-net-based medical image segmentation.arXiv preprint arXiv:1809.10486(2018)

  4. [4]

    & Shen, D.Deep learning-based feature representation for ad/mci classification, 583–590 (Springer, 2013)

    Suk, H.-I. & Shen, D.Deep learning-based feature representation for ad/mci classification, 583–590 (Springer, 2013)

  5. [5]

    & Disease Neuroimaging Initiative, f

    Ebrahimi, A., Luo, S. & Disease Neuroimaging Initiative, f. t. A. Convolutional neural networks for alzheimer’s disease detection on mri images.Journal of Medical Imaging8, 024503–024503 (2021)

  6. [6]

    An efficient method for early alzheimer’s disease detection based on mri images using deep convolutional neural networks.Frontiers in Artificial Intelligence8, 1563016 (2025)

    Dardouri, S. An efficient method for early alzheimer’s disease detection based on mri images using deep convolutional neural networks.Frontiers in Artificial Intelligence8, 1563016 (2025)

  7. [7]

    H.et al.Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker.NeuroImage163, 115–124 (2017)

    Cole, J. H.et al.Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker.NeuroImage163, 115–124 (2017)

  8. [8]

    F., Vedaldi, A

    Peng, H., Gong, W., Beckmann, C. F., Vedaldi, A. & Smith, S. M. Accurate brain age prediction with lightweight deep neural networks.Medical image analysis68, 101871 (2021)

  9. [9]

    M.et al.Mri signatures of brain age and disease over the lifespan based on a deep brain network and 14 468 individuals worldwide.Brain143, 2312–2324 (2020)

    Bashyam, V. M.et al.Mri signatures of brain age and disease over the lifespan based on a deep brain network and 14 468 individuals worldwide.Brain143, 2312–2324 (2020)

  10. [10]

    Kickingereder, P.et al.Radiomic profiling of glioblastoma: identifying an imaging predictor of patient survival with improved performance over established clinical and radiologic risk models.Radiology280, 880–889 (2016)

  11. [11]

    Macyszyn, L.et al.Imaging patterns predict patient survival and molecular subtype in glioblastoma via machine learning techniques.Neuro-oncology18, 417–425 (2015)

  12. [12]

    & Merhof, D.Segmentation of brain tumors and patient survival prediction: Methods for the brats 2018 challenge, 3–12 (Springer, 2018)

    Weninger, L., Rippel, O., Koppers, S. & Merhof, D.Segmentation of brain tumors and patient survival prediction: Methods for the brats 2018 challenge, 3–12 (Springer, 2018)

  13. [13]

    F., Kohl, S

    Isensee, F., Jaeger, P. F., Kohl, S. A., Petersen, J. & Maier-Hein, K. H. nnu-net: a self-configuring method for deep learning-based biomedical image segmentation.Nature methods18, 203–211 (2021)

  14. [14]

    Liu, Z.et al.Deep learning based brain tumor segmentation: a survey.Complex & intelligent systems 9, 1001–1026 (2023)

  15. [15]

    & Liu, M

    Guan, H. & Liu, M. Domain adaptation for medical image analysis: a survey.IEEE Transactions on Biomedical Engineering69, 1173–1185 (2021)

  16. [16]

    S., Oh, K., Shin, Y., Mazurowski, M

    Yoon, J. S., Oh, K., Shin, Y., Mazurowski, M. A. & Suk, H.-I. Domain generalization for medical image analysis: A review.Proceedings of the IEEE112, 1583–1609 (2024)

  17. [17]

    Eidex, Z.et al.Deep learning in mri-guided radiation therapy: A systematic review.Journal of applied clinical medical physics25, e14155 (2024)

  18. [18]

    & Hinton, G.A simple framework for contrastive learning of visual representations, 1597–1607 (PmLR, 2020)

    Chen, T., Kornblith, S., Norouzi, M. & Hinton, G.A simple framework for contrastive learning of visual representations, 1597–1607 (PmLR, 2020)

  19. [19]

    Chen, X., Fan, H., Girshick, R. & He, K. Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297(2020). 19

  20. [20]

    Grill, J.-B.et al.Bootstrap your own latent-a new approach to self-supervised learning.Advances in neural information processing systems33, 21271–21284 (2020)

  21. [21]

    Masked autoencoders are scalable vision learners, 16000–16009 (2022)

    He, K.et al. Masked autoencoders are scalable vision learners, 16000–16009 (2022)

  22. [22]

    Simmim: A simple framework for masked image modeling, 9653–9663 (2022)

    Xie, Z.et al. Simmim: A simple framework for masked image modeling, 9653–9663 (2022)

  23. [23]

    BEiT: BERT Pre-Training of Image Transformers

    Bao, H., Dong, L., Piao, S. & Wei, F. Beit: Bert pre-training of image transformers.arXiv preprint arXiv:2106.08254(2021)

  24. [24]

    Emerging properties in self-supervised vision transformers, 9650–9660 (2021)

    Caron, M.et al. Emerging properties in self-supervised vision transformers, 9650–9660 (2021)

  25. [25]

    Oquab, M.et al.Dinov2: Learning robust visual features without supervision.arXiv preprint arXiv:2304.07193(2023)

  26. [26]

    & Arbel, T.Building a general simclr self-supervised foundation model across neurological diseases to advance 3d brain mri diagnoses, 1310–1319 (2025)

    Kaczmarek, E., Szeto, J., Nichyporuk, B. & Arbel, T.Building a general simclr self-supervised foundation model across neurological diseases to advance 3d brain mri diagnoses, 1310–1319 (2025)

  27. [27]

    Tak, D.et al.A generalizable foundation model for analysis of human brain mri.Nature Neuroscience 1–12 (2026)

  28. [28]

    & Nielsen, M

    Munk, A., Ambsdorf, J., Llambias, S. & Nielsen, M. Amaes: Augmented masked autoencoder pre- training on public brain mri data for 3d-native segmentation.arXiv preprint arXiv:2408.00640 (2024)

  29. [29]

    & Moyal, E

    Robinet, L., Berjaoui, A. & Moyal, E. C.-J. Multimodal masked autoencoder pre-training for 3d mri-based brain tumor analysis with missing modalities.arXiv preprint arXiv:2505.00568(2025)

  30. [30]

    Multi-modal vision pre-training for medical image analysis, 5164–5174 (2025)

    Rui, S.et al. Multi-modal vision pre-training for medical image analysis, 5164–5174 (2025)

  31. [31]

    Mazher, M., Parker, G. J. & Alexander, D. C. Towards generalisable foundation models for brain mri. arXiv preprint arXiv:2510.23415(2025)

  32. [32]

    F., Smith, S

    Yang, C., Feng, J., Beckmann, C. F., Smith, S. M. & Gong, W. Genbrain: A generative foundation model of multimodal brain imaging.medRxiv2025–12 (2025)

  33. [33]

    Allah, A. M. G., Sarhan, A. M. & Elshennawy, N. M. Edge u-net: Brain tumor segmentation using mri based on deep u-net model with boundary information.Expert Systems with Applications213, 118833 (2023)

  34. [34]

    B., Fox, N

    Frisoni, G. B., Fox, N. C., Jack Jr, C. R., Scheltens, P. & Thompson, P. M. The clinical use of structural mri in alzheimer disease.Nature reviews neurology6, 67–77 (2010)

  35. [35]

    A., Small, S

    Feng, X., Provenzano, F. A., Small, S. A. & Initiative, A. D. N. A deep learning mri approach out- performs other biomarkers of prodromal alzheimer’s disease.Alzheimer’s research & therapy14, 45 (2022)

  36. [36]

    Sarica, A.et al.Explainability of random survival forests in predicting conversion risk from mild cognitive impairment to alzheimer’s disease.Brain informatics10, 31 (2023)

  37. [37]

    Zhang, Y., Teng, Q., Liu, Y., Liu, Y. & He, X. Diagnosis of alzheimer’s disease based on regional attention with smri gray matter slices.Journal of neuroscience methods365, 109376 (2022)

  38. [38]

    Liew, S.-L.et al.A large, open source dataset of stroke anatomical brain images and manual lesion segmentations.Scientific data5, 180011 (2018)

  39. [39]

    H.et al.Predicting survival in glioblastoma with multimodal neuroimaging and machine learning.Journal of Neuro-oncology164, 309–320 (2023)

    Luckett, P. H.et al.Predicting survival in glioblastoma with multimodal neuroimaging and machine learning.Journal of Neuro-oncology164, 309–320 (2023)

  40. [40]

    Sim´ eoni, O.et al.Dinov3.arXiv preprint arXiv:2508.10104(2025). 20

  41. [41]

    & Yang, X

    Li, Y., Wu, Y., Lai, Y., Hu, M. & Yang, X. Meddinov3: How to adapt vision foundation models for medical image segmentation?arXiv preprint arXiv:2509.02379(2025)

  42. [42]

    Baid, U.et al.The rsna-asnr-miccai brats 2021 benchmark on brain tumor segmentation and radiogenomic classification.arXiv preprint arXiv:2107.02314(2021)

  43. [43]

    W.et al.The brain tumor segmentation-metastases (brats-mets) challenge 2023: Brain metastasis segmentation on pre-treatment mri.arxivarXiv–2306 (2024)

    Moawad, A. W.et al.The brain tumor segmentation-metastases (brats-mets) challenge 2023: Brain metastasis segmentation on pre-treatment mri.arxivarXiv–2306 (2024)

  44. [44]

    LaBella, D.et al.The asnr-miccai brain tumor segmentation (brats) challenge 2023: Intracranial meningioma.arXiv preprint arXiv:2305.07642(2023)

  45. [45]

    Synapse (2024)

    BraTS-ISBI 2024 – generalizability across tumors challenge (BraTS-GoAT). Synapse (2024). URL https://www.synapse.org/Synapse:syn52939291. Synapse ID: syn52939291

  46. [46]

    Di Martino, A.et al.The autism brain imaging data exchange: towards a large-scale evaluation of the intrinsic brain architecture in autism.Molecular psychiatry19, 659–667 (2014)

  47. [47]

    C.et al.Alzheimer’s disease neuroimaging initiative (adni) clinical characterization

    Petersen, R. C.et al.Alzheimer’s disease neuroimaging initiative (adni) clinical characterization. Neurology74, 201–209 (2010)

  48. [48]

    Marcus, D. S.et al.Open access series of imaging studies (oasis): cross-sectional mri data in young, middle aged, nondemented, and demented older adults.Journal of cognitive neuroscience19, 1498–1507 (2007)

  49. [49]

    https://brain-development.org/ixi-dataset/ (2015)

    IXI – information extraction from images. https://brain-development.org/ixi-dataset/ (2015)

  50. [50]

    N., Weiss, Y., Yamasaki, B

    Wang, J., Lytle, M. N., Weiss, Y., Yamasaki, B. L. & Booth, J. R. A longitudinal neuroimaging dataset on language processing in children ages 5, 7, and 9 years old.Scientific Data9, 4 (2022)

  51. [51]

    & Saxe, R

    Richardson, H., Lisandrelli, G., Riobueno-Naylor, A. & Saxe, R. Mri data of 3–12 year old children and adults during viewing of a short animated film.Openneuro(2019)

  52. [52]

    Liew, S.-L.et al.A large, curated, open-source stroke neuroimaging dataset to improve lesion segmentation algorithms.Scientific data9, 320 (2022)

  53. [53]

    The Cancer Imaging Archive (2022)

    Calabrese, E.et al.The university of california san francisco preoperative diffuse glioma MRI (UCSF- PDGM). The Cancer Imaging Archive (2022). Version 5

  54. [54]

    F.et al.The brain tumor segmentation (brats) challenge 2023: focus on pediatrics (cbtn-connect-dipgr-asnr-miccai brats-peds).ArXivarXiv–2305 (2024)

    Kazerooni, A. F.et al.The brain tumor segmentation (brats) challenge 2023: focus on pediatrics (cbtn-connect-dipgr-asnr-miccai brats-peds).ArXivarXiv–2305 (2024)

  55. [55]

    Adewole, M.et al.The brain tumor segmentation (brats) challenge 2023: Glioma segmentation in sub-saharan africa patient population (brats-africa).ArXivarXiv–2305 (2023)

  56. [56]

    The Cancer Imaging Archive (2021)

    Bakas, S.et al.Multi-parametric magnetic resonance imaging (mpMRI) scans for de novo glioblastoma (GBM) patients from the university of pennsylvania health system (UPENN-GBM). The Cancer Imaging Archive (2021). Version 2

  57. [57]

    & Hinton, G.Similarity of neural network representations revisited, 3519–3529 (PMlR, 2019)

    Kornblith, S., Norouzi, M., Lee, H. & Hinton, G.Similarity of neural network representations revisited, 3519–3529 (PMlR, 2019)

  58. [58]

    Bommasani, R.et al.On the opportunities and risks of foundation models.arXiv preprint arXiv:2108.07258(2021)

  59. [59]

    R., Guttag, J

    Balakrishnan, G., Zhao, A., Sabuncu, M. R., Guttag, J. & Dalca, A. V. Voxelmorph: a learning framework for deformable medical image registration.IEEE transactions on medical imaging38, 1788–1800 (2019)

  60. [60]

    21 Series B: Biological Sciences356, 1293–1322 (2001)

    Mazziotta, J.et al.A probabilistic atlas and reference system for the human brain: International consortium for brain mapping (icbm).Philosophical Transactions of the Royal Society of London. 21 Series B: Biological Sciences356, 1293–1322 (2001)

  61. [61]

    Snoek, L.et al.The amsterdam open mri collection, a set of multimodal mri datasets for individual difference analyses.Scientific data8, 85 (2021)

  62. [62]

    Marek, K.et al.The parkinson progression marker initiative (ppmi).Progress in neurobiology95, 629–635 (2011)

  63. [63]

    Scarpace, L.et al.The cancer genome atlas glioblastoma multiforme collection (tcga-gbm).The Cancer Imaging Archive(2016)

  64. [64]

    The Clinical Proteomic Tumor Analysis Consortium Glioblastoma Multiforme Collection (CPTAC-GBM)

    National Cancer Institute Clinical Proteomic Tumor Analysis Consortium (CPTAC). The Clinical Proteomic Tumor Analysis Consortium Glioblastoma Multiforme Collection (CPTAC-GBM). The Cancer Imaging Archive (2018). Version 16

  65. [65]

    & Bortfeld, T

    Shusharina, N. & Bortfeld, T. Glioma image segmentation for radiotherapy: RT targets, barriers to cancer spread, and organs at risk (GLIS-RT). The Cancer Imaging Archive (2021)

  66. [66]

    E., Jain, R., Mikkelsen, T

    Scarpace, L., Flanders, A. E., Jain, R., Mikkelsen, T. & Andrews, D. W. Data from REMBRANDT. The Cancer Imaging Archive (2019)

  67. [67]

    B.et al.An anatomic transcriptional atlas of human glioblastoma.Science360, 660–663 (2018)

    Puchalski, R. B.et al.An anatomic transcriptional atlas of human glioblastoma.Science360, 660–663 (2018)

  68. [68]

    C.et al.The 2024 brain tumor segmentation (brats) challenge: Glioma segmentation on post-treatment mri.arXiv preprint arXiv:2405.18368(2024)

    de Verdier, M. C.et al.The 2024 brain tumor segmentation (brats) challenge: Glioma segmentation on post-treatment mri.arXiv preprint arXiv:2405.18368(2024)

  69. [69]

    LaBella, D.et al.A multi-institutional meningioma mri dataset for automated multi-sequence image segmentation.Scientific data11, 496 (2024)

  70. [70]

    The Cancer Imaging Archive (2023)

    Vassantachart, A.et al.Segmentation and classification of grade I and II meningiomas from mag- netic resonance imaging: An open annotated dataset (Meningioma-SEG-CLASS). The Cancer Imaging Archive (2023)

  71. [71]

    Juvekar, P.et al.Remind: The brain resection multimodal imaging database.Scientific data11, 494 (2024)

  72. [72]

    Zbontar, J.et al.fastmri: An open dataset and benchmarks for accelerated mri.arXiv preprint arXiv:1811.08839(2018)

  73. [73]

    & Coombs, L

    Kinahan, P., Muzi, M., Bialecki, B., Herman, B. & Coombs, L. Data from ACRIN-DSC-MR-Brain. The Cancer Imaging Archive (2019)

  74. [74]

    R.et al.Acrin 6684 assessment of tumor hypoxia in glioblastoma using 18f- fluoromisonidazole with pet and mri

    Gerstner, E. R.et al.Acrin 6684 assessment of tumor hypoxia in glioblastoma using 18f- fluoromisonidazole with pet and mri. (2012)

  75. [75]

    Human brain mapping40, 4952–4964 (2019)

    Isensee, F.et al.Automated brain extraction of multisequence mri using artificial neural networks. Human brain mapping40, 4952–4964 (2019)

  76. [76]

    Zhou, J.et al.ibot: Image bert pre-training with online tokenizer.arXiv preprint arXiv:2111.07832 (2021). 22