Kenyon-Dean, K., Wang, Z. J., Urbanik, J., Donhauser, K., Hartford, J., Saberian, S., Sahin, N., Bendidi, I., Celik, S., Fay, M., ... & Kraus, O. (2024). ViTally Consistent: Scaling Biological Representation Learning for Cell Microscopy. Advances in Neural Information Processing Systems, 38.
This research paper aims to address the challenges of extracting meaningful and consistent biological features from large-scale cell microscopy images for downstream analysis in drug discovery and molecular biology research.
The researchers developed MAE-G/8, a 1.9 billion-parameter ViT-G/8 Masked Autoencoder, trained on a curated dataset called Phenoprints-16M, consisting of 16 million statistically significant positive cell microscopy image crops. They compared MAE-G/8's performance against several baseline models, including Dino-v2 backbones, weakly supervised and MAE ViT models pretrained on ImageNet, and a smaller MAE model trained on a different microscopy dataset. The evaluation involved linear probing tasks for gene perturbation and functional group classification, as well as whole-genome benchmarking using biological relationship recall and replicate consistency metrics.
The authors conclude that scaling training compute and parameters of self-supervised learning models for microscopy, specifically using Masked Autoencoders, significantly benefits downstream biological analysis. They propose a three-step approach for training and extracting optimal representations from self-supervised models trained on experimental data: (1) curate the training set for consistency, (2) train a scaled transformer-based model using self-supervised learning, and (3) evaluate the performance of each block to identify the optimal layer for representation.
This research significantly contributes to the field of computational biology by presenting a novel and highly effective approach for learning biologically meaningful representations from large-scale cell microscopy images. The proposed methodology and the development of MAE-G/8 have the potential to accelerate drug discovery and advance our understanding of biological processes.
While the study demonstrates the effectiveness of MAE-G/8, the authors acknowledge the computational demands of training and evaluating such large models. Future research could explore methods for improving the efficiency of these models or investigate alternative self-supervised learning techniques that might be more computationally tractable.
翻譯成其他語言
從原文內容
arxiv.org
深入探究