Core Concepts
MEDBind integrates CXR, ECG, and text data for improved medical insights.
Abstract
Yuan Gao et al. introduce MEDBind for tri-modality binding in medical data.
The model enhances zero-shot learning and downstream tasks with competitive performance.
Contrastive loss functions like TMCL and EMCL are key to the model's success.
MEDBind improves cross-modality retrieval and classification tasks significantly.
The framework is scalable and open for future expansion to include additional modalities.
Stats
"We trained models for 150 epochs with batch size 128."
"Input dimensions were 224×224 for CXR and 12×1000 for ECG."
"The final embedding size was set to 256."
Quotes
"We present MEDBind (Medical Electronic patient recorD), which learns joint embeddings across CXR, ECG, and medical text."
"Contributions: MEDBind is the first tri-modality framework that employs contrastive learning to fuse CXR, ECG, and medical texts into a unified representation space."
"MEDBind outperformed all separately trained VLPM in total RSUM modality-to-text retrieval."
"MEDBindBD consistently beat MEDBindNM across all datasets and outperformed other models in three out of four datasets."
"MEDBindBD outperformed MEDBindNM and encoder zoo in cross-modality zero-shot classification tasks."