OpenPMC
Collection
Advancing Medical Representation Learning Through High-Quality Data
•
4 items
•
Updated
Arxiv: Arxiv | Code: Open-PMC Github | Dataset: Hugging Face
This model is a checkpoint trained on the Open-PMC dataset. It utilizes a Vision Transformer (ViT-base16) as the backbone for visual feature extraction and PubMedBERT for processing text data. The model is trained using Contrastive Learning with the vanilla Info-NCE loss to learn meaningful representations across different modalities.
The model was trained using the mmlearn framework, which is designed for multimodal learning. You can find more information and access the framework here.
Please visit out GitHub for information on how to run benchmarking using this checkpoint