MMedS-Llama3

💻Github Repo 🖨️arXiv Paper

The official codes for "Towards Evaluating and Building Versatile Large Language Models for Medicine"

Introduction

This repository hosts MMedS-Llama-3-8B. Its foundation model, MMed-Llama-3-8B, is a multilingual medical language model which has undergone additional continuous pretraining on MMedC. Furthermore, the model has been fine-tuned under supervision using MedS-Ins, a comprehensive dataset designed specifically for supervised fine-tuning (SFT), featuring 13.5 million samples across 122 tasks. For more details, please refer to our paper.

Usage

The model can be loaded as follows:

import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Henrychur/MMed-Llama-3-8B-EnIns")
model = AutoModelForCausalLM.from_pretrained("Henrychur/MMed-Llama-3-8B-EnIns", torch_dtype=torch.float16)
  • Inference format is the same as Llama 3, you can check the inference code here.
Downloads last month
58
Safetensors
Model size
8.03B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for Henrychur/MMedS-Llama-3-8B

Unable to build the model tree, the base model loops to the model itself. Learn more.

Datasets used to train Henrychur/MMedS-Llama-3-8B

Space using Henrychur/MMedS-Llama-3-8B 1