Skip to content

MedRegA: Interpretable Bilingual Multimodal Large Language Model for Diverse Biomedical Tasks

Notifications You must be signed in to change notification settings

xmed-lab/MedRegA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

MedRegA: Interpretable Bilingual Multimodal Large Language Model for Diverse Biomedical Tasks

MedRegA, an interpretable bilingual generalist model for diverse biomedical tasks, represented by its outstanding ability to leverage regional information. MedRegA can perceive 8 modalities covering almost all the body parts, showcasing significant versatility.

Overview

💡We establish Region-Centric tasks with a large-scale dataset, MedRegInstruct, where each sample is paired with coordinates of body structures or lesions.

💡Based on the proposed dataset, we develop a Region-Aware medical MLLM, MedRegA, as a bilingual generalist medical AI system to perform both image-level and region-level medical vision-language tasks, demonstrating impressive versatility.

Schedule

  • Release the model.
  • Release the demo code.
  • Release the evaluation code.
  • Release the training code.
  • Release the data.

Environment

Please refer to InternVL Installation to build the environment.

Demo

Run the demo:

torchrun --nproc-per-node=1 src/demo.py

Acknowledgement

We refer to the codes from InternVL. Thank the authors for releasing their code.

About

MedRegA: Interpretable Bilingual Multimodal Large Language Model for Diverse Biomedical Tasks

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages