MedRegA, an interpretable bilingual generalist model for diverse biomedical tasks, represented by its outstanding ability to leverage regional information. MedRegA can perceive 8 modalities covering almost all the body parts, showcasing significant versatility.
💡We establish Region-Centric tasks with a large-scale dataset, MedRegInstruct, where each sample is paired with coordinates of body structures or lesions.
💡Based on the proposed dataset, we develop a Region-Aware medical MLLM, MedRegA, as a bilingual generalist medical AI system to perform both image-level and region-level medical vision-language tasks, demonstrating impressive versatility.
- Release the model.
- Release the demo code.
- Release the evaluation code.
- Release the training code.
- Release the data.
Please refer to InternVL Installation to build the environment.
Run the demo:
torchrun --nproc-per-node=1 src/demo.py
We refer to the codes from InternVL. Thank the authors for releasing their code.