Skip to content

uni-medical/GMAI-VL-R1

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

3 Commits
Β 
Β 

Repository files navigation

GMAI-VL-R1

πŸ”¬ Official codebase for the paper:
GMAI-VL-R1: Harnessing Reinforcement Learning for Multimodal Medical Reasoning
arXiv: 2504.01886 Β· Yanzhou Su, Tianbin Li, Jiyao Liu, Chenglong Ma, Junzhi Ning, Cheng Tang, Sibo Ju, Jin Ye, Pengcheng Chen, Ming Hu, et al.


πŸ“˜ Introduction

This repository contains the implementation of GMAI-VL-R1, a framework for multimodal medical reasoning powered by reinforcement learning.

We aim to explore how large vision-language models (VLMs) can improve medical visual understanding and reasoning with rule-based reward design and high-quality medical datasets.

Our method incorporates recent advancements in reinforcement learning for vision-language models and adapts them to the challenging domain of medical imaging.


🧠 Built Upon

Our work builds upon and integrates ideas from several open-source projects:

We thank the authors of these projects for their valuable contributions to the community.


πŸ“Š Dataset

We use a high-quality medical reasoning dataset:

πŸ“ Dataset Name: GMAI-Reasoning10K
πŸ“„ Format: JSONL
πŸ“· Modalities: X-ray, CT, MRI, OCT, Ultrasound, etc.
πŸ”’ Samples: 10,000 high-quality multiple-choice questions designed for reinforcement learning training
πŸ“š Source: Aggregated from 95 public medical datasets (e.g., Kaggle, GrandChallenge)


πŸ“œ Citation

If you find our work helpful, please consider citing our paper:

@article{su2025gmai,
  title={Gmai-vl-r1: Harnessing reinforcement learning for multimodal medical reasoning},
  author={Su, Yanzhou and Li, Tianbin and Liu, Jiyao and Ma, Chenglong and Ning, Junzhi and Tang, Cheng and Ju, Sibo and Ye, Jin and Chen, Pengcheng and Hu, Ming and others},
  journal={arXiv preprint arXiv:2504.01886},
  year={2025}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published