多模态多智能体强化学习用于放射学报告生成:类似放射科医师的工作流程与临床可验证奖励

📄 中文摘要

提出了一种新颖的多模态多智能体强化学习框架MARL-Rad,用于放射学报告的生成。该框架协调区域特定的智能体和一个全局整合智能体,通过临床可验证的奖励进行优化。与以往的单模型强化学习或独立训练模型的后处理智能体化不同,该方法通过强化学习联合训练多个智能体,并优化整个智能体系统。在MIMIC-CXR和IU X-ray数据集上的实验表明,MARL-Rad在临床有效性(CE)指标上持续改善,如RadGraph、CheXbert和GREEN分数,达到了最先进的CE性能。进一步分析确认,MARL-Rad增强了侧别一致性,并生成了更高质量的报告。

📄 English Summary

Multi-Modal Multi-Agent Reinforcement Learning for Radiology Report Generation: Radiologist-Like Workflow with Clinically Verifiable Rewards

A novel multi-modal multi-agent reinforcement learning framework, MARL-Rad, is proposed for radiology report generation. This framework coordinates region-specific agents and a global integrating agent, optimized through clinically verifiable rewards. Unlike previous approaches that rely on single-model reinforcement learning or post-hoc agentization of independently trained models, this method jointly trains multiple agents and optimizes the entire agent system through reinforcement learning. Experiments on the MIMIC-CXR and IU X-ray datasets demonstrate that MARL-Rad consistently improves clinical efficacy (CE) metrics, such as RadGraph, CheXbert, and GREEN scores, achieving state-of-the-art CE performance. Further analyses confirm that MARL-Rad enhances laterality consistency and produces higher quality reports.

Powered by Cloudflare Workers + Payload CMS + Claude 3.5

数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等