Cited 0 times in 
Cited 0 times in 
Advancing Text-Driven Chest X-Ray Generation with Policy-Based Reinforcement Learning
| DC Field | Value | Language |
|---|---|---|
| dc.date.accessioned | 2025-07-09T08:38:59Z | - |
| dc.date.available | 2025-07-09T08:38:59Z | - |
| dc.date.issued | 2024-10 | - |
| dc.identifier.issn | 0302-9743 | - |
| dc.identifier.uri | https://ir.ymlib.yonsei.ac.kr/handle/22282913/206552 | - |
| dc.description.abstract | Recent advances in text-conditioned image generation diffusion models have begun paving the way for new opportunities in the modern medical domain, in particular, particularly in generating Chest X-rays (CXRs) from diagnostic reports. Nonetheless, to further drive the diffusion models to generate CXRs that faithfully reflect the complexity and diversity of real data, it has become evident that a nontrivial learning approach is needed. In light of this, we propose CXRL, a framework motivated by the potential of reinforcement learning (RL). Specifically, we integrate a policy gradient RL approach with well-designed multiple distinctive CXR-domain specific reward models. This approach guides the diffusion denoising trajectory, achieving precise CXR posture and pathological details. Here, considering the complex medical image environment, we present “RL with Comparative Feedback” (RLCF) for the reward mechanism, a human-like comparative evaluation that is known to be more effective and reliable in complex scenarios compared to direct evaluation. Our CXRL framework includes jointly optimizing learnable adaptive condition embeddings (ACE) and the image generator, enabling the model to produce more accurate and higher perceptual CXR quality. Our extensive evaluation of the MIMIC-CXR-JPG dataset demonstrates the effectiveness of our RL-based tuning approach. Consequently, our CXRL generates pathologically realistic CXRs, establishing a new standard for generating CXRs with high fidelity to real-world clinical scenarios. Project page: https://micv-yonsei.github.io/cxrl2024/. | - |
| dc.description.statementOfResponsibility | restriction | - |
| dc.language | English | - |
| dc.publisher | Springer | - |
| dc.relation.isPartOf | Lecture Notes in Computer Science | - |
| dc.rights | CC BY-NC-ND 2.0 KR | - |
| dc.title | Advancing Text-Driven Chest X-Ray Generation with Policy-Based Reinforcement Learning | - |
| dc.type | Article | - |
| dc.contributor.college | College of Medicine (의과대학) | - |
| dc.contributor.department | Others | - |
| dc.contributor.googleauthor | Woojung Han | - |
| dc.contributor.googleauthor | Chanyoung Kim | - |
| dc.contributor.googleauthor | Dayun Ju | - |
| dc.contributor.googleauthor | Yumin Shim & Seong Jae Hwang | - |
| dc.identifier.doi | 10.1007/978-3-031-72384-1_6 | - |
| dc.relation.journalcode | J02160 | - |
| dc.identifier.url | https://link.springer.com/chapter/10.1007/978-3-031-72384-1_6 | - |
| dc.citation.volume | 15003 | - |
| dc.citation.startPage | 56 | - |
| dc.citation.endPage | 66 | - |
| dc.identifier.bibliographicCitation | Lecture Notes in Computer Science, Vol.15003 : 56-66, 2024-10 | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.