Recognition: unknown
CXRMate-2: Structured Multimodal Temporal Embeddings and Tractable Reinforcement Learning for Clinically Acceptable Chest X-ray Radiology Report Generation
Pith reviewed 2026-05-10 03:36 UTC · model grok-4.3
The pith
Chest X-ray report model reaches 45% acceptability in blinded radiologist ratings with no preference difference on seven of eight findings.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
CXRMate-2 achieves statistically significant gains over prior models on GREEN and RadGraph-XL across MIMIC-CXR, CheXpert Plus, and ReXgradient, then demonstrates in a randomised blinded review by three consultant radiologists that its reports on 120 studies were rated acceptable (preferred or equal) in 45% of cases, with no statistically significant preference difference for seven of the eight analysed findings. The model enables this outcome through structured multimodal temporal embeddings that support efficient GRPO reinforcement learning and a reward function aimed at semantic alignment with radiologist text. The authors conclude that higher recall of subtle findings remains the main gap
What carries the argument
Structured multimodal temporal embeddings with high-resolution visual feature compression that together allow efficient unified conditioning of an LLM decoder on visual, textual, and temporal context from a study and its prior, enabling tractable GRPO reinforcement learning.
If this is right
- Statistically significant metric gains on GREEN (11.2%) and RadGraph-XL (24.4%) over MedGemma 1.5 on MIMIC-CXR.
- Generated reports are consistently preferred for readability while radiologist reports are preferred for recall.
- The remaining barrier to non-inferiority is detection of subtle findings.
- The approach positions CXR RRG for prospective evaluation inside assistive, radiologist-led workflows.
Where Pith is reading between the lines
- If recall of subtle findings can be improved without losing readability, the model could move from 45% acceptability toward routine assistive use in high-volume reporting.
- The same temporal-embedding structure might transfer to other longitudinal imaging tasks such as CT follow-up where prior context matters.
- Real-world deployment would require monitoring whether radiologists adjust their own reporting style when assisted by the model.
Load-bearing premise
The blinded qualitative ratings by three consultant radiologists on 120 MIMIC-CXR studies represent what would be found in wider clinical practice and that the GRPO reward function improves true semantic alignment without hidden post-hoc adjustments.
What would settle it
A follow-up study with at least 500 diverse cases and more than three radiologists that finds acceptability below 30% or statistically significant preference gaps on four or more findings would refute the claim of a clear pathway to clinical acceptability.
Figures
read the original abstract
Chest X-ray (CXR) radiology report generation (RRG) models have shown rapid progress on automated metrics, yet their clinical utility remains uncertain due to limited qualitative evaluation by radiologists. We present CXRMate-2, a state-of-the-art CXR RRG model that enables tractable reinforcement learning (RL) through structured multimodal temporal embeddings and high-resolution visual feature compression, for efficient, unified conditioning of an LLM decoder on visual, textual, and temporal context from a study and its prior. This enables group relative policy optimisation (GRPO), where a proposed reward function is used to improve semantic alignment with radiologist reports. Across the MIMIC-CXR, CheXpert Plus, and ReXgradient datasets, CXRMate-2 achieves statistically significant improvements over strong benchmarks, including gains of 11.2% and 24.4% in GREEN and RadGraph-XL, respectively, on MIMIC-CXR relative to MedGemma 1.5 (4B). To directly compare CXRMate-2 against radiologist reporting, we conduct a blinded, randomised qualitative retrospective evaluation. Three consultant radiologists compare generated and radiologist reports across 120 studies from the MIMIC-CXR test set. Generated reports were deemed acceptable (defined as preferred or rated equally to radiologist reports) in 45% of ratings, with no statistically significant difference in preference rates for seven of the eight analysed findings. Preferences for radiologist reports were driven primarily by higher recall, while generated reports were consistently preferred for readability. Together, these results define a clear pathway to clinically acceptable CXR RRG. Improving recall and the detection of subtle findings represents the primary remaining barrier to non-inferiority with radiologist reporting, positioning CXR RRG for prospective evaluation in assistive, radiologist-led workflows.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript presents CXRMate-2, a chest X-ray radiology report generation model that uses structured multimodal temporal embeddings and high-resolution visual feature compression to enable tractable reinforcement learning via group relative policy optimization (GRPO) with a custom reward function for improved semantic alignment. It reports statistically significant gains on automated metrics including 11.2% in GREEN and 24.4% in RadGraph-XL on MIMIC-CXR relative to MedGemma 1.5 (4B), with similar results on CheXpert Plus and ReXgradient. A blinded randomized qualitative evaluation by three consultant radiologists on 120 MIMIC-CXR studies finds generated reports acceptable (preferred or equal) in 45% of ratings, with no statistically significant preference differences versus radiologist reports for seven of eight findings; radiologist reports are preferred for recall while generated reports score higher on readability. The work concludes that improving recall of subtle findings is the main remaining barrier to non-inferiority.
Significance. If the results hold, the paper is significant for directly linking metric improvements to a blinded radiologist preference study that quantifies clinical acceptability and isolates recall as the primary gap. The technical contributions around temporal embeddings and tractable GRPO provide a concrete mechanism for aligning generated reports with radiologist style, and the identification of readability advantages offers a practical insight for assistive deployment. These elements, combined with multi-dataset evaluation, strengthen the case for progressing CXR report generation toward prospective clinical testing.
major comments (2)
- [Qualitative evaluation (blinded retrospective study on 120 MIMIC-CXR cases)] The interpretation of no statistically significant difference in preference rates for seven of the eight findings as supporting clinical acceptability (45% acceptable rate) is load-bearing for the central claim yet rests on a sample of 120 studies rated by three radiologists without reported power analysis, equivalence testing (e.g., TOST), or confidence intervals on the preference proportions. Non-significance may reflect insufficient sensitivity rather than true parity, especially given the abstract's note that radiologist preference is driven by higher recall of subtle findings.
- [Methods (GRPO and reward function)] The reward function design for GRPO, including its precise formulation, how it avoids embedding data-driven fitting or post-hoc tuning, and its relation to the automated metrics, is insufficiently detailed. This is critical because the semantic alignment improvements and the 45% acceptability claim depend on this component; without explicit equations or ablation on reward components, reproducibility and assessment of selection effects in the 120-study subset are compromised.
minor comments (2)
- [Experimental setup] Clarify the exact data splits used for training, validation, and the 120-study qualitative subset, including any exclusion criteria or selection effects that might affect generalizability to broader clinical practice.
- [Abstract and results] The abstract states gains relative to MedGemma 1.5 (4B); providing parameter counts and architectural details for all baselines in a table would improve context for the reported metric improvements.
Simulated Author's Rebuttal
We thank the referee for their constructive and detailed feedback. We address each major comment below with point-by-point responses and indicate the revisions planned for the next manuscript version.
read point-by-point responses
-
Referee: [Qualitative evaluation (blinded retrospective study on 120 MIMIC-CXR cases)] The interpretation of no statistically significant difference in preference rates for seven of the eight findings as supporting clinical acceptability (45% acceptable rate) is load-bearing for the central claim yet rests on a sample of 120 studies rated by three radiologists without reported power analysis, equivalence testing (e.g., TOST), or confidence intervals on the preference proportions. Non-significance may reflect insufficient sensitivity rather than true parity, especially given the abstract's note that radiologist preference is driven by higher recall of subtle findings.
Authors: We acknowledge the referee's concern regarding the statistical interpretation of the qualitative results. The 120-study sample follows common practice in radiology AI reader studies, but we agree that confidence intervals and power considerations should be reported explicitly. In the revised manuscript we will add 95% confidence intervals for all preference proportions and include a post-hoc power discussion based on the observed effect sizes. While we did not perform TOST equivalence testing, the consistent pattern (no significant differences on seven of eight findings, with the sole driver of radiologist preference being recall of subtle findings) supports our conclusion that recall remains the primary barrier. We will expand the limitations section to qualify the acceptability claim and note the retrospective, exploratory nature of the evaluation. revision: partial
-
Referee: [Methods (GRPO and reward function)] The reward function design for GRPO, including its precise formulation, how it avoids embedding data-driven fitting or post-hoc tuning, and its relation to the automated metrics, is insufficiently detailed. This is critical because the semantic alignment improvements and the 45% acceptability claim depend on this component; without explicit equations or ablation on reward components, reproducibility and assessment of selection effects in the 120-study subset are compromised.
Authors: We thank the referee for identifying this gap in methodological detail. In the revised manuscript we will provide the complete mathematical formulation of the reward function, including each term (semantic alignment via RadGraph-XL and GREEN components, readability, and length penalties) and the fixed weighting scheme. The design uses clinically motivated, pre-defined metrics rather than any fitting or tuning on the test or qualitative-evaluation sets, thereby avoiding data-driven selection effects. We will also add ablation results isolating the contribution of each reward component to both automated metric gains and the qualitative acceptability rates. These additions will directly support reproducibility and allow readers to assess any influence on the 120-study subset. revision: yes
Circularity Check
No significant circularity; derivation and claims rest on independent evaluation
full rationale
The paper introduces structured embeddings and GRPO with a proposed reward function to align generated reports to radiologist references, then reports gains on GREEN and RadGraph-XL plus a separate blinded radiologist preference study on 120 MIMIC-CXR cases. No equation or step reduces to its own inputs by construction: the reward drives RL training but the primary clinical-acceptability claim (45% acceptable, no sig diff on 7/8 findings) is measured by external human raters rather than the automated metrics or reward itself. Automated metric gains are presented as empirical outcomes, not tautological predictions. Self-citations are absent from the provided text and the evaluation design supplies independent grounding.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
Bannur, Shruthi and Bouzid, Kenza and Castro, Daniel C. and Schwaighofer, Anton and Bond-Taylor, Sam and Ilse, Maximilian and Pérez-García, Fernando and Salvatelli, Valentina and Sharma, Harshita and Meissen, Felix and Ranjit, Mercy and Srivastav, Shaury and Gong, Julia and Falck, Fabian and Oktay, Ozan and Thieme, Anja and Lungren, Matthew P. and Wetsche...
-
[2]
Understanding and. RadioGraphics , author =. 2022 , pages =. doi:10.1148/rg.220037 , number =
-
[3]
Journal of Medical Imaging and Radiation Oncology , author =
Occupational burnout among radiographers, sonographers and radiologists in. Journal of Medical Imaging and Radiation Oncology , author =. 2017 , keywords =. doi:https://doi.org/10.1111/1754-9485.12547 , abstract =
-
[4]
Nicolson, Aaron and Zhuang, Shengyao and Dowling, Jason and Koopman, Bevan , editor =. The. Proceedings of the 63rd. 2025 , pages =. doi:10.18653/v1/2025.acl-long.9 , abstract =
-
[5]
Zhao, Weike and Wu, Chaoyi and Zhang, Xiaoman and Zhang, Ya and Wang, Yanfeng and Xie, Weidi , editor =. Proceedings of the 2024. 2024 , pages =. doi:10.18653/v1/2024.emnlp-main.836 , abstract =
-
[6]
Clinical. Radiology , author =. 2025 , pages =. doi:10.1148/radiol.250568 , abstract =
-
[7]
Korean J Radiol , author =
Artificial. Korean J Radiol , author =. 2025 , keywords =
2025
-
[8]
Value of. Radiology , author =. 2025 , pages =. doi:10.1148/radiol.241646 , abstract =
-
[9]
Generative. JAMA Network Open , author =. 2023 , pages =. doi:10.1001/jamanetworkopen.2023.36100 , abstract =
-
[10]
Johnson, Alistair and Bulgarelli, Lucas and Pollard, Tom and Celi, Leo Anthony and Mark, Roger and Horng, Steven , year =. doi:10.13026/5NTK-KM72 , abstract =
-
[11]
MIMIC-IV , a freely accessible electronic health record dataset,
Scientific Data , author =. 2023 , pages =. doi:10.1038/s41597-022-01899-x , abstract =
-
[12]
doi:10.71718/6NVZ-PM34 , abstract =
Chambon, Pierre and Delbrouck, Jean-Benoit and Sounack, Thomas and Huang, Shih-Cheng and Chen, Zhihong and Varma, Maya and Truong, Steven QH and Chuong, Chu The and Langlotz, Curtis P , year =. doi:10.71718/6NVZ-PM34 , abstract =
-
[13]
Qiao Jin, Bhuwan Dhingra, Zhengping Liu, William W Cohen, and Xinghua Lu
Scientific Data , author =. 2019 , pages =. doi:10.1038/s41597-019-0322-0 , abstract =
-
[14]
Zhang, Xiaoman and Acosta, Julian N and Miller, Josh and Huang, Ouwen and Rajpurkar, Pranav , year =
-
[15]
Hu, Edward and Shen, Yelong and Wallis, Phillip and Allen-Zhu, Zeyuan and Li, Yuanzhi and Wang, Shean and Wang, Lu and Chen, Weizhu , year =
-
[16]
Delbrouck, Jean-Benoit and Xu, Justin and Moll, Johannes and Thomas, Alois and Chen, Zhihong and Ostmeier, Sophie and Azhar, Asfandyar and Li, Kelvin Zhenghao and Johnston, Andrew and Bluethgen, Christian and Reis, Eduardo Pontes and Muneer, Mohamed S and Varma, Maya and Langlotz, Curtis , editor =. Automated. Proceedings of the 63rd. 2025 , pages =. doi:...
-
[17]
Flamingo: a visual language model for few-shot learning , isbn =
Alayrac, Jean-Baptiste and Donahue, Jeff and Luc, Pauline and Miech, Antoine and Barr, Iain and Hasson, Yana and Lenc, Karel and Mensch, Arthur and Millicah, Katie and Reynolds, Malcolm and Ring, Roman and Rutherford, Eliza and Cabi, Serkan and Han, Tengda and Gong, Zhitao and Samangooei, Sina and Monteiro, Marianne and Menick, Jacob and Borgeaud, Sebasti...
-
[18]
Gemini 2.5:
Comanici, Gheorghe and Bieber, Eric and Schaekermann, Mike and Pasupat, Ice and Sachdeva, Noveen and Dhillon, Inderjit and Blistein, Marcel and. Gemini 2.5:. 2025 , note =
2025
-
[19]
Miura, Yasuhide and Zhang, Yuhao and Tsai, Emily and Langlotz, Curtis and Jurafsky, Dan , editor =. Improving. Proceedings of the 2021. 2021 , pages =. doi:10.18653/v1/2021.naacl-main.416 , abstract =
-
[20]
Xu Chu, Xinrong Chen, Guanyu Wang, Zhijie Tan, Kui Huang, Wenyu Lv, Tong Mo, and Weiping Li
Delbrouck, Jean-Benoit and Chambon, Pierre and Bluethgen, Christian and Tsai, Emily and Almusa, Omar and Langlotz, Curtis , editor =. Improving the. Findings of the. 2022 , pages =. doi:10.18653/v1/2022.findings-emnlp.319 , abstract =
-
[21]
Proceedings of the
Jain, Saahil and Agrawal, Ashwin and Saporta, Adriel and Truong, Steven and Duong, Du Nguyen Duong Nguyen and Bui, Tan and Chambon, Pierre and Zhang, Yuhao and Lungren, Matthew and Ng, Andrew and Langlotz, Curtis and Rajpurkar, Pranav and Rajpurkar, Pranav , editor =. Proceedings of the
-
[22]
Tanno, Ryutaro and Barrett, David G. T. and Sellergren, Andrew and Ghaisas, Sumedh and Dathathri, Sumanth and See, Abigail and Welbl, Johannes and Lau, Charles and Tu, Tao and Azizi, Shekoofeh and Singhal, Karan and Schaekermann, Mike and May, Rhys and Lee, Roy and Man, SiWai and Mahdavi, Sara and Ahmed, Zahra and Matias, Yossi and Barral, Joelle and Esla...
-
[23]
arXiv preprint arXiv:2311.13668 , year=
Hyland, Stephanie L. and Bannur, Shruthi and Bouzid, Kenza and Castro, Daniel C. and Ranjit, Mercy and Schwaighofer, Anton and Pérez-García, Fernando and Salvatelli, Valentina and Srivastav, Shaury and Thieme, Anja and Codella, Noel and Lungren, Matthew P. and Wetscherek, Maria Teodora and Oktay, Ozan and Alvarez-Valle, Javier , month = apr, year =. doi:1...
-
[24]
Aldhafeeri, Faten M. , month = jan, year =. Governing. Diagnostics , publisher =. doi:10.3390/diagnostics15182300 , abstract =
-
[25]
British Journal of Radiology , author =. 2023 , pages =. doi:10.1259/bjr.20230023 , abstract =
-
[26]
Implementing. Mayo Clinic Proceedings: Digital Health , author =. 2025 , pages =. doi:10.1016/j.mcpdig.2024.100188 , abstract =
-
[27]
Dratsch, Thomas and Chen, Xue and Rezazade Mehrizi, Mohammad and Kloeckner, Roman and Mähringer-Kunz, Aline and Püsken, Michael and Baeßler, Bettina and Sauer, Stephanie and Maintz, David and Pinto dos Santos, Daniel , month = may, year =. Automation. Radiology , publisher =. doi:10.1148/radiol.222176 , abstract =
-
[28]
Journal of Safety Science and Resilience , author =
Exploring the risks of automation bias in healthcare artificial intelligence applications:. Journal of Safety Science and Resilience , author =. 2024 , keywords =. doi:10.1016/j.jnlssr.2024.06.001 , abstract =
-
[29]
Korean Journal of Radiology , author =
Real-. Korean Journal of Radiology , author =. 2025 , pages =. doi:10.3348/kjr.2025.0962 , abstract =
-
[30]
How to. RadioGraphics , author =. 2020 , pages =. doi:10.1148/rg.2020200020 , language =
-
[31]
Nature Machine Intelligence , author =
Exploring scalable medical image encoders beyond text supervision , volume =. Nature Machine Intelligence , author =. 2025 , pages =. doi:10.1038/s42256-024-00965-w , language =
-
[32]
Journal of the Royal Society of Medicine , author =
Lateral. Journal of the Royal Society of Medicine , author =. 2005 , pages =. doi:10.1177/014107680509800705 , language =
-
[33]
Informatics in Medicine Unlocked , author =
Longitudinal data and a semantic similarity reward for chest. Informatics in Medicine Unlocked , author =. 2024 , pages =. doi:10.1016/j.imu.2024.101585 , language =
-
[34]
Kevin and Xiao, Li , editor =
Wu, Xian and Yang, Shuxin and Qiu, Zhaopeng and Ge, Shen and Yan, Yangtian and Wu, Xingwang and Zheng, Yefeng and Zhou, S. Kevin and Xiao, Li , editor =. Proceedings of the 29th. 2022 , pages =
2022
-
[35]
, volume =
The chest radiograph. , volume =. The Ulster medical journal , author =. 2012 , keywords =
2012
-
[36]
Bannur, Shruthi and Hyland, Stephanie and Liu, Qianchu and Pérez-García, Fernando and Ilse, Maximilian and Castro, Daniel C. and Boecking, Benedikt and Sharma, Harshita and Bouzid, Kenza and Thieme, Anja and Schwaighofer, Anton and Wetscherek, Maria and Lungren, Matthew P. and Nori, Aditya and Alvarez-Valle, Javier and Oktay, Ozan , month = jun, year =. L...
-
[37]
Liu, Shuai and Cho, Hyundong and Freedman, Marjorie and Ma, Xuezhe and May, Jonathan , editor =. Proceedings of the 61st. 2023 , pages =. doi:10.18653/v1/2023.acl-long.468 , abstract =
-
[38]
Pragmatic
Nguyen, Dang and Chen, Chacha and He, He and Tan, Chenhao , month = dec, year =. Pragmatic. Proceedings of the 3rd
-
[39]
Qin, Zhen and Han, Xiaodong and Sun, Weixuan and Li, Dongxu and Kong, Lingpeng and Barnes, Nick and Zhong, Yiran , editor =. The. Proceedings of the 2022. 2022 , pages =. doi:10.18653/v1/2022.emnlp-main.473 , abstract =
-
[40]
Nicolson, Aaron and Liu, Jinghui and Dowling, Jason and Nguyen, Anthony and Koopman, Bevan , editor =. e-. Proceedings of the 23rd. 2024 , pages =. doi:10.18653/v1/2024.bionlp-1.8 , abstract =
-
[41]
Clinically
Liu, Guanxiong and Hsu, Tzu-Ming Harry and McDermott, Matthew and Boag, Willie and Weng, Wei-Hung and Szolovits, Peter and Ghassemi, Marzyeh , month = oct, year =. Clinically. Proceedings of the 4th
-
[42]
Sellergren, Andrew and Kazemzadeh, Sahar and Jaroensri, Tiam and Kiraly, Atilla and Traverse, Madeleine and Kohlberger, Timo and Xu, Shawn and Jamil, Fayaz and Hughes, Cían and Lau, Charles and Chen, Justin and Mahvar, Fereshteh and Yatziv, Liron and Chen, Tiffany and Sterling, Bram and Baby, Stefanie Anna and Baby, Susanna Maria and Lai, Jeremy and Schmi...
work page internal anchor Pith review Pith/arXiv arXiv doi:10.48550/arxiv.2507.05201
-
[43]
and Marcheret, Etienne and Mroueh, Youssef and Ross, Jerret and Goel, Vaibhava , month = jul, year =
Rennie, Steven J. and Marcheret, Etienne and Mroueh, Youssef and Ross, Jerret and Goel, Vaibhava , month = jul, year =. Self-. 2017. doi:10.1109/CVPR.2017.131 , abstract =
-
[44]
Shao, Zhihong and Wang, Peiyi and Zhu, Qihao and Xu, Runxin and Song, Junxiao and Bi, Xiao and Zhang, Haowei and Zhang, Mingchuan and Li, Y. K. and Wu, Y. and Guo, Daya , month = apr, year =. doi:10.48550/arXiv.2402.03300 , abstract =
work page internal anchor Pith review Pith/arXiv arXiv doi:10.48550/arxiv.2402.03300
-
[45]
https://doi.org/10.1016/j.neucom.2023.127063
Neurocomput. , author =. doi:10.1016/j.neucom.2023.127063 , number =
-
[46]
Papineni, Kishore and Roukos, Salim and Ward, Todd and Zhu, Wei-Jing , editor =. Bleu: a. Proceedings of the 40th. 2002 , pages =. doi:10.3115/1073083.1073135 , urldate =
-
[47]
Lin, Chin-Yew and Och, Franz Josef , month = jul, year =. Automatic. Proceedings of the 42nd. doi:10.3115/1218955.1219032 , urldate =
-
[48]
and Artzi, Yoav , month = sep, year =
Zhang, Tianyi and Kishore, Varsha and Wu, Felix and Weinberger, Kilian Q. and Artzi, Yoav , month = sep, year =
-
[49]
Boecking, Benedikt and Usuyama, Naoto and Bannur, Shruthi and Castro, Daniel C. and Schwaighofer, Anton and Hyland, Stephanie and Wetscherek, Maria and Naumann, Tristan and Nori, Aditya and Alvarez-Valle, Javier and Poon, Hoifung and Oktay, Ozan , editor =. Making the. Computer. 2022 , note =. doi:10.1007/978-3-031-20059-5_1 , language =
-
[50]
Xu, Justin and Zhang, Xi and Abderezaei, Javid and Bauml, Julie and Boodoo, Roger and Haghighi, Fatemeh and Ganjizadeh, Ali and Brattain, Eric and Van Veen, Dave and Meng, Zaiqiao and Eyre, David W and Delbrouck, Jean-Benoit , editor =. Proceedings of the 2025. 2025 , pages =. doi:10.18653/v1/2025.emnlp-demos.40 , abstract =
-
[51]
Smit, Akshay and Jain, Saahil and Rajpurkar, Pranav and Pareek, Anuj and Ng, Andrew and Lungren, Matthew , year =. Combining. Proceedings of the 2020. doi:10.18653/v1/2020.emnlp-main.117 , abstract =
-
[52]
Delbrouck, Jean-Benoit and Chambon, Pierre and Chen, Zhihong and Varma, Maya and Johnston, Andrew and Blankemeier, Louis and Van Veen, Dave and Bui, Tan and Truong, Steven and Langlotz, Curtis , editor =. Findings of the. 2024 , pages =. doi:10.18653/v1/2024.findings-acl.765 , abstract =
-
[53]
Ostmeier, Sophie and Xu, Justin and Chen, Zhihong and Varma, Maya and Blankemeier, Louis and Bluethgen, Christian and Michalson, Arne Edward and Moseley, Michael and Langlotz, Curtis and Chaudhari, Akshay S and Delbrouck, Jean-Benoit , editor =. Findings of the. 2024 , pages =. doi:10.18653/v1/2024.findings-emnlp.21 , abstract =
-
[54]
and Rajpurkar, Pranav , month = mar, year =
Zhou, Hong-Yu and Acosta, Julián Nicolás and Adithan, Subathra and Datta, Suvrankar and Topol, Eric J. and Rajpurkar, Pranav , month = mar, year =. NEJM AI , publisher =. doi:10.1056/AIoa2500595 , number =
-
[55]
Zhang, Xi and Meng, Zaiqiao and Lever, Jake and Ho, Edmond S. L. , editor =. Libra:. Findings of the. 2025 , pages =. doi:10.18653/v1/2025.findings-acl.888 , abstract =
-
[56]
Lin, Qika and Zhu, Yifan and Pu, Bin and Huang, Ling and Luo, Haoran and Ma, Jingying and Wu, Feng and He, Kai and Xu, Jiaxing and Peng, Zhen and Zhao, Tianzhe and Xu, Fangzhi and Zhang, Jian and Ou, Zhonghong and Cambria, Erik and Mishra, Swapnil and Feng, Mengling , month = mar, year =. Toward. Proceedings of the
-
[57]
What’s in the image? a deep-dive into the vision of vision language models
Liu, Kang and Ma, Zhuoqi and Kang, Xiaolu and Li, Yunan and Xie, Kun and Jiao, Zhicheng and Miao, Qiguang , month = jun, year =. Enhanced. 2025. doi:10.1109/CVPR52734.2025.00968 , urldate =
-
[58]
Lin, Qika and Zhu, Yifan and Pu, Bin and Huang, Ling and Luo, Haoran and Ma, Jingying and Wu, Feng and He, Kai and Xu, Jiaxing and Peng, Zhen and Zhao, Tianzhe and Xu, Fangzhi and Zhang, Jian and Ou, Zhonghong and Cambria, Erik and Mishra, Swapnil and Feng, Mengling , month = mar, year =. Toward. doi:10.48550/arXiv.2509.03906 , abstract =
-
[59]
Gaussian Error Linear Units (GELUs)
Hendrycks, Dan and Gimpel, Kevin , month = jun, year =. Gaussian. doi:10.48550/arXiv.1606.08415 , abstract =
-
[60]
Decoupled
Loshchilov, Ilya and Hutter, Frank , year =. Decoupled. International
-
[61]
Dasegowda, Giridhar and Kalra, Mannudeep K. and Abi-Ghanem, Alain S. and Arru, Chiara D. and Bernardo, Monica and Saba, Luca and Segota, Doris and Tabrizi, Zhale and Viswamitra, Sanjaya and Kaviani, Parisa and Karout, Lina and Dreyer, Keith J. , month = jan, year =. Suboptimal. Diagnostics , publisher =. doi:10.3390/diagnostics13030412 , abstract =
-
[62]
Schroeder, Will and Martin, Ken and Lorensen, Bill , year =. The
-
[63]
Evaluating progress in automatic chest. Patterns , author =. 2023 , keywords =. doi:10.1016/j.patter.2023.100802 , abstract =
-
[64]
Reporting of. JAMA , author =. 2012 , pages =. doi:10.1001/jama.2012.87802 , abstract =
-
[65]
InThe Thirty-ninth Annual Conference on Neural Information Processing Systems
Chen, Zhihong and Varma, Maya and Delbrouck, Jean-Benoit and Paschali, Magdalini and Blankemeier, Louis and Veen, Dave Van and Valanarasu, Jeya Maria Jose and Youssef, Alaa and Cohen, Joseph Paul and Reis, Eduardo Pontes and Tsai, Emily B. and Johnston, Andrew and Olsen, Cameron and Abraham, Tanishq Mathew and Gatidis, Sergios and Chaudhari, Akshay S. and...
-
[66]
doi:10.13026/4jqj-jw95 , journal =
Johnson, Alistair and Pollard, Tom and Mark, Roger and Berkowitz, Seth and Horng, Steven , month = jul, year =. doi:10.13026/4jqj-jw95 , journal =
-
[67]
Liu, Kang and Ma, Zhuoqi and Fang, Zikang and Li, Yunan and Xie, Kun and Miao, Qiguang , month = mar, year =. Proceedings of the. doi:10.1609/aaai.v40i9.37657 , number =
-
[68]
doi:10.13026/jsn5-t979 , journal =
Johnson, Alistair and Lungren, Matthew and Peng, Yifan and Lu, Zhiyong and Mark, Roger and Berkowitz, Seth and Horng, Steven , month = mar, year =. doi:10.13026/jsn5-t979 , journal =
-
[69]
Artificial Intelligence in Medicine , author =
Improving chest. Artificial Intelligence in Medicine , author =. 2023 , keywords =. doi:https://doi.org/10.1016/j.artmed.2023.102633 , abstract =
-
[70]
Sellergren, Andrew and Gao, Chufan and Mahvar, Fereshteh and Kohlberger, Timo and Jamil, Fayaz and Traverse, Madeleine and Tono, Alberto and Sadjad, Bashir and Yang, Lin and Lau, Charles and Yatziv, Liron and Chen, Tiffany and Sterling, Bram and Philbrick, Kenneth and Tiwari, Richa and Liu, Yun and Jajoo, Madhuram and Sankarapu, Chandrashekar and Vispute,...
work page internal anchor Pith review Pith/arXiv arXiv doi:10.48550/arxiv.2604.05081
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.