期刊文献+

MKGViLT:visual-and-language transformer based on medical knowledge graph embedding

在线阅读 下载PDF
导出
摘要 Medical visual question answering(MedVQA)aims to enhance diagnostic confidence and deepen patientsunderstanding of their health conditions.While the Transformer architecture is widely used in multimodal fields,its application in MedVQA requires further enhancement.A critical limitation of contemporary MedVQA systems lies in the inability to integrate lifelong knowledge with specific patient data to generate human-like responses.Existing Transformer-based MedVQA models require enhancing their capabitities for interpreting answers through the applications of medical image knowledge.The introduction of the medical knowledge graph visual language transformer(MKGViLT),designed for joint medical knowledge graphs(KGs),addresses this challenge.MKGViLT incorporates an enhanced Transformer structure to effectively extract features and combine modalities for MedVQA tasks.The MKGViLT model delivers answers based on richer background knowledge,thereby enhancing performance.The efficacy of MKGViLT is evaluated using the SLAKE and P-VQA datasets.Experimental results show that MKGViLT surpasses the most advanced methods on the SLAKE dataset.
作者 CUI Wencheng SHI Wentao SHAO Hong 崔文成
出处 《High Technology Letters》 2025年第1期73-85,共13页 高技术通讯(英文版)
基金 Supported by the National Natural Science Foundation of China(No.62001313) the Liaoning Professional Talent Protect(No.XLYC2203046) the Shenyang Municipal Medical Engineering Cross Research Foundation of China(No.22-321-32-09).
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部