Multi-Modal Knowledge-Aware Attention Network for Question Answering

被引:0
|
作者
Zhang Y. [1 ,2 ]
Qian S. [2 ]
Fang Q. [2 ]
Xu C. [1 ,2 ]
机构
[1] University of Chinese Academy of Sciences, Beijing
[2] National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences, Beijing
来源
Jisuanji Yanjiu yu Fazhan/Computer Research and Development | 2020年 / 57卷 / 05期
基金
中国国家自然科学基金;
关键词
Attention; Deep learning; Information retrieval; Medical question answering system; Multi-modal knowledge graph;
D O I
10.7544/issn1000-1239.2020.20190474
中图分类号
学科分类号
摘要
With the popularity of the Internet, more people choose to search online to find the solutions when they feel sick. With the emergence of reliable medical question answering websites, e.g. Chunyu Doctor, XYWY, patients can communicate with the doctor one-one at home. However, existing question answering methods focus on word-level interaction or semantics, but rarely notice the hidden rationale with doctors' commonsense, while in the real scenes, doctors need to acquire plenty of domain knowledge to give advice to the patients. This paper proposes a novel multi-modal knowledge-aware attention network (MKAN) to effectively exploit multi-modal knowledge graph for medical question answering. The incorporation of multi-modal information can provide more fine-grained information. This information shows how entities in the medical graph are related. Our model first generates multi-modal entity representation with a translation-based method, and then defines question-answer interactions as the paths in the multi-modal knowledge graph that connect the entities in the question and answer. Furthermore, to discriminate the importance of paths, we propose an attention network. We build a large-scale multi-modal medical knowledge graph based on Symptom-in-Chinese, as well as one real-world medical question answering datasets based on Chunyu Doctor website. Extensive experiments strongly evidence that our proposed model obtains significant performance compared with state-of-the arts. © 2020, Science Press. All right reserved.
引用
收藏
页码:1037 / 1045
页数:8
相关论文
共 30 条
  • [1] Tay Y., Tuan L.A., Hui S.C., Cross temporal recurrent networks for ranking question answer pairs, Proc of the 32nd AAAI Conf on Artificial Intelligence, pp. 5512-5519, (2018)
  • [2] Tay Y., Phan M.C., Tuan L.A., Et al., Learning to rank question answer pairs with holographic dual LSTM architecture, Proc of the 40th Int Conf ACM SIGIR on Research and Development in Information Retrieval, pp. 695-704, (2017)
  • [3] Peng L., Lan Y., Guo J., Et al., Text matching as image recognition, Proc of the 30th AAAI Conf on Artificial Intelligence, pp. 2793-2799, (2016)
  • [4] Xiong C., Dai Z., Callan J., Et al., End-to-end neural ad-hoc ranking with kernel pooling, Proc of the 40th ACM SIGIR Int Conf on Research and Development in Information Retrieval, pp. 55-64, (2017)
  • [5] Choi E., Bahadori M.T., Song L., Et al., GRAM: Graph-based attention model for healthcare representation learning, Proc of the 23rd ACM SIGKDD Int Conf on Knowledge Discovery and Data Mining, pp. 787-795, (2017)
  • [6] Goodwin T.R., Harabagiu S.M., Knowledge representations and inference techniques for medical question answering, ACM Transactions on Intelligent Systems and Technology, 9, 2, (2017)
  • [7] Ma F., You Q., Xiao H., Et al., KAME: Knowledge-based attention model for diagnosis prediction in healthcare, Proc of the 27th ACM Int Conf on Information and Knowledge Management, pp. 743-752, (2018)
  • [8] Shen Y., Deng Y., Yang M., Et al., Knowledge-aware attentive neural network for ranking question answer pairs, Proc of the 41st Int Conf ACM SIGIR on Research Development in Information Retrieval, pp. 901-904, (2018)
  • [9] Nie L., Wang M., Zhang L., Et al., Disease inference from health-related questions via sparse deep learning, IEEE Transactions on Knowledge and Data Engineering, 27, 8, pp. 2107-2119, (2015)
  • [10] Guo D., Li M., Yu Y., Et al., Disease inference with symptom extraction and bidirectional recurrent neural network, Proc of the 12th IEEE Int Conf on Bioinformatics and Biomedicine, pp. 864-868, (2018)