We introduce Medical Evidence Dependency(MD)-informed Self-Attention
, a Neuro-Symbolic Model for understanding free-text medical evidence in literature. We hypothesize this method can get the best of both: the high capacity of neural networks and the rigor, semantic clarity and reusability of symbolic logic.
- Citation: Kang, T., Turfah, A. Kim, J. Perotte, A. and Weng, C. (2021). A Neuro-Symbolic Method for Understanding Free-text Medical Evidence. Journal of the American Medical Informatics Association (in press)
- Contact: Tian Kang (tk2624@cumc.columbia.edu)
- Affiliation: Department of Biomedical Informatics, Columbia Univerisity (Dr. Chunhua Weng's lab)
MDAtt.py
generate Medical Evidence Dependency-information attention head.
MED_modeling.py
: modified from bert/modeling.py
(attention_layer, transformer and bert classes).
run_MDAttBert.py
: run BERT with Medical Evidence Dependency-information attention head.
Model We develop a symbolic compositional representation called Medical evidence Dependency
(MD) to represent the basic medical evidence entities and relations following the PICO framework widely adopted among clinicians for searching evidence. We use Transformer as the backbone and train one head in the Multi-Head Self-Attention to attend to MD
and to pass linguistic and domain knowledge onto later layers (MD-informed
). We integrated MD-informed Attention into BioBERT and evaluated it on two public MRC benchmarks for medical evidence from literature: i.e., Evidence Inference 2.0 and PubMedQA.
Medical Evidence Dependency (MD) and Proposition
Medical Evidence Dependency (MD) Matrix
Medical Evidence Dependency (MD)-informed Self Attention
Results The integration of MD-informed Attention
head improves BioBERT substantially for both benchmarks—as large as by +30% in the F1 score—and achieves the new state-of-the-art performance on the Evidence Inference 2.0. By visualizing the weights learned from MD-informed Attention head, we find the model can capture clinically meaningful relations separated by long passages of text.