Learning Relation Entailment with Structured and Textual Information

Zhengbao JiangJun ArakiDonghan Yu, Ruohong Zhang, Wei Xu, Yiming YangGraham Neubig.

doi:10.24432/C5MS3J

TL;DR

Learning to predict relation entailment using both structured and textual information
Relations among words and entities are important for semantic understanding of text, but previous work has largely not considered relations between relations, or meta-relations. In this paper, we specifically examine relation entailment, where the existence of one relation can entail the existence of another relation. Relation entailment allows us to construct relation hierarchies, enabling applications in representation learning, question answering, relation extraction, and summarization. To this end, we formally define the new task of predicting relation entailment and construct a dataset by expanding the existing Wikidata relation hierarchy without expensive human intervention. We propose several methods that incorporate both structured and textual information to represent relations for this task. Experiments and analysis demonstrate that this task is challenging, and we provide insights into task characteristics that may form a basis for future work. The dataset and code have been released at https://github.com/jzbjyb/RelEnt.

Citation

@inproceedings{
jiang2020learning,
title={Learning Relation Entailment with Structured and Textual Information},
author={Zhengbao Jiang and Jun Araki and Donghan Yu and Ruohong Zhang and Wei Xu and Yiming Yang and Graham Neubig},
booktitle={Automated Knowledge Base Construction},
year={2020},
url={https://openreview.net/forum?id=ToTf_MX7Vn},
doi={10.24432/C5MS3J}
}

Sponsors