Skip to main content
Unpublished Paper
Modeling Relations and Their Mentions without Labeled Text
  • Sebastian Riedel
  • Limin Yao
  • Andrew McCallum, University of Massachusetts - Amherst
Several recent works on relation extraction have been applying the distant supervision paradigm: instead of relying on annotated text to learn how to predict relations, they employ existing knowledge bases (KBs) as source of supervision. Crucially, these approaches are trained based on the assumption that each sentence which mentions the two related entities is an expression of the given relation. Here we argue that this leads to noisy patterns that hurt precision, in particular if the knowledge base is not directly related to the text we are working with. We present a novel approach to distant supervision that can alleviate this problem based on the following two ideas: First, we use a factor graph to explicitly model the decision whether two entities are related, and the decision whether this relation is mentioned in a given sentence; second, we apply constraint-driven semi-supervision to train this model without any knowledge about which sentences express the relations in our training KB. We apply our approach to extract relations from the New York Times corpus and use Freebase as knowledge base. When compared to a state-of-the art approach for relation extraction under distant supervision, we achieve 31% error reduction.
Publication Date
This is the pre-published version harvested from CIIR.
Citation Information
Sebastian Riedel, Limin Yao and Andrew McCallum. "Modeling Relations and Their Mentions without Labeled Text" (2010)
Available at: