Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently,...

9
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, pages 3821–3829, Hong Kong, China, November 3–7, 2019. c 2019 Association for Computational Linguistics 3821 Improving Distantly-Supervised Relation Extraction with Joint Label Embedding Linmei Hu 1 , Luhao Zhang 1 , Chuan Shi 1 * , Liqiang Nie 2 , Weili Guan 3 , Cheng Yang 1 1 Beijing University of Posts and Telecommunications, China 2 Shan Dong University, China 3 Hewlett Packard Enterprise Singapore, Singapore {hulinmei,zhangluhao,shichuan}@bupt.edu.cn {nieliqiang, honeyguan, albertyang33}@gmail.com Abstract Distantly-supervised relation extraction has proven to be effective to find relational facts from texts. However, the existing approaches treat labels as independent and meaningless one-hot vectors, which cause a loss of poten- tial label information for selecting valid in- stances. In this paper, we propose a novel multi-layer attention-based model to improve relation extraction with joint label embedding. The model makes full use of both structural information from Knowledge Graphs and tex- tual information from entity descriptions to learn label embeddings through gating integra- tion, while avoiding the imposed noise with an attention mechanism. Then the learned label embeddings are used as another atten- tion over the instances (whose embeddings are also enhanced with the entity descriptions) for improving relation extraction. Extensive ex- periments demonstrate that our model signifi- cantly outperforms state-of-the-art methods. 1 Introduction Knowledge Graphs (KGs) such as Freebase and DBpedia have shown their strong power in many natural language processing tasks including ques- tion answering (Zhang et al., 2018) and dialog generation (Zhou et al., 2018). However, these KGs are far from complete. Relation extraction, which aims to fill this gap by extracting semantic relationships between entity pairs from plain texts, is thus of great importance. Most existing supervised relation extraction methods require a large number of labeled train- ing data, which is time-consuming and laborious. Distant supervision has been proposed by (Mintz et al., 2009) to address the challenge. It assumes that if two entities have a relation in KGs, then all sentences mentioning the two entities express * Corresponding author: Chuan Shi. Romania Its capital and largest city is Bucharest include Cluj-1DSRFDォ Knowledge Graph Entity Description Entity Description A bag of sentences Method (2) (a) Existing Methods (b) Our Proposed Method Label Embedding Classifier Att Att Classifier Method (3) Knowledge Graph A bag of sentences Att Textual Relation Embedding Classifier Method (1) Att Textual Relation Embedding Figure 1: Illustration of comparison of existing meth- ods and our proposed method. this relation. Thus, distant supervision can auto- matically generate a large number of labeled data without labor cost. Simultaneously, it often suf- fers from wrong labeling problem (Surdeanu et al., 2012; Zeng et al., 2015). Recently, significant progress has been made in applying deep neural networks for relation extrac- tion under distant supervision (Zeng et al., 2014, 2015; Feng et al., 2017). To alleviate the wrong labeling problem in distant supervision, attention models have been proposed to select valid in- stances (Ji et al., 2017). As shown in Figure 1 (a), they can be divided into three categories: (1) typ- ical attention models without external information (Lin et al., 2016; Du et al., 2018), (2) attention models using KGs (Han et al., 2018a), (3) atten- tion models using side information such as entity descriptions (Vashishth et al., 2018; Ji et al., 2017). However, they all have flaws in selecting valid in- stances due to failing to exploit potential label in- formation. They treat labels as independent and meaningless one-hot vectors, which cause a loss

Transcript of Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently,...

Page 1: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processingand the 9th International Joint Conference on Natural Language Processing, pages 3821–3829,Hong Kong, China, November 3–7, 2019. c©2019 Association for Computational Linguistics

3821

Improving Distantly-Supervised Relation Extraction withJoint Label Embedding

Linmei Hu1, Luhao Zhang1, Chuan Shi1 ∗, Liqiang Nie2, Weili Guan3, Cheng Yang1

1Beijing University of Posts and Telecommunications, China2Shan Dong University, China

3Hewlett Packard Enterprise Singapore, Singapore{hulinmei,zhangluhao,shichuan}@bupt.edu.cn

{nieliqiang, honeyguan, albertyang33}@gmail.com

Abstract

Distantly-supervised relation extraction hasproven to be effective to find relational factsfrom texts. However, the existing approachestreat labels as independent and meaninglessone-hot vectors, which cause a loss of poten-tial label information for selecting valid in-stances. In this paper, we propose a novelmulti-layer attention-based model to improverelation extraction with joint label embedding.The model makes full use of both structuralinformation from Knowledge Graphs and tex-tual information from entity descriptions tolearn label embeddings through gating integra-tion, while avoiding the imposed noise withan attention mechanism. Then the learnedlabel embeddings are used as another atten-tion over the instances (whose embeddings arealso enhanced with the entity descriptions) forimproving relation extraction. Extensive ex-periments demonstrate that our model signifi-cantly outperforms state-of-the-art methods.

1 Introduction

Knowledge Graphs (KGs) such as Freebase andDBpedia have shown their strong power in manynatural language processing tasks including ques-tion answering (Zhang et al., 2018) and dialoggeneration (Zhou et al., 2018). However, theseKGs are far from complete. Relation extraction,which aims to fill this gap by extracting semanticrelationships between entity pairs from plain texts,is thus of great importance.

Most existing supervised relation extractionmethods require a large number of labeled train-ing data, which is time-consuming and laborious.Distant supervision has been proposed by (Mintzet al., 2009) to address the challenge. It assumesthat if two entities have a relation in KGs, thenall sentences mentioning the two entities express

∗Corresponding author: Chuan Shi.

Romania … Its capital and

largest city is Bucharest …

include Cluj-Napoca…

Knowledge Graph

Entity Description

Entity Description

A bag of sentences

Method (2)

(a) Existing Methods

(b) Our Proposed Method

Label Embedding

Classifier

Att

Att

Classifier

Method (3)

Knowledge GraphA bag of sentences

Att

Textual Relation

Embedding

Classifier

Method (1)

Att

Textual Relation

Embedding

Figure 1: Illustration of comparison of existing meth-ods and our proposed method.

this relation. Thus, distant supervision can auto-matically generate a large number of labeled datawithout labor cost. Simultaneously, it often suf-fers from wrong labeling problem (Surdeanu et al.,2012; Zeng et al., 2015).

Recently, significant progress has been made inapplying deep neural networks for relation extrac-tion under distant supervision (Zeng et al., 2014,2015; Feng et al., 2017). To alleviate the wronglabeling problem in distant supervision, attentionmodels have been proposed to select valid in-stances (Ji et al., 2017). As shown in Figure 1 (a),they can be divided into three categories: (1) typ-ical attention models without external information(Lin et al., 2016; Du et al., 2018), (2) attentionmodels using KGs (Han et al., 2018a), (3) atten-tion models using side information such as entitydescriptions (Vashishth et al., 2018; Ji et al., 2017).However, they all have flaws in selecting valid in-stances due to failing to exploit potential label in-formation. They treat labels as independent andmeaningless one-hot vectors, which cause a loss

Page 2: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3822

of potential label information. Label embeddingsaim to learn representations of labels based on re-lated information. The label embeddings can beused to attend over the bag of instances for relationclassification. Additionally, they don’t take advan-tage of both structural information from KGs andtextual information from entity descriptions andignore the imposed noise.

In this paper, we propose a novel multi-layerattention-based model RELE (Relation Extractionwith Joint Label Embedding) to improve rela-tion extraction. Our model integrates both struc-tural information from KGs and textual informa-tion from entity descriptions with a gating mech-anism to learn label embeddings, while avoidingthe imposed noise (highlighted in green) in entitydescriptions with an attention mechanism. Thenthe label embeddings are used as another anten-tion over the bag of instances to select valid onesfor improving relation extraction. Note that wealso enhance the instance embedding with entitydescriptions. The contributions of this paper canbe summarized as follows:

• We propose a novel multi-layer attention-based model RELE to improve distantly su-pervised relation extraction with joint labelembedding. The label embeddings can beused to attend over the bag of instances forrelation classification.

• RELE makes full use of both structural infor-mation from KGs and textual information ofentity descriptions to learn label embeddingsthrough gating integration, while avoidingthe imposed noise with attention.

• Extensive experiments on two benchmarkdatasets have demonstrated that our modelsignificantly outperforms state-of-the-artmethods on distantly-supervised relationextraction.

2 Related Work

Our work is mainly related to distant supervision,neural relation extraction and label embedding.

Distant Supervision. Most supervised rela-tion extraction methods require large-scale labeledtraining data which are expensive. Distant Super-vision proposed by (Mintz et al., 2009) is an ef-fective method to automatically label large-scaletraining data under the assumption that if two

entities have a relation in a KG, then all sen-tences mentioning those entities express this rela-tion. The assumption does not work in all casesand causes the mislabeling problem.

MultiR (Hoffmann et al., 2011) and MIMLRE(Surdeanu et al., 2012) introduce multi-instancelearning where the instances mentioning the sameentity pair are processed at a bag level. How-ever, these methods rely heavily on handcraftedfeatures.

Neural Relation Extraction. With the devel-opment of deep learning, neural networks haveproven to be efficient to automatically extract validfeatures from sentences in recent years. Some re-searches (Zeng et al., 2014, 2015) adopt Convo-lution Neural Networks (CNN) to learn sentencerepresentations automatically. To alleviate themislabeling problem, attention mechanisms (Linet al., 2016; Du et al., 2018) have been employed.Apart from that, some studies apply other relevantinformation to improve relation extraction (Zenget al., 2017; Vashishth et al., 2018; Han et al.,2018b,a; Ji et al., 2017). For example, RESIDE(Vashishth et al., 2018) utilizes the available sideinformation from knowledge bases, including en-tity types and relation alias information. Han et al.(2018a) proposed a joint representation learningframework of KGs and instances, which lever-ages the KG embeddings to select valid instances.APCNN+D (Ji et al., 2017) exploits the entity de-scriptions as background knowledge for selectionof valid instances, and ignores the imposed noise.

Different from the existing works, we proposea novel multi-layer attention-based model RELEwith joint label embedding. Our model makesfull use of both structural information from KGsand textual information from entity descriptions tolearn label embeddings through gating integration,while avoiding the imposed noise with an atten-tion mechanism. The label embeddings are thenused as another attention over the bag of instancesto select valid instances for relation classification.

Label Embedding. Label embedding has beenwidely exploited in computer vision including im-age classification (Akata et al., 2016) and textrecognition (Rodriguez-Serrano et al., 2015). Re-cently, LEAM (Wang et al., 2018) successfully ap-plies label embedding in text classification, whichobtains each label’s embedding by its correspond-ing text descriptions. In this work, we are the firstto apply it for relation extraction. We propose

Page 3: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3823

a novel multi-layer attention-based model to im-prove relation extraction with joint label embed-ding.

3 Preliminaries

In this section, we briefly introduce some nota-tions and concepts used in this paper.

For convenience, we denote a KG as G ={(h, r, t)}, which contains considerable triplets(h, r, t) where h and t are respectively head entityand tail entity, and r denotes the relation. Theirembeddings are denoted as (h, r, t).

Formally, given a pair of entities (h, t) in aKG G and a bag of instances (sentences) B ={s1, s2, · · · , sm}, where each instance si contains(h, t), the task of relation extraction is to train aclassifier based on B to predict the relation labely of (h, t) from a predefined relation set. If norelation exists, we simply assign NA to it.

To improve relation extraction, we make fulluse of both the KG G and entity descriptionsD = {d1, d2, · · · , dn} to learn label embeddingswhich can benefit selection of valid instances. Foreach entity ei, we take the first paragraph of itscorresponding Wikipedia page as its descriptiontext di = {w1, w2, · · · , wl}, where wi ∈ V de-notes the description word, l is the length and V isthe vocabulary.

4 Our Proposed Model

In this section, we will detail our proposed multi-layer attention-based model RELE for relationextraction with joint label embedding. Existingmethods for relation extraction take labels as inde-pendent and meaningless one-hot vectors, whichcause a loss of potential label information for se-lecting valid instances. Additionally, they don’ttake full advantage of both structural informationfrom KGs and textual information from entity de-scriptions and ignore the imposed noisy informa-tion.

As shown in Figure 2, our model is based ona multi-layer attention, containing two parts: 1)label embedding (shown in the right) and 2) neu-ral relation classification (shown in the left) . Theformer makes full use of structural informationfrom KGs and textual information from entity de-scriptions to learn label embeddings through gat-ing mechanism, while avoiding the imposed noisewith an attention mechanism. The latter leveragesthe label embeddings as another attention over the

𝒅𝒆

𝒔𝒎Entity

Descriptions

𝒔𝟏 𝒕 𝒉

CNNCNN

𝒓𝒉𝒕

𝒓𝒉𝒕′

𝒓𝒔

𝒅 = (𝒘𝟏,⋯ ,𝒘𝒍 )

Knowledge

GraphSentences

𝒔𝟏′ 𝒔𝒎

Conv Conv…

Attention 𝜶

Attention 𝝀

⨁ ⨁ G

Gating

Mechanism

Concatenation

Label

EmbeddingClassifierClassifier

Figure 2: Illustration of our multi-layer attention basedmodel RELE.

instances to select valid instances for improvingneural relation extraction. Note that we also usethe entity descriptions to enhance the representa-tions of instances. We detail the two parts as fol-lows.

4.1 Joint Label Embedding

Label information plays a vital role in selectingvalid instances for improving relation extraction.We make full use of both structural informationfrom KGs (Han et al., 2018a) and textual infor-mation from entity descriptions (Ji et al., 2017) tolearn label embeddings via gating integration. En-tity descriptions provide rich background knowl-edge for entities (Newman-Griffis et al., 2018) andare supposed to benefit the label embedding andrelation extraction. Nevertheless, as shown in Fig-ure 1, entity descriptions may also contain irrel-evant and even misleading information. There-fore, we propose to use KG embeddings to at-tend over the entity descriptions, alleviating theimposed noise. Then a gating mechanism is usedto integrate both the KGs and entity descriptionsfor learning label embeddings.

KG Embedding. We use TransE (Bordes et al.,2013) for KG embedding. Given a triplet (h, r, t),the model aims to learn low-dimensional represen-tations vector for entities h, t and the relationshipr into the same vector space, and regards a rela-tionship r as a translation from the head entity hto tail entity t, assuming the embedding t should

Page 4: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3824

be close to h+ r if (h, r, t) exists. The score func-tion is defined as :

f(h, r, t) = −‖h + r− t‖22. (1)

Note that since the true relations in test set are un-known, we simply represent the relation by:

rht = t− h. (2)

In this way, we can also get the relation embed-dings given the entity pairs during testing.

Entity Description Embedding. Then we usethe representations of relations rht as attentionover the words of an entity description to reducethe weights of noisy words. Formally, for each en-tity e, we learn the representation of its descriptiond = (w1, w2, · · · , wl) as follows:

xi = CNN(wi− c−12, · · · ,wi+ c−1

2), (3)

xi = tanh(Wx · xi + bx), (4)

αi =exp(xi · rht)∑i=1 exp(xi · rht)

, (5)

de =

l∑i=1

αixi, (6)

where CNN(·) denotes a convolution layer withwindow size c over the word sequence. xi ∈ RDh

is the hidden representation of the word wi. Wx isthe weight matrix and bx is the bias vector. αi isthe attention weight of the word wi, which is com-puted based on the relation embedding rht. Fi-nally, the text description embedding de is com-puted by the weighted average of words.

Gating Integration. We apply a gating mecha-nism (Xu et al., 2016) to integrate the textual entitydescription embedding de and the structural infor-mation (entity embedding e) from KGs:

e′ = g � e + (1− g)� de, (7)

where g ∈ RDw is a gating vector for integration,e′ ∈ RDw represents the final integrated entityembedding and � represents Hadamard product.Consequently, we compute the final label embed-ding l:

l = t′ − h′. (8)

Label Classifier. Ideally, each label embeddingis supposed to act as an “anchor” for each relationclass. To achieve this goal, we consider to train

the learned label embeddings l to be easily clas-sified as the correct relation class. Therefore, weuse softmax to get the predicted probabilities ofthe relation classes:

P(y|D,G) = Softmax(Mkl + bk), (9)

where Mk is the transformation matrix, bk is thebias.

4.2 Neural Relation ExtractionAfter obtaining the embeddings of labels and en-tity descriptions, we leverage them to advance theneural relation extraction. We first use the en-tity descriptions to enhance instance embeddings.Then we leverage the label embeddings to attendover the instances to select valid instances for re-lation classification.

Instance Embedding. We enrich the repre-sentation of an instance with the pair of entitydescriptions. Firstly, for each word w ∈ s ={w1, · · · , wn}, its embedding wi is initialized asfollows:

wi = wi ⊕ pi1 ⊕ pi2, (10)

where wi is the pre-trained word vector of wi, pi1

and pi2 are its position embeddings to incorporaterelative distances to two target entities into twoDp-dimensional vectors respectively (Zeng et al.,2014). The symbol ⊕ represents concatenationoperator.

Then, we choose CNN (Zeng et al., 2014) withwindow size c as our encoder to learn the instanceembedding considering the text of the instance it-self.

zi = CNN(wi− c−12, · · · , wi+ c−1

2), (11)

[s]j = max{[z1]j , · · · , [zn]j}, (12)

where s ∈ RDh is the sentence (instance) embed-ding, [·]j is the j-th value of a vector and functionmax denotes max-pooling.

Finally, we concatenate the original instanceembedding s with the entity descriptions (dh, dt),obtaining the new instance representation s′. For-mally,

s′ = s⊕ dh ⊕ dt. (13)

Attention over Instances. To alleviate thewrong labeling problem of distant supervision, weleverage the label embedding l as attention over

Page 5: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3825

instances to reduce the weights of noisy instancesin the sentence bag B = {s1, · · · , sm}. Then therepresentation of textual relation feature s from thebagB can be calculated as weighted average of theinstance embeddings s′i:

si = tanh(Wss′i + bs), (14)

λi =exp(l · si)∑mj=1 exp(l · si)

, (15)

s =

m∑i=1

λis′i, (16)

where Ws is the weight matrix and bs is the biasvector. λi is the attention score of the instance si,computed based on the label embedding l.

Relation Classifier. Finally, to compute theconfidence of each relation class, we feed the rep-resentation of the textual relation s into a softmaxclassifier after being processed by a linear trans-formation. Formally,

P(y|B) = Softmax(Mss + bs), (17)

where Ms is the transformation matrix, and bs isthe bias.

4.3 Model TrainingThe objective function of our joint model RELEincludes two parts, the loss of label classifier L1

and the loss of relation classifier L2.Assuming that there are N bags in training set{B1, B2, · · · , BN}, and their corresponding la-bels {y1, y2, · · · , yN}, we exploit cross entropyfor the loss function of label classifier L1:

L1 = −N∑i=1

log P(yi|D,G), (18)

The loss L1 aims to train the label embedding tobe classified to the correct relation class.

Similarly, for the loss of relation classifier L2,we also exploit cross entropy and get:

L2 = −N∑i=1

log P(yi|Bi), (19)

Finally, we aim to minimize the loss function Lwith L2-norm:

minL = L1 + L2 + η‖Θ‖2, (20)

where η is the regularization coefficient and Θdenotes the parameters in our model. Stochasticgradient descent (SGD) is used to optimize ourmodel.

5 Experiments

5.1 DatasetsIn our experiments, we evaluate our model overthe NYT-FB60K dataset (Han et al., 2018a) andGIDS-FB8K dataset (Jat et al., 2018). In the fol-lowing, we detail each dataset.

• NYT-FB60K. The dataset NYT-FB60K (Hanet al., 2018a) includes three parts: knowl-edge graphs (FB60K extended from Riedeldataset (Riedel et al., 2010), containing1,324 relations, 69,512 entities, and 335,350facts), text corpus (whose sentences are fromRiedel dataset, containing 570,088 sentences,63,696 entities, and 56 relations) and entitydescriptions (which are the first paragraphsof the entities’ Wikipedia pages, containingaround 80 words on average).

• GIDS-FB8K. We construct the dataset GIDS-FB8K based on GIDS dataset (Jat et al.,2018). It also contains three parts: knowl-edge graphs (FB8K extended from GIDSdataset, containing 208 relations, 8,917 en-tities, and 38,509 facts), text corpus (whosesentences are from GIDS dataset, contain-ing 16,960 sentences, 14,261 entities, and 5relations) and entity descriptions (which arethe first paragraphs of the entities’ Wikipediapages, containing around 80 words on aver-age).

5.2 BaselinesWe compare our model with the state-of-the-artbaselines:

• Mintz (Mintz et al., 2009). A multi-class lo-gistic regression model under distant supervi-sion.

• MultiR (Hoffmann et al., 2011). A prob-abilistic graphical model for multi-instancelearning.

• MIMLRE (Surdeanu et al., 2012). A graph-ical model for multi-instance multi-labellearning.

• CNN+ATT (Lin et al., 2016). A CNN modelwith instance-level attention.

• APCNN+D (Ji et al., 2017) A PiecewiseCNN model with instance-level attention us-ing entity descriptions. As no code available,we implemented it by ourselves.

Page 6: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3826

0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40Recall

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0Pr

ecisi

onRELEJointD+KATTAPCNN+DMIMLRE

MintzMultiRCNN+ATTRESIDE

(a) NYT-FB60K Dataset

0.0 0.2 0.4 0.6 0.8 1.0Recall

0.20.30.40.50.60.70.80.91.0

Prec

ision

RELERESIDEJointD+KATTCNN+ATTAPCNN+D

(b) GIDS-FB8K Dataset

Figure 3: Precision-recall curves for different methods. RELE achieves higher precision over the entire range ofrecall compared to all the baselines on both datasets.

Parameter Value

Word/Entity/Relation Dimension Dw 50Position Dimension Dp 5Hidden Layer Dimension Dh 230Kernel Size c 3Learning Rate α 0.5Regularization Coefficient η 0.0001Dropout Probability p 0.5

Table 1: Parameter Settings.

• JointD+KATT (Han et al., 2018a). A jointmodel for knowledge graph embedding andrelation extraction.

• Reside (Vashishth et al., 2018). A neural net-work based model which makes use of rel-evant side information and employs GraphConvolution Networks for encoding syntac-tic information of instances.

5.3 Evaluation Metrics

Following previous studies (Lin et al., 2016),our model is evaluated held-out, which comparesthe relations discovered from test corpus withthose in Freebase. We report the Precision-Recallcurve and top-N precision (P@N) metric for NYT-FB60K dataset.

For GIDS-FB8K dataset, we report thePrecision-Recall curve, F1 score and Mean Av-erage Precision (MAP). We do not use the top-Nprecision (P@N) metric for the dataset which issmall containing only 5 relation classes.

5.4 Parameter SettingsFor all the models, we use the pre-trained wordembeddings with word2vec tool∗ on NYT corpusfor initialization. The embeddings of entities men-tioned in datasets are pre-trained through TransEmodel. We select the learning rate α among{0.1, 0.01, 0.005, 0.001} for minimizing the loss.For other parameters, we simply follow the set-tings used in (Lin et al., 2016; Han et al., 2018a)so that it can be fairly compared with these base-lines. Table 1 shows all the parameters used in ourexperiment.

5.5 Experiment Results5.5.1 Precision-Recall Curves on Both

DatasetsFigure 3 shows the comparison results in termsof Precision-Recall Curves on NYT-FB60K andGIDS-FB8K datasets. Overall, we observe that:(1) As shown in Figure 3(a), the neural networkbased approaches have more obvious advantagesthan Mintz, MultiR, and MIMLRE, illustrating thelimitation of human-designed features and the ad-vancement of neural networks in relation extrac-tion. (2) APCNN+D using entity descriptions,and JointD+KATT exploiting KGs both outper-form CNN+ATT on both datasets, showing thatentity descriptions and KGs are both useful for im-proving the performance of relation extraction un-der distant supervision. RESIDE achieves betterperformance than APCNN+D and JointD+KATT.It is probably because that RESIDE utilizes moreavailable side information, including entity typesand relation alias information. (3) Our model

∗https://code.google.com/p/word2vec/

Page 7: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3827

Test Setttings One Two All

P@N(%) 100 300 500 Mean 100 300 500 Mean 100 300 500 Mean

CNN+ATT 67.3 61.1 56.5 61.6 73.3 62.1 53.5 63.0 74.2 62.4 55.9 64.1APCNN+D 78.0 67.7 60.6 68.7 77.0 63.3 55.6 65.3 77.0 66.7 63.0 68.9JointD+KATT 80.0 70.0 64.0 71.3 81.0 67.0 59.6 69.2 84.0 68.7 63.7 71.0RESIDE 79.0 67.6 57.4 68.0 83.0 70.6 62.6 72.0 84.0 78.5 69.8 77.4

RELE 87.0 76.6 67.0 76.8 88.0 73.3 63.2 78.7 88.0 78.6 69.8 78.8

Table 2: Evaluation results P@N of different models using different number of sentences in bags on NYT-FB60Kdataset. Here, One, Two and All represent the number of sentences randomly selected from a bag.

JointD+KATT APCNN+D RESIDE RELE73

76

79

82

85

88F1(%)MAP(%)

Figure 4: Comparison on GIDS-FB8K dataset.

RELE achieves the best performance compared toall the baselines. We believe the reason is that wemake use of potential label information throughjoint label embedding. The learned label embed-dings are of high quality since we fully exploitboth the structural information from KGs and tex-tual information from texts via gating integrationwhile avoiding the imposed noise by an attentionmechanism.

5.5.2 P@N Evaluation on NYT-FB60KDataset

As shown in Table 2, we report Precision@Nof different neural network based approaches onNYT-FB60K dataset. To verify the performanceof our model on those entity pairs with few in-stances, we randomly select one, two and allinstances for each entity pair, following previ-ous studies (Du et al., 2018; Vashishth et al.,2018). As we can observe: (1) APCNN+D andJointD+KATT both outperform CNN+ATT in allcases, demonstrating the effectiveness of entity de-scriptions and KGs. RESIDE achieves better per-formance than APCNN+D and JointD+KATT byincorporating more side information (e.g., entitytypes). (2) Our model RELE significantly out-performs all the baselines. The reason is that

P@N(%) 100 300 500 Mean

RELE w/o LE 74.2 62.4 55.9 64.1RELE w/o ATTe 82.0 75.6 69.6 75.7RELE w/o LC 81.0 74.0 69.0 74.7RELE 88.0 78.6 69.8 78.8

Table 3: Evaluation results P@N of variant models onNYT-FB60K dataset.

our model learns high quality of label embeddingswhich play a critical role in relation extraction.

5.5.3 Results on GIDS-FB8K Dataset

Based on the results on the NYT-FB60K dataset,we choose APCNN+D, JointD+KATT and RE-SIDE as representative baselines to compareagainst our models on the GIDS-FB8K dataset interms of F1 and MAP. As shown in Figure 4, ourmodel consistently achieves better performance,which verifies the effectiveness of our model withjoint label embedding. The overall results onGIDS-FB8K dataset show that our model can bewell applied to smaller-scale datasets.

5.5.4 Comparison of Variant Models

In order to verify the effectiveness of differentmodules of our model, we design three variantmodels:

• RELE w/o LE removes the label embed-ding from RELE, which degenerates toCNN+ATT.

• RELE w/o ATTe removes the attention ofthe KG over entity descriptions during la-bel embedding. We use max-pooling instead,which does not consider the noise in the en-tity descriptions.

Page 8: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3828

𝑠1: In Bucharest , Romania , president Traian Basescu said Sunday night that the entry into …

𝑠2: Last year, it opened offices in Warsaw and Bucharest , the capital of Romania.

𝑠3: … by the presence of two films from Romania , the way I spent … east of Bucharest …

𝑠4:… Ervin a of New York City , on April 18th , age 98 , formerly of Bucharest , Romania .

Triplet(Romania, Bucharest, /location/country/capital)JointD+

KATT(×)RELE

(√)APCNN+D

(×)

0

1

Figure 5: A case study for predicting the relation between “Bucharest” and “Romania”. The baselines all predictwrongly while our model gives the right result. The left shows the weights assigned to different sentences bydifferent models. Our model always gives higher weights to correct sentences (shown in red).

Romania is a country located at the crossroads of Central , Eastern ,

and Southeastern Europe . It borders the Black Sea to the southeast ,

… , Romania is the 12th largest country and also the 7th most

populous ... Its capital and largest city is Bucharest , and other major

urban areas include Cluj-Napoca ...

(a) Description text of “Romania”

x

Bucharest is the capital and largest city of Romania , as well as its

cultural , industrial , and financial centre . It is located in the southeast

of the country , … ,on the banks of the Dambovita River , less than 60 km

north of the Danube River and the Bulgarian border.

x (b) Description text of “Bucharest”

Figure 6: Visualization of attention values of words in the descriptions of the entities “Romania” and “Bucharest”.

• RELE w/o LC removes the label classifier.It does not train the label embeddings to beclassified to the correct classes.

As shown in Table 3, without label embedding,the performance of RELE w/o LE drops signifi-cantly (more than 10%). It demonstrates the ef-fectiveness of our joint label embedding. If we donot consider the imposed noise in the entity de-scriptions, the performance of RELE w/o ATTe

decreases by around 3% on mean P@N. It demon-strates that the attention of KGs over entity de-scriptions is important for learning high-quality la-bel embeddings. We also explore the performanceof RELE w/o LC which does not train the labelembeddings to be classified to the correct classesand find that label classifier plays a critical role inlabel embedding. The performance decreased byaround 4% on mean P@N, without label classifier.

5.5.5 Case StudyFigure 5 illustrates an example from the testset of NYT-FB60K. As we can observe, rep-resentative attention-based baselines APCNN+Dand JointD+KATT both predict wrong rela-tion labels for the entities, while our modelRELE correctly predicts the relation label “/loca-tion/country/capital”. That is because the base-lines assign relatively low weights to correct sen-tences (s1 and s2) and achieve inferior representa-tions of the textual relation for classification. Ourmodel RELE assigns higher weights to all the cor-rect sentences s1, s2 and s4, demonstrating that

RELE learns high-quality label embeddings.We also provide the insights of the attention of

KGs over entity descriptions about “Romania” and“Bucharest”. As shown in Figure 6, the words“capital”, “largest”. and “centre” which are re-lated to the relation are given higher weights. Itdemonstrates that the attention of KGs over entitydescriptions can help reduce the noise in entity de-scriptions and thus improve the label embeddings.

6 Conclusion

In this work, we consider leveraging potentiallabel information to select valid instances fordistantly-supervised relation extraction. We pro-pose a novel multi-layer attention-based modelRELE to improve relation extraction with joint la-bel embedding. Our model takes full advantage ofboth structural information from KGs and textualinformation from entity descriptions to learn la-bel embeddings, while avoiding the imposed noisewith an attention mechanism. The label embed-dings are trained to be classified to the correct re-lation classes. Then, the learned label embeddingsare used as another attention over the bag to selectvalid instances for relation extraction. Extensiveexperiments have demonstrated that our modelsignificantly outperforms state-of-the-art methods.

In the future, we will explore other useful in-formation (e.g., correlations among the relationsfrom KGs) available to improve the label embed-dings.

Page 9: Improving Distantly-Supervised Relation Extraction with ... · 2012;Zeng et al.,2015). Recently, significant progress has been made in applying deep neural networks for relation

3829

Acknowledgement

This work is supported by the National NaturalScience Foundation of China (No. 61806020,61772082, 61702296), the National Key Re-search and Development Program of China(2017YFB0803304), the Beijing Municipal Nat-ural Science Foundation (4182043), the CCF-Tencent Open Fund, and the Fundamental Re-search Funds for the Central Universities.

ReferencesZeynep Akata, Florent Perronnin, Zaıd Harchaoui, and

Cordelia Schmid. 2016. Label-embedding for imageclassification. IEEE Transactions on Pattern Analy-sis and Machine Intelligence, 38:1425–1438.

Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko.2013. Translating embeddings for modeling multi-relational data. In NIPS, pages 2787–2795.

Jinhua Du, Jingguang Han, Andy Way, and DadongWan. 2018. Multi-level structured self-attentions fordistantly supervised relation extraction. In EMNLP,pages 2216–2225.

Xiaocheng Feng, Jiang Guo, Bing Qin, Ting Liu, andYongjie Liu. 2017. Effective deep memory networksfor distant supervised relation extraction. In IJCAI,pages 4002–4008.

Xu Han, Zhiyuan Liu, and Maosong Sun. 2018a. Neu-ral knowledge acquisition via mutual attention be-tween knowledge graph and text. In AAAI, pages4832–4839.

Xu Han, Pengfei Yu, Zhiyuan Liu, Maosong Sun, andPeng Li. 2018b. Hierarchical relation extractionwith coarse-to-fine grained attention. In EMNLP,pages 2236–2245.

Raphael Hoffmann, Congle Zhang, Xiao Ling, LukeZettlemoyer, and Daniel S Weld. 2011. Knowledge-based weak supervision for information extractionof overlapping relations. In ACL, pages 541–550.

Sharmistha Jat, Siddhesh Khandelwal, and ParthaTalukdar. 2018. Improving distantly supervised re-lation extraction using word and entity based atten-tion. arXiv preprint arXiv:1804.06987.

Guoliang Ji, Kang Liu, Shizhu He, Jun Zhao, et al.2017. Distant supervision for relation extractionwith sentence-level attention and entity descriptions.In AAAI, pages 3060–3066.

Yankai Lin, Shiqi Shen, Zhiyuan Liu, Huanbo Luan,and Maosong Sun. 2016. Neural relation extractionwith selective attention over instances. In ACL, vol-ume 1, pages 2124–2133.

Mike Mintz, Steven Bills, Rion Snow, and Dan Juraf-sky. 2009. Distant supervision for relation extrac-tion without labeled data. In ACL/IJCNLP, pages1003–1011.

Denis Newman-Griffis, Albert M Lai, and Eric Fosler-Lussier. 2018. Jointly embedding entities and textwith distant supervision. ACL 2018, page 195.

Sebastian Riedel, Limin Yao, and Andrew McCallum.2010. Modeling relations and their mentions with-out labeled text. In ECML/PKDD, pages 148–163.

Jose A Rodriguez-Serrano, Albert Gordo, and FlorentPerronnin. 2015. Label embedding: A frugal base-line for text recognition. International Journal ofComputer Vision, 113(3):193–207.

Mihai Surdeanu, Julie Tibshirani, Ramesh Nallapati,and Christopher D Manning. 2012. Multi-instancemulti-label learning for relation extraction. InEMNLP-CoNLL, pages 455–465.

Shikhar Vashishth, Rishabh Joshi, Sai Suman Prayaga,Chiranjib Bhattacharyya, and Partha Talukdar. 2018.Reside: Improving distantly-supervised neural rela-tion extraction using side information. In EMNLP,pages 1257–1266.

Guoyin Wang, Chunyuan Li, Wenlin Wang, YizheZhang, Dinghan Shen, Xinyuan Zhang, RicardoHenao, and Lawrence Carin. 2018. Joint embeddingof words and labels for text classification. In ACL,volume 1, pages 2321–2331.

Jiacheng Xu, Kan Chen, Xipeng Qiu, and XuanjingHuang. 2016. Knowledge graph representation withjointly structural and textual encoding. In IJCAI,pages 1318–1324.

Daojian Zeng, Kang Liu, Yubo Chen, and Jun Zhao.2015. Distant supervision for relation extractionvia piecewise convolutional neural networks. InEMNLP, pages 1753–1762.

Daojian Zeng, Kang Liu, Siwei Lai, Guangyou Zhou,and Jun Zhao. 2014. Relation classification via con-lutional deep neural network. In COLING, pages2335–2344.

Wenyuan Zeng, Yankai Lin, Zhiyuan Liu, andMaosong Sun. 2017. Incorporating relation paths inneural relation extraction. In EMNLP, pages 1768–1777.

Yuyu Zhang, Hanjun Dai, Zornitsa Kozareva, Alexan-der J Smola, and Le Song. 2018. Variational reason-ing for question answering with knowledge graph.In AAAI, pages 6069–6076.

Hao Zhou, Tom Young, Minlie Huang, Haizhou Zhao,Jingfang Xu, and Xiaoyan Zhu. 2018. Com-monsense knowledge aware conversation generationwith graph attention. In IJCAI, pages 4623–4629.