Entity-aware Attention for Relation Classification
This repository contains the official TensorFlow implementation of the following paper:
Semantic Relation Classification via Bidirectional LSTM Networks with Entity-aware Attention using Latent Entity Typing
Joohong Lee, Sangwoo Seo, Yong Suk Choi
https://arxiv.org/abs/1901.08163Abstract: Classifying semantic relations between entity pairs in sentences is an important task in Natural Language Processing (NLP). Most previous models for relation classification rely on the high-level lexical and syntactic features obtained by NLP tools such as WordNet, dependency parser, part-of-speech (POS) tagger, and named entity recognizers (NER). In addition, state-of-the-art neural models based on attention mechanisms do not fully utilize information of entity that may be the most crucial features for relation classification. To address these issues, we propose a novel end-to-end recurrent neural model which incorporates an entity-aware attention mechanism with a latent entity typing (LET) method. Our model not only utilizes entities and their latent types as features effectively but also is more interpretable by visualizing attention mechanisms applied to our model and results of LET. Experimental results on the SemEval-2010 Task 8, one of the most popular relation classification task, demonstrate that our model outperforms existing state-of-the-art models without any high-level features.
Usage
Train / Test
- Train data is located in "SemEval2010_task8_all_data/SemEval2010_task8_training/TRAIN_FILE.TXT".
- You can apply some pre-trained word embeddings: word2vec, glove100, glove300, and elmo. The pre-trained files should be located in
resource/
. Check this code. - In every evaluation step, the test performance is evaluated by test dataset located in "SemEval2010_task8_all_data/SemEval2010_task8_testing_keys/TEST_FILE_FULL.TXT".
Display help message:
$ python train.py --help
Train Example:
$ python train.py --embeddings glove300
Visualization
SemEval-2010 Task #8
- Given: a pair of nominals
- Goal: recognize the semantic relation between these nominals.
- Example:
- "There were apples, pears and oranges in the bowl."
β CONTENT-CONTAINER(pears, bowl) - βThe cup contained tea from dried ginseng.β
β ENTITY-ORIGIN(tea, ginseng)
- "There were apples, pears and oranges in the bowl."
The Inventory of Semantic Relations
- Cause-Effect(CE): An event or object leads to an effect(those cancers were caused by radiation exposures)
- Instrument-Agency(IA): An agent uses an instrument(phone operator)
- Product-Producer(PP): A producer causes a product to exist (a factory manufactures suits)
- Content-Container(CC): An object is physically stored in a delineated area of space (a bottle full of honey was weighed) Hendrickx, Kim, Kozareva, Nakov, O SΒ΄ eaghdha, Pad Β΄ o,Β΄ Pennacchiotti, Romano, Szpakowicz Task Overview Data Creation Competition Results and Discussion The Inventory of Semantic Relations (III)
- Entity-Origin(EO): An entity is coming or is derived from an origin, e.g., position or material (letters from foreign countries)
- Entity-Destination(ED): An entity is moving towards a destination (the boy went to bed)
- Component-Whole(CW): An object is a component of a larger whole (my apartment has a large kitchen)
- Member-Collection(MC): A member forms a nonfunctional part of a collection (there are many trees in the forest)
- Message-Topic(CT): An act of communication, written or spoken, is about a topic (the lecture was about semantics)
- OTHER: If none of the above nine relations appears to be suitable.
Distribution for Dataset
-
SemEval-2010 Task #8 Dataset [Download]
Relation Train Data Test Data Total Data Cause-Effect 1,003 (12.54%) 328 (12.07%) 1331 (12.42%) Instrument-Agency 504 (6.30%) 156 (5.74%) 660 (6.16%) Product-Producer 717 (8.96%) 231 (8.50%) 948 (8.85%) Content-Container 540 (6.75%) 192 (7.07%) 732 (6.83%) Entity-Origin 716 (8.95%) 258 (9.50%) 974 (9.09%) Entity-Destination 845 (10.56%) 292 (10.75%) 1137 (10.61%) Component-Whole 941 (11.76%) 312 (11.48%) 1253 (11.69%) Member-Collection 690 (8.63%) 233 (8.58%) 923 (8.61%) Message-Topic 634 (7.92%) 261 (9.61%) 895 (8.35%) Other 1,410 (17.63%) 454 (16.71%) 1864 (17.39%) Total 8,000 (100.00%) 2,717 (100.00%) 10,717 (100.00%)