TY - JOUR
T1 - A joint neural model for fine-grained named entity classification of wikipedia articles
AU - Suzuki, Masatoshi
AU - Matsuda, Koji
AU - Sekine, Satoshi
AU - Okazaki, Naoaki
AU - Inui, Kentaro
N1 - Funding Information:
Manuscript received February 15, 2017. Manuscript revised June 15, 2017. Manuscript publicized September 15, 2017. †The authors are with Tohoku University, Sendai-shi, 980– 8579 Japan. ††The author is with Language Craft Inc, Tokyo, 152–0031 Japan. †††The authors are with RIKEN Center for Advanced Intelligence Project, Tokyo, 103–0027 Japan. ††††The author is with Tokyo Institute of Technology, Tokyo, 152–8550 Japan. ∗This paper is originally published in Proceedings of the 30th Pacific Asia Conference on Language, Information and Computation. Major changes are made to Sect. 3 where we introduce the NE type set and explain assumptions in NE type annotation, and Sect. 5.2 where we report the results of t-test on F1 values. All the figures and tables are from the original paper [1]. This work was partially supported by Research and Development on Real World Big Data Integration and Analysis, MEXT and JSPS KAKENHI Grant 15H05318 and 15H01702. a) E-mail: m.suzuki@ecei.tohoku.ac.jp b) E-mail: matsuda@ecei.tohoku.ac.jp c) E-mail: sekine@languagecarft.com d) E-mail: okazaki@c.titech.ac.jp e) E-mail: inui@ecei.tohoku.ac.jp DOI: 10.1587/transinf.2017SWP0005
Publisher Copyright:
Copyright © 2018 The Institute of Electronics, Information and Communication Engineers
PY - 2018/1
Y1 - 2018/1
N2 - This paper addresses the task of assigning labels of fine-grained named entity (NE) types to Wikipedia articles. Information of NE types are useful when extracting knowledge of NEs from natural language text. It is common to apply an approach based on supervised machine learning to named entity classification. However, in a setting of classifying into fine-grained types, one big challenge is how to alleviate the data sparseness problem since one may obtain far fewer instances for each fine-grained types. To address this problem, we propose two methods. First, we introduce a multi-task learning framework, in which NE type classifiers are all jointly trained with a neural network. The neural network has a hidden layer, where we expect that effective combinations of input features are learned across different NE types. Second, we propose to extend the input feature set by exploiting the hyperlink structure of Wikipedia. While most of previous studies are focusing on engineering features from the articles’ contents, we observe that the information of the contexts the article is mentioned can also be a useful clue for NE type classification. Concretely, we propose to learn article vectors (i.e. entity embeddings) from Wikipedia’s hyperlink structure using a Skip-gram model. Then we incorporate the learned article vectors into the input feature set for NE type classification. To conduct large-scale practical experiments, we created a new dataset containing over 22,000 manually labeled articles. With the dataset, we empirically show that both of our ideas gained their own statistically significant improvement separately in classification accuracy. Moreover, we show that our proposed methods are particularly effective in labeling infrequent NE types. We’ve made the learned article vectors publicly available. The labeled dataset is available if one contacts the authors.
AB - This paper addresses the task of assigning labels of fine-grained named entity (NE) types to Wikipedia articles. Information of NE types are useful when extracting knowledge of NEs from natural language text. It is common to apply an approach based on supervised machine learning to named entity classification. However, in a setting of classifying into fine-grained types, one big challenge is how to alleviate the data sparseness problem since one may obtain far fewer instances for each fine-grained types. To address this problem, we propose two methods. First, we introduce a multi-task learning framework, in which NE type classifiers are all jointly trained with a neural network. The neural network has a hidden layer, where we expect that effective combinations of input features are learned across different NE types. Second, we propose to extend the input feature set by exploiting the hyperlink structure of Wikipedia. While most of previous studies are focusing on engineering features from the articles’ contents, we observe that the information of the contexts the article is mentioned can also be a useful clue for NE type classification. Concretely, we propose to learn article vectors (i.e. entity embeddings) from Wikipedia’s hyperlink structure using a Skip-gram model. Then we incorporate the learned article vectors into the input feature set for NE type classification. To conduct large-scale practical experiments, we created a new dataset containing over 22,000 manually labeled articles. With the dataset, we empirically show that both of our ideas gained their own statistically significant improvement separately in classification accuracy. Moreover, we show that our proposed methods are particularly effective in labeling infrequent NE types. We’ve made the learned article vectors publicly available. The labeled dataset is available if one contacts the authors.
KW - Multi-task learning
KW - Named entity classification
KW - Neural network
KW - Wikipedia
UR - http://www.scopus.com/inward/record.url?scp=85040253504&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85040253504&partnerID=8YFLogxK
U2 - 10.1587/transinf.2017SWP0005
DO - 10.1587/transinf.2017SWP0005
M3 - Article
AN - SCOPUS:85040253504
SN - 0916-8532
VL - E101D
SP - 73
EP - 81
JO - IEICE Transactions on Information and Systems
JF - IEICE Transactions on Information and Systems
IS - 1
ER -