TY - GEN
T1 - Word rotator's distance
AU - Yokoi, Sho
AU - Takahashi, Ryo
AU - Akama, Reina
AU - Suzuki, Jun
AU - Inui, Kentaro
N1 - Funding Information:
We appreciate the helpful comments from the anonymous reviewers. We thank Emad Kebriaei for indicating how to normalize the sentence vectors. We also thank Benjamin Heinzerling, Masashi Yoshikawa, Hiroki Ouchi, Sosuke Kobayashi, Paul Reisert, Ana Brassard, and Shun Kiyono for constructive comments on the manuscript, and Masatoshi Suzuki and Goro Kobayashi for technical support. This work was supported by JSPS KAKENHI Grants JP19J21913 and JP19H04162.
Publisher Copyright:
© 2020 Association for Computational Linguistics
PY - 2020
Y1 - 2020
N2 - One key principle for assessing textual similarity is measuring the degree of semantic overlap between two texts by considering the word alignment. Such alignment-based approaches are both intuitive and interpretable; however, they are empirically inferior to the simple cosine similarity between general-purpose sentence vectors. We focus on the fact that the norm of word vectors is a good proxy for word importance, and the angle of them is a good proxy for word similarity. Alignment-based approaches do not distinguish the norm and direction, whereas sentence-vector approaches automatically use the norm as the word importance. Accordingly, we propose decoupling word vectors into their norm and direction then computing the alignment-based similarity with the help of earth mover's distance (optimal transport), which we refer to as word rotator's distance. Furthermore, we demonstrate how to “grow” the norm and direction of word vectors (vector converter); this is a new systematic approach derived from the sentence-vector estimation methods, which can significantly improve the performance of the proposed method. On several STS benchmarks, the proposed methods outperform not only alignment-based approaches but also strong baselines.
AB - One key principle for assessing textual similarity is measuring the degree of semantic overlap between two texts by considering the word alignment. Such alignment-based approaches are both intuitive and interpretable; however, they are empirically inferior to the simple cosine similarity between general-purpose sentence vectors. We focus on the fact that the norm of word vectors is a good proxy for word importance, and the angle of them is a good proxy for word similarity. Alignment-based approaches do not distinguish the norm and direction, whereas sentence-vector approaches automatically use the norm as the word importance. Accordingly, we propose decoupling word vectors into their norm and direction then computing the alignment-based similarity with the help of earth mover's distance (optimal transport), which we refer to as word rotator's distance. Furthermore, we demonstrate how to “grow” the norm and direction of word vectors (vector converter); this is a new systematic approach derived from the sentence-vector estimation methods, which can significantly improve the performance of the proposed method. On several STS benchmarks, the proposed methods outperform not only alignment-based approaches but also strong baselines.
UR - http://www.scopus.com/inward/record.url?scp=85103505211&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85103505211&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85103505211
T3 - EMNLP 2020 - 2020 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference
SP - 2944
EP - 2960
BT - EMNLP 2020 - 2020 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference
PB - Association for Computational Linguistics (ACL)
T2 - 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020
Y2 - 16 November 2020 through 20 November 2020
ER -