TY - JOUR
T1 - Relation is an option for processing context information
AU - Yamada, Kazunori D.
AU - Baladram, M. Samy
AU - Lin, Fangzhou
N1 - Publisher Copyright:
Copyright © 2022 Yamada, Baladram and Lin.
PY - 2022/10/11
Y1 - 2022/10/11
N2 - Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., include the Attention. Various improvements have been made to enhance its performance since Attention is a powerful component to realize artificial intelligence. The time complexity of Attention depends on the square of the input sequence length. Developing methods to improve the time complexity of Attention is one of the most popular research topics. Attention is a mechanism that conveys contextual information of input sequences to downstream networks. Thus, if one wants to improve the performance of processing contextual information, the focus should not be confined only on improving Attention but also on devising other similar mechanisms as possible alternatives. In this study, we devised an alternative mechanism called “Relation” that can understand the context information of sequential data. Relation is easy to implement, and its time complexity depends only on the length of the sequences; a comparison of the performance of Relation and Attention on several benchmark datasets showed that the context processing capability of Relation is comparable to that of Attention but with less computation time. Processing contextual information at high speeds would be useful because natural language processing and biological sequence processing sometimes deal with very long sequences. Hence, Relation is an ideal option for processing context information.
AB - Attention mechanisms are one of the most frequently used architectures in the development of artificial intelligence because they can process contextual information efficiently. Various artificial intelligence architectures, such as Transformer for processing natural language, image data, etc., include the Attention. Various improvements have been made to enhance its performance since Attention is a powerful component to realize artificial intelligence. The time complexity of Attention depends on the square of the input sequence length. Developing methods to improve the time complexity of Attention is one of the most popular research topics. Attention is a mechanism that conveys contextual information of input sequences to downstream networks. Thus, if one wants to improve the performance of processing contextual information, the focus should not be confined only on improving Attention but also on devising other similar mechanisms as possible alternatives. In this study, we devised an alternative mechanism called “Relation” that can understand the context information of sequential data. Relation is easy to implement, and its time complexity depends only on the length of the sequences; a comparison of the performance of Relation and Attention on several benchmark datasets showed that the context processing capability of Relation is comparable to that of Attention but with less computation time. Processing contextual information at high speeds would be useful because natural language processing and biological sequence processing sometimes deal with very long sequences. Hence, Relation is an ideal option for processing context information.
KW - Attention
KW - Relation
KW - Transformer
KW - artificial intelligence
KW - multilayer perceptron
KW - neural networks
KW - time complexity
UR - http://www.scopus.com/inward/record.url?scp=85140386375&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85140386375&partnerID=8YFLogxK
U2 - 10.3389/frai.2022.924688
DO - 10.3389/frai.2022.924688
M3 - Article
AN - SCOPUS:85140386375
SN - 2624-8212
VL - 5
JO - Frontiers in Artificial Intelligence
JF - Frontiers in Artificial Intelligence
M1 - 924688
ER -