Skip to main content

Long short-term attention

Publication ,  Chapter
Zhong, G; Lin, X; Chen, K; Li, Q; Huang, K
January 1, 2020

Attention is an important cognition process of humans, which helps humans concentrate on critical information during their perception and learning. However, although many machine learning models can remember information of data, they have no the attention mechanism. For example, the long short-term memory (LSTM) network is able to remember sequential information, but it cannot pay special attention to part of the sequences. In this paper, we present a novel model called long short-term attention (LSTA), which seamlessly integrates the attention mechanism into the inner cell of LSTM. More than processing long short term dependencies, LSTA can focus on important information of the sequences with the attention mechanism. Extensive experiments demonstrate that LSTA outperforms LSTM and related models on the sequence learning tasks.

Duke Scholars

DOI

Publication Date

January 1, 2020

Volume

11691 LNAI

Start / End Page

45 / 54

Related Subject Headings

  • Artificial Intelligence & Image Processing
  • 46 Information and computing sciences
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Zhong, G., Lin, X., Chen, K., Li, Q., & Huang, K. (2020). Long short-term attention (Vol. 11691 LNAI, pp. 45–54). https://doi.org/10.1007/978-3-030-39431-8_5
Zhong, G., X. Lin, K. Chen, Q. Li, and K. Huang. “Long short-term attention,” 11691 LNAI:45–54, 2020. https://doi.org/10.1007/978-3-030-39431-8_5.
Zhong G, Lin X, Chen K, Li Q, Huang K. Long short-term attention. In 2020. p. 45–54.
Zhong, G., et al. Long short-term attention. Vol. 11691 LNAI, 2020, pp. 45–54. Scopus, doi:10.1007/978-3-030-39431-8_5.
Zhong G, Lin X, Chen K, Li Q, Huang K. Long short-term attention. 2020. p. 45–54.

DOI

Publication Date

January 1, 2020

Volume

11691 LNAI

Start / End Page

45 / 54

Related Subject Headings

  • Artificial Intelligence & Image Processing
  • 46 Information and computing sciences