微信公众号搜"智元新知"关注
微信扫一扫可直接关注哦!

Attention machenism

from attention mechanism

Attention is one component of a network’s architecture, and is in charge of managing and quantifying the interdependence.

  1. Between the input and output elements (General Attention)
  2. Within the input elements (Self-Attention)

While Attention does have its application in other fields of deep learning such as Computer Vision, its main breakthrough and success come from its application in Natural Language Processing (NLP) tasks. This is due to the fact that Attention was introduced to address the problem of long sequences in Machine Translation, which is also a problem for most other NLP tasks as well.

版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 [email protected] 举报,一经查实,本站将立刻删除。

相关推荐