Is Attention Linear a Transformation

前端 未结 0 927
失恋的感觉
失恋的感觉 2020-11-28 14:16

In sample for pytorch for SeqtoSeq attention is calculated with linear layer followed by softmax.

embedded = self.embedding(input).view(1, 1, -1)
embedded = s         


        
相关标签:
回答
  • 消灭零回复
提交回复
热议问题