Commit dae1ac87 by 20200318029

homework5

parent 6fcc50c5
......@@ -1038,7 +1038,7 @@
" # Self-Attention:注意self-attention的q,k和v均为decoder hidden\n",
" x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, tgt_mask))\n",
" # Context-Attention:注意context-attention的q为decoder hidden,而k和v为encoder hidden\n",
" x = self.sublayer[1](x, lambda x: self.self_attn(x, m, m, src_mask))\n",
" x = self.sublayer[1](x, lambda x: self.src_attn(x, m, m, src_mask))\n",
" return self.sublayer[2](x, self.feed_forward)"
]
},
......@@ -9340,4 +9340,4 @@
},
"nbformat": 4,
"nbformat_minor": 2
}
}
\ No newline at end of file
......@@ -742,7 +742,7 @@ class DecoderLayer(nn.Module):
# Self-Attention:注意self-attention的q,k和v均为decoder hidden
x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, tgt_mask))
# Context-Attention:注意context-attention的q为decoder hidden,而k和v为encoder hidden
x = self.sublayer[1](x, lambda x: self.self_attn(x, m, m, src_mask))
x = self.sublayer[1](x, lambda x: self.src_attn(x, m, m, src_mask))
return self.sublayer[2](x, self.feed_forward)
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment