Commit dae1ac87 by 20200318029

homework5

parent 6fcc50c5
...@@ -1038,7 +1038,7 @@ ...@@ -1038,7 +1038,7 @@
" # Self-Attention:注意self-attention的q,k和v均为decoder hidden\n", " # Self-Attention:注意self-attention的q,k和v均为decoder hidden\n",
" x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, tgt_mask))\n", " x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, tgt_mask))\n",
" # Context-Attention:注意context-attention的q为decoder hidden,而k和v为encoder hidden\n", " # Context-Attention:注意context-attention的q为decoder hidden,而k和v为encoder hidden\n",
" x = self.sublayer[1](x, lambda x: self.self_attn(x, m, m, src_mask))\n", " x = self.sublayer[1](x, lambda x: self.src_attn(x, m, m, src_mask))\n",
" return self.sublayer[2](x, self.feed_forward)" " return self.sublayer[2](x, self.feed_forward)"
] ]
}, },
...@@ -9340,4 +9340,4 @@ ...@@ -9340,4 +9340,4 @@
}, },
"nbformat": 4, "nbformat": 4,
"nbformat_minor": 2 "nbformat_minor": 2
} }
\ No newline at end of file
...@@ -742,7 +742,7 @@ class DecoderLayer(nn.Module): ...@@ -742,7 +742,7 @@ class DecoderLayer(nn.Module):
# Self-Attention:注意self-attention的q,k和v均为decoder hidden # Self-Attention:注意self-attention的q,k和v均为decoder hidden
x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, tgt_mask)) x = self.sublayer[0](x, lambda x: self.self_attn(x, x, x, tgt_mask))
# Context-Attention:注意context-attention的q为decoder hidden,而k和v为encoder hidden # Context-Attention:注意context-attention的q为decoder hidden,而k和v为encoder hidden
x = self.sublayer[1](x, lambda x: self.self_attn(x, m, m, src_mask)) x = self.sublayer[1](x, lambda x: self.src_attn(x, m, m, src_mask))
return self.sublayer[2](x, self.feed_forward) return self.sublayer[2](x, self.feed_forward)
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment