From 8dc85f54939e87ea2900f477f7a8c96e35c15acc Mon Sep 17 00:00:00 2001 From: Michael Darmanis <49272306+mdarm@users.noreply.github.com> Date: Sun, 7 Apr 2024 12:58:04 +0300 Subject: [PATCH] Fixed minor grammatic mistake. --- the_annotated_transformer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/the_annotated_transformer.py b/the_annotated_transformer.py index 4aa1d46..5342fc3 100644 --- a/the_annotated_transformer.py +++ b/the_annotated_transformer.py @@ -433,7 +433,7 @@ def forward(self, x, memory, src_mask, tgt_mask): # # We also modify the self-attention sub-layer in the decoder stack to # prevent positions from attending to subsequent positions. This -# masking, combined with fact that the output embeddings are offset by +# masking, combined with the fact that output embeddings are offset by # one position, ensures that the predictions for position $i$ can # depend only on the known outputs at positions less than $i$.