-
Notifications
You must be signed in to change notification settings - Fork 65
Open
Description
Lines 230 to 235 in 067f8a3
# (B, Seq_Len, Dim) + (B, Seq_Len, Dim) --> (B, Seq_Len, Dim) | |
h = x + self.attention.forward( | |
self.attention_norm(x), start_pos, freqs_complex | |
) | |
# (B, Seq_Len, Dim) + (B, Seq_Len, Dim) --> (B, Seq_Len, Dim) | |
out = h + self.feed_forward.forward(self.ffn_norm(h)) |
No need to use forward method?
I mean, we could use nn.Module directly.
h = x + self.attention(self.attention_norm(x), start_pos, freqs_complex)
out = h + self.feed_forward(self.ffn_norm(h))
Metadata
Metadata
Assignees
Labels
No labels