Transformer block: communication followed by computation. More...
Public Member Functions | |
__init__ (self, n_embd, n_head) | |
forward (self, x) | |
Public Attributes | |
sa = MultiHeadAttention(n_head, head_size) | |
ffwd = FeedFoward(n_embd) | |
ln1 = nn.LayerNorm(n_embd) | |
ln2 = nn.LayerNorm(n_embd) | |
Transformer block: communication followed by computation.
Definition at line 142 of file working_gpt.py.
working_gpt.Block.__init__ | ( | self, | |
n_embd, | |||
n_head ) |
Definition at line 145 of file working_gpt.py.
References __init__().
Referenced by __init__().
working_gpt.Block.forward | ( | self, | |
x ) |
working_gpt.Block.ffwd = FeedFoward(n_embd) |
Definition at line 150 of file working_gpt.py.
Referenced by forward().
working_gpt.Block.ln1 = nn.LayerNorm(n_embd) |
Definition at line 151 of file working_gpt.py.
Referenced by forward().
working_gpt.Block.ln2 = nn.LayerNorm(n_embd) |
Definition at line 152 of file working_gpt.py.
Referenced by forward().
working_gpt.Block.sa = MultiHeadAttention(n_head, head_size) |
Definition at line 149 of file working_gpt.py.
Referenced by forward().