Transformer block: communication followed by computation. More...
Public Member Functions | |
| __init__ (self, n_embd, n_head) | |
| forward (self, x) | |
Public Attributes | |
| sa = MultiHeadAttention(n_head, head_size) | |
| ffwd = FeedFoward(n_embd) | |
| ln1 = nn.LayerNorm(n_embd) | |
| ln2 = nn.LayerNorm(n_embd) | |
Transformer block: communication followed by computation.
Definition at line 142 of file working_gpt.py.
| working_gpt.Block.__init__ | ( | self, | |
| n_embd, | |||
| n_head ) |
Definition at line 145 of file working_gpt.py.
References __init__().
Referenced by __init__().
| working_gpt.Block.forward | ( | self, | |
| x ) |
| working_gpt.Block.ffwd = FeedFoward(n_embd) |
Definition at line 150 of file working_gpt.py.
Referenced by forward().
| working_gpt.Block.ln1 = nn.LayerNorm(n_embd) |
Definition at line 151 of file working_gpt.py.
Referenced by forward().
| working_gpt.Block.ln2 = nn.LayerNorm(n_embd) |
Definition at line 152 of file working_gpt.py.
Referenced by forward().
| working_gpt.Block.sa = MultiHeadAttention(n_head, head_size) |
Definition at line 149 of file working_gpt.py.
Referenced by forward().