Attention # Self-Attention & Multi-Head Attention Coming soon — Q/K/V, scaled dot-product, FlashAttention, FlexAttention.