Output is added back to the block input through the residual path.
Diagram: Standard LLaMA transformer architecture – stacked blocks with
RMSNorm, multi-head self-attention with RoPE, and a SwiGLU MLP, drawn in a
style similar to the original figure.