Skip to content

Latest commit

 

History

History
11 lines (7 loc) · 338 Bytes

README.md

File metadata and controls

11 lines (7 loc) · 338 Bytes

COOL Attention

Attention layers are the basic components of the LLMs. The project focus on the basic implementation of those attention mechanism in system verilog.

Attention Mechanism

The attention mechanism can represented as follows,

$$ \text{Attention}(Q, K, V) = \text{softmax}\left(\frac{QK^T}{\sqrt{d_k}}\right) V $$