transformer/multi_head_attention library

Classes

MultiHeadAttention
Combines multiple self-attention heads to run in parallel.