A Faster Pytorch Implementation of Multi-Head Self-Attention
attention attention-mechanism multihead-attention self-attention multi-head-attention multi-head multi-head-self-attention multihead-self-attention transformer-attention pytorch-self-attention
-
Updated
May 27, 2022 - Jupyter Notebook