This project aims to implement the Scaled-Dot-Product Attention layer and the Multi-Head Attention layer using various Positional Encoding methods.
-
Updated
Jun 27, 2022 - Python
This project aims to implement the Scaled-Dot-Product Attention layer and the Multi-Head Attention layer using various Positional Encoding methods.
A framework for training and evaluating a transformer with scaled dot product attention on a tensorflow dataset.
Add a description, image, and links to the scaled-dot-product topic page so that developers can more easily learn about it.
To associate your repository with the scaled-dot-product topic, visit your repo's landing page and select "manage topics."