References
- Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, 2017, Attention Is All You Need: https://arxiv.org/abs/1706.03762
- Hugging Face Transformer Usage: https://huggingface.co/transformers/usage.html
- Manuel Romero Notebook with link to explanations by Raimi Karim: https://colab.research.google.com/drive/1rPk3ohrmVclqhH7uQ7qys4oznDdAhpzF
- Google language research: https://research.google/teams/language/
- Google Brain Trax documentation: https://trax-ml.readthedocs.io/en/latest/
- Hugging Face research: https://huggingface.co/transformers/index.html
- The Annotated Transformer: http://nlp.seas.harvard.edu/2018/04/03/attention.html
- Jay Alammar, The Illustrated Transformer: http://jalammar.github.io/illustrated-transformer/