Skip to main content
metafunctor
Research · Coding
Home
Posts
Series
Academic
Papers
Publications
Research
Creative
Writing
Media
Projects
About
Search
Home
/
Recommended Media
/
Generating Long Sequences with Sparse Transformers
Back to Media
Generating Long Sequences with Sparse Transformers
Child, Gray, Radford, Sutskever
paper
completed
ai-ml
Year
2019
External Link
https://arxiv.org/pdf/1904.10509
sparse attention
long sequences
efficiency
from:language-models
Notes
Sparse attention patterns for long-range dependencies. O(n√n) attention.
View Resource
All Media
More in Ai-Ml
More in Ai-Ml
An Introduction to Statistical Learning (ISLR2)
James, Witten, Hastie, Tibshirani · book
Artificial Intelligence: A Modern Approach
Russell & Norvig · book
Attention Is All You Need
Vaswani, Shazeer, Parmar, Uszkoreit, Jones, Gomez, Kaiser, Polosukhin · paper
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Devlin, Chang, Lee, Toutanova · paper