Reformer matches standard Transformer accuracy on long sequences while using far less memory and running faster via LSH attention and reversible residual layers.
Sainbayar Sukhbaatar, Edouard Grave, Piotr Bojanowski, and Armand Joulin
1 Pith paper cite this work. Polarity classification is still indexing.
1
Pith paper citing it
fields
cs.LG 1years
2020 1verdicts
ACCEPT 1representative citing papers
citing papers explorer
-
Reformer: The Efficient Transformer
Reformer matches standard Transformer accuracy on long sequences while using far less memory and running faster via LSH attention and reversible residual layers.