arXivNotes
arXivNotes copied to clipboard
🚧 2019: Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context
Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context Zihang Dai, Zhilin Yang, Yiming Yang, Jaime Carbonell, Quoc V. Le, Ruslan Salakhutdinov ACL 2019 long paper. Code and pretrained models are available at this https URL https://arxiv.org/abs/1901.02860