1. Attention is all you need;Vaswani;Adv. Neural Inf. Process. Syst.,2017
2. E. Egonmwan, Y. Chali, Transformer and seq2seq model for paraphrase generation, in: Proceedings of the Workshop on Neural Generation and Translation, 2019, pp. 249–255.
3. Fine-grained style control in transformer-based text-to-speech synthesis;Chen,2022
4. Emformer: Efficient memory transformer based acoustic model for low latency streaming speech recognition;Shi,2021
5. A structured self-attentive sentence embedding;Lin,2017