1. Language models are few-shot learners;Brown,2020
2. What does bert look at? an analysis of bert’s attention;Clark,2019
3. Clark, Christopher, Lee, Kenton, Chang, Ming-Wei, Kwiatkowski, Tom, Collins, Michael, Toutanova, Kristina, 2019a. BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). pp. 2924–2936.
4. The pascal recognising textual entailment challenge;Dagan,2005
5. Transformer-xl: Attentive language models beyond a fixed-length context;Dai,2019