What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?

Summary

This is a publication. If there is no link to the publication on this page, you can try the pre-formated search via the search engines listed on this page.

Authors: Thomas Wang, Adam Roberts, Daniel Hesslow, Teven Le Scao, Hyung Won Chung, Iz Beltagy, Julien Launay, Colin Raffel

Journal title: Proceedings of the 39 th International Conference on Machine Learning

Journal publisher: MLR press

Published year: 2022

ISSN: 2640-3498