Scaling Instruction-Finetuned Language Models

Andreas Stephan, 15. Dec 2022

A current paradigm in AI seems to be “The bigger the better”. In this work the authors explore scaling of tasks, instructions and chain of thought prompts in order to test the generalization capabilities for decoder- and encoder/decoder transformers on different types of tasks.

References: https://arxiv.org/abs/2210.11416