The total compute used to tune GPT-3 turned out to be a mere 7 per cent of the compute used to pretrain the model.