23 Feb Large tech firms and the language models they associate with Avi Gopani bert gopher gpt-3 switch transformer gopher glam wu dao 2.0 turing nlg macaw jurassic Wu Dao 2.0 is built on 1.75 trillion parameters, easily surpassing GPT-3 or Google’s Switch Transformer.