Archives for Google research on memorizing transformers
By presenting a simple extension to the transformer, known as kNN-augmented attention, the research found that it could increase the length of the context in a language model.
By presenting a simple extension to the transformer, known as kNN-augmented attention, the research found that it could increase the length of the context in a language model.

