Link: Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention

A new paper by researchers at Google claims to give large language models (LLMs) the ability to work with text of infinite length. The paper introduces Infini-attention, a technique that configures language models in a way that extends their “context window” while keeping memory and compute requirements constant. #


Yoooo, this is a quick note on a link that made me go, WTF? Find all past links here.