Skip to content
Home ยป Revolutionizing Long Context Modeling: Titans and MIRAS from Transformers to Associative Memory

Revolutionizing Long Context Modeling: Titans and MIRAS from Transformers to Associative Memory

Google Research has introduced two innovative concepts called Titans and MIRAS, which aim to enhance the capabilities of sequence models by providing them with effective long-term memory. This development comes as a response to the limitations of existing models like Transformers, which struggle with maintaining context over long sequences.

Titans is a new architecture that integrates a deep neural memory with a Transformer-style backbone. It allows the model to learn and adapt during inference, enabling it to store important information while processing data. MIRAS, on the other hand, offers a broader framework that views many sequence models as forms of associative memory, focusing on how they learn and forget over time.

Traditional Transformers use attention mechanisms that help them learn from context but become inefficient as the context length increases. The cost of processing grows significantly, making it challenging to work with long sequences. While some models, like efficient recurrent neural networks, handle longer sequences better, they often lose valuable information in the process.

The Titans architecture addresses these issues by combining short-term memory through attention with a separate long-term memory module. This allows the model to maintain a precise focus on the current data while also storing and recalling relevant past information. The long-term memory is updated during the model’s operation, ensuring that it retains only the most surprising and impactful tokens.

Experimental results have shown that Titans outperforms other models on various benchmarks, including language modeling and commonsense reasoning tasks. It has been particularly effective in recalling information from extremely long contexts, surpassing even larger models like GPT-4 while using fewer parameters.

MIRAS expands on these ideas by providing a unified framework for understanding sequence models as associative memories. It breaks down the design of these models into four key components: memory structure, attentional bias, retention gate, and memory algorithm. This perspective allows for the creation of new attention-free models that still perform well on language tasks.

Overall, Titans and MIRAS represent significant advancements in the field of sequence modeling, offering solutions to long-standing challenges in maintaining context and memory in AI systems. As researchers continue to explore these concepts, the potential for more efficient and capable AI models appears promising.