Synthesis Blog

Together, we’re building the future of computer vision and machine learning.
Featured Post

Linear Attention and Mamba: New Power to Old Ideas

We have already discussed how to extend the context size for modern Transformer architectures, but today we explore a different direction of this research. In the quest to handle longer sequences and larger datasets, Transformers are turning back to the classics: the memory mechanisms of RNNs, associative memory, and even continuous dynamical systems. From linear attention to Mamba, modern models are blending old and new ideas to bring forth a new paradigm of sequence modeling, and this paradigm is exactly what we discuss today.

Continue reading
All Posts
November 20, 2024

We have already discussed how to extend the context size for modern Transformer architectures, but today we explore a different…

November 7, 2024

Although deep learning is a very new branch of computer science, foundations of neural networks have been in place since…

September 25, 2024

OpenAI’s o1-preview has been all the buzz lately. While this model is based on the GPT-4o general architecture, it boasts…

September 18, 2024

We continue our series on LLMs and various ways to make them better. We have already discussed ways to increase…

August 13, 2024

We continue our series on generative AI. We have discussed Transformers, large language models, and some specific aspects of Transformers…

July 2, 2024

One of the most striking AI advances this spring was OpenAI's Sora, a video generation model that sets new standards…

All Series