<img height="1" width="1" style="display:none" src="https://www.facebook.com/tr?id=145304570664993&amp;ev=PageView&amp;noscript=1">

91ƵAPP

Papers wide (2)

Jul 04, 2024

Mamba-2 & Matmul-free Models: June Papers of the Month

Written By:

Luke Prince, Charlie Blake, Alberto Cattaneo, Douglas Orr

We're Hiring

Join us and build the next generation AI stack - including silicon, hardware and software - the worldwide standard for AI compute

Join our team

Improving transformers is now not “just one area” of machine learning research. This is illustrated by the breadth of papers we got excited about this month, all of which claim to improve upon some aspect of the transformer, but in very different ways.

First, Mamba-2 explores the connection between structured state space models and attention, resulting in a new architecture, Mamba-2. (The paper isn’t short, so you get value-for-money with this summary!)

SµPar builds upon the maximal update parameterisation to transfer hyperparameters across different sparsity levels, promising predictable training of sparse models.

CoPE identifies deficiencies in current relative positional encodings, which are critical for turning transformers from set models into sequence models, and introduces a new & richer form of encoding.

Finally, “matmul-free LMs” follow the trajectory of BitNet and BitNet b1.58, removing all matrix multiplies from a transformer LM forward pass (in doing so, they make it an RNN), promising compression & compute efficiency.

I hope you enjoy these as much as we did. If you have thoughts or questions, keep the conversation going on X.

Here’s our summary of this month’s chosen papers:

Authors: Tri Dao and Albert Gu (Princeton University and Carnegie Mellon University)

The key idea

If you stare for a while at the mapping from SSM inputs to outputs, it looks a lot like a more expressive form of causal linear attention. If you simplify this general form, you get a new SSM building block for Mamba-2 that scales across devices better than the original, while improving performance on associative recall tasks and maintaining language modelling performance comparable to Transformers up to 2.7B parameters.

FIG-Schema (1)

The authors demonstrate equivalence between linear attention and state space models under specific conditions for attention masking structure and state-space rank.

Background

State-space models have been used for decades to model continuous-time signals. In their simplest form, they are just a linear time-invariant system mapping some hidden variable ℎ𝑡 that to an observed variable 𝑦𝑡, where ℎ𝑡 acts as a filter over some input signal 𝑥𝑡, i.e.,

𝑡=𝐴ℎ𝑡−1+𝐵𝑥𝑡
𝑦𝑡=𝐶𝑡

Where 𝐴,𝐵,𝐶, are matrices parameterising the system in question.

Structured state-space models of sequences in deep learning, e.g. Mamba, make 𝐴,𝐵,𝐶 matrices change over time and depend on the input.

The authors note that by induction

Screenshot 2024-07-04 15 00 42-1

Which can be vectorised and written as a matrix multiplication 𝑦=𝑀𝑥 mapping sequence 𝑦 to sequence 𝑥 via matrix 𝑀 where

Screenshot 2024-07-04 15 00 48

This makes 𝑀 a semiseparable matrix.

FIG-SSM (1)

They also show that masked linear attention can also be written in this form, i.e., 𝑦=(𝐿∘𝑄𝐾)𝑣=𝑀𝑣, and that for a choice of structure for 𝐿 (to make 𝑀 semiseparable), and A, these can be made equivalent. In particular, when 𝐴 is of the scalar-diagonal form 𝑎𝐼, it is equivalent to a weighted cumulative sum where weights are the product of 𝑎 over time, and when 𝐿 is a multiplicative causal mask with a relative positional encoding.

FIG-Attention (1)

Using these insights, the authors propose the fundamental building block of Mamba-2: the State-Space Dual (SSD) layer, connecting state-space models to linear attention.

Their method

By restricting A to be in scalar-diagonal form, the general SSM layer of Mamba can be drastically simplified, permitting a more straightforward implementation that targets GPU tensor cores (improving throughput) via batched matrix multiplications rather than scan operations.

ALGO-SSD (1)

This also reduces space complexity to be linear (rather than quadratic) in hidden size.

TBL-Complexity-1

They also add a normalisation layer before the final output projection as Mamba was found to be unstable with larger state sizes, and permit B and C to be shared across heads, analogous to multi-/grouped-value attention.

FIG-SSD-1

By allowing convolution blocks and normalisation to be grouped within devices, this block can be parallelised more easily, requiring only an all-reduce for the final output projection. SSD states may also be passed sequentially between devices to allow sequences to be split across devices.

FIG-Parallelisation

Results

The authors demonstrate improved associative recall on synthetic tasks due to larger state size permitted by improved space complexity. Mamba-2 could still hit a wall when model state capacity is saturated over longer sequence lengths however, while attention should be more robust.

FIG-Recall

It’s faster than flash attention at 2k sequence length (Mamba only at 16k), and remains faster with larger state sizes due to memory requirements linear with state size (quadratic with Mamba).

FIG-Throughput

Scaling laws look comparable to Llama-like transformers (Transformer++) up to 1.3B. It is possibly worse at 2.7B as they appear to have trained up to this size but only compare to Pythia rather than Transformer++. Extrapolating from the scaling laws figure, it looks as though there might be a crossover soon after 1.3B.

FIG-Scaling

The authors also explore hybrid models blending SSD layers with attention and MLPs, demonstrating that 10% attention layers works well. released a few weeks after this one conducts a more thorough investigation.

Takeaways

The release of the original Mamba paper sparked a wave of exploration into applying Mamba to various domains and tweaking the architecture to improve performance, enabled in particular by open-sourcing optimised kernels for the scan operation at the heart. By removing the need for this more sophisticated scan operation entirely and instead being able to rely on batched matrix multiplications for acceleration, we can also speed up the cycle of experimentation. We might expect to see further improvements to this architecture at increasing scale.

Full paper:

Authors: Nolan Dey, Shane Bergsma, Joel Hestness (Cerebras Systems)

The key idea

Introducing sparsity into a model causes its learning dynamics to change, meaning that the optimal hyperparameters (especially learning rate) may be altered. Many previous studies have failed to account for this effect. By applying the principles of µP we can create sparse models with stable hyperparameters.

figure_1 (1)-1

Background

has suggested that a simple and quite effective approach to sparsity is to prune a random selection of weight elements at initialisation. This kind of sparsity is known to affect learning dynamics, yet the community has largely persisted with fixing the dense LR across different sparsity levels, largely due to the expense of re-sweeping it with every sparsity change.

Recently µP has emerged as a method for ensuring consistent learning dynamics across models of different widths. This works by adjusting multipliers in the model, initialisation and learning rate such that activations and updates are invariant to changes in width.

Their method

Applying sparsity to a matrix multiplication is comparable to a matrix multiplication of reduced width. The authors leverage this to come up with a version of µP for sparsity, named SµPar (‘soo-pahr’). Where 𝑚 is the model width and 𝑚𝜌 the model density (1 - sparsity), their rules are:

table_1

The paper contains some mathematical justification for this, which can make the method seem quite complex, but in reality the above table reflects the simplicity of the method: use µP with your sparse model as though it had its “effective width” of 𝑚⋅𝑚𝜌 for the sparse layers.

Their aim is that

desiderata

meaning they can scale with width and sparsity. In this way, they can make their models both wider and sparser at once and still keep the same learning dynamics.

Results

Based on the fact they can scale width and sparsity simultaneously, they show the effect under this sparsity setup of keeping the number of non-sparse parameters fixed, and scaling the number of total parameters in combination with sparsity:

figure_9

This is a neat result. The LR is much more stable, and there are clear gains to be had from increased sparsity (with a constant memory footprint).

The only negative here really is that the benefit is not particularly substantial. They plot the final validation loss for their full LLM training runs at different densities and even for the aggressive 1/128 sparsity the gain is relatively modest:

figure_3

Takeaways

Nevertheless, this approach to stable sparsity is principled and worth adopting. One can see it being particularly useful for more extreme or unusual forms of sparsity, where the hyperparameters may shift further. It also has a lot of overlap with the recent paper, which also uses µP-style rules for different kinds of structured matrices. In general, the idea of controlling for learning dynamics when testing new ideas seems like it could become standard in years ahead.

Full paper:

Authors: Olga Golovneva, et al. (Meta (FAIR))

The key idea

Transformers rely on Position Encoding (PE) to inject information about the position of tokens in a sequence into the attention block, which by construction is order-invariant. This paper proposes Contextual Position Encoding (CoPE), a flexible, context-dependent technique for measuring positional distances at higher abstraction levels than just counting tokens, improving performance on language modelling and addressing common failures of LLMs in counting-based tasks.

cope_overview (1)

Background

PE introduces learnable position embeddings 𝐞𝑖, into the attention block

Screenshot 2024-07-23 at 12.30.50

with 𝐞𝑖,=𝐞[𝑖]for Absolute PE, or 𝐞𝑖,=𝐞[𝑖−]in the case of Relative PE. A clear limitation of this setup is that positions are always measured in terms of tokens, which - depending on the task - might not be the best unit of measure. For instance, state-of-the-art LLMs (like GPT4) are observed to often fail at simple counting tasks that require them to attend only to tokens or words within specific chunks of text, like sentences or paragraphs, that can have highly variable lengths.

Their method

In CoPE, the distance of token with respect to the query position 𝑖(with <𝑖) is measured based on the context of the intermediate tokens, through a soft gate:

Screenshot 2024-07-23 at 12.30.58

In the attention computation we then use 𝐞𝑖,=ư𝑝𝑖,] or, in the case where 𝑝𝑖, is not an integer, an interpolation of 𝐞[⌊𝑝𝑖,⌋] and 𝐞[⌈𝑝𝑖,.

Relative PE can be seen as the limit case where all 𝑔𝑖,𝑡=1. More generally though, thanks to context-awareness, 𝑝𝑖, could be the count of a specific word, or the number of sentences, between token positions and 𝑖, or any other measure that the model finds useful to track. Note that, by construction, each attention head and each layer will compute a different 𝑝𝑖,, thus allowing the model to represent different levels of position abstraction at the same time.

cope_contextualised_attention (1)

Results

CoPE is tested on a variety of artificial tasks (like selective copying/counting and the Flip-Flop task) where standard PE methods perform poorly. For all of them, CoPE yields strong improvements and better generalisation to out-of-distribution data.

cope_results (1)

Moreover, CoPE improves in perplexity over Relative PE for language and code modelling with small Transformers (20M-100M parameters), also showing better generalisation to longer contexts than the ones seen in training.

cope_ppl (1)

Takeaways

Despite the limited scale of experiments, the results show a promising step in the direction of making the reasoning and abstraction abilities of Transformers even more flexible. It will be interesting to see how CoPE performs on larger models, and quantify the trade-off between performance gains and additional computation costs on real-world downstream tasks.

Full paper:

Authors: Rui-Jie Zhu, et al. (University of California Santa Cruz, Soochow University, University of California Davis, LuxiTech)

The key idea

Building upon , which quantises all parameter matrices in a LM into a ternary format, the authors describe a “matmul-free” LM where all forward pass matrix multiplies are ternary.

The authors achieve this by replacing self-attention with a structured-recurrence RNN, which contains only parametric matmuls and elementwise operations, and replace these parametric matmuls with ternary matmuls (shown below).

ternary-matmul (1)

Their method

Following BitNet b1.58, forward-pass weights are quantised to {-1, 0, +1} using absmean quantisation, and activations to int8 using absmax quantisation:
algorithm (1)
In the backward pass, the straight-through estimator replaces these with the identity function, such that the weight gradient and master weights are maintained in higher precision.

The authors replace attention with the Matmul-free Linear Gated Recurrent Unit (MLGRU),

mlgru (1)

The MLGRU maps a sequence of inputs 𝒙𝑡 to a sequence of outputs 𝒐𝑡. First, compute three gates: forget gate 𝒇, output gate 𝒈 and candidate 𝒄, which are ternary-weight projections of the input with sigmoid, sigmoid and SiLU nonlinearities respectively. Then use the forget gate to interpolate between the previous hidden state 𝒉 and the candidate. Finally, use the output gate to mask the hidden state before projecting via a final ternary matmul.

FPGA implementation

While ternary weights provide an advantage of reducing memory transfers when running on modern ML hardware, they are not supported by matrix compute units, so the energy benefits of ternary quantisation are not realised. To illustrate the potential of the matmul-free LM, the authors implement a custom FPGA accelerator in SystemVerilog, implementing a small special-purpose instruction set. They deploy the RTL on a D5005 Stratix 10, which runs a width-512 single-layer forward pass in 43 ms.

While the authors acknowledge that this is a limited and preliminary result, their extrapolations to incorporate bursting over the DDR4 interface, using vendor IP and adding pipelining show promise (24 tokens/s of a 1.3B model at 13 W). The number of cores may also be increased, yielding higher throughput (and power).

Results

Results compare well against a Transformer++ baseline when trained on SlimPajama. The limited training duration makes it hard to compare the baselines with state-of-the-art LMs trained on this dataset, but the baseline is competitive with that of BitNet b1.58.

zeroshot (1)

Takeaways

We’re excited to see this line of work continue, as it challenges our preconceptions regarding continuous optimisation in deep learning and offers the promise of reaching new levels of practical efficiency.

Full paper:

Discover more on the , and subscribe to the Papers of the Month newsletter.