cover

How Mamba and Hyena Are Changing the Way AI Learns and Remembers

18 Dec 2024

Explore detailed experimental results from cutting-edge AI architectures, including Selective Copying, Induction Heads, and innovative models like Mamba and Hye

cover

Hardware-Aware Algorithm for Selective State Space Models

18 Dec 2024

Selective SSMs achieve up to 7× faster processing than attention and optimal memory efficiency with kernel fusion and recomputation on modern GPUs.

cover

The Math Behind Selective State Space Models

18 Dec 2024

Explore the mechanics of Selective SSMs, focusing on discretization, learnable biases, and zero-order hold formulas for efficient AI modeling.

cover

AI That Remembers

18 Dec 2024

Discover how new AI models handle longer tasks, improve learning, and bring smarter, faster performance to the next generation of technology.

cover

Why Selection Mechanisms Are Key to the Future of Sequence Modeling

18 Dec 2024

Explore the concept of selection mechanisms in neural networks, distinguishing them from traditional gating, hypernetworks, and data-dependence.

cover

Why Mamba Could Be the Future of Big Data Processing

18 Dec 2024

Mamba introduces a selection mechanism to structured state space models, enabling context-dependent reasoning with linear scaling in sequence length.

cover

Why Scaling Mamba Beyond Small Models Could Lead to New Challenges

17 Dec 2024

Mamba introduces transformative concepts in language modeling with its selective SSMs, but the scaling of these models faces engineering challenges.

cover

How Selective State Space Models Boost Mamba’s Performance

17 Dec 2024

Mamba’s architecture is enhanced by the selective SSM layer, yielding significant performance improvements.

cover

How Mamba’s Design Makes AI Up to 40x Faster

17 Dec 2024

Mamba outperforms FlashAttention-2 and Transformers in both speed and memory efficiency, offering up to 20-40x faster inference throughput.