FlashAttention

FlashAttention

Update: 2025-03-05
Share

Description

FlashAttention is an IO-aware attention mechanism designed to be fast and memory-efficient, especially for long sequences. Its core innovation is tiling, where input sequences are divided into blocks processed within the fast on-chip SRAM, significantly reducing reads and writes to the slower HBM. This contrasts with standard attention, which materializes the entire attention matrix in HBM. By minimizing HBM access and recomputing the attention matrix in the backward pass, FlashAttention achieves faster Transformer training and a linear memory footprint, outperforming many approximate attention methods that overlook memory access costs.

Comments 
In Channel
Kimi K2

Kimi K2

2025-07-2215:30

MeanFlow

MeanFlow

2025-07-1006:47

Mamba

Mamba

2025-07-1008:14

LLM Alignment

LLM Alignment

2025-06-1420:06

Why We Think

Why We Think

2025-05-2014:20

Deep Research

Deep Research

2025-05-1211:35

vLLM

vLLM

2025-05-0413:06

DeepSeek-Prover-V2

DeepSeek-Prover-V2

2025-05-0111:04

DeepSeek-Prover

DeepSeek-Prover

2025-05-0108:37

Agent AI Overview

Agent AI Overview

2025-03-1721:06

FlashAttention-3

FlashAttention-3

2025-03-0713:43

FlashAttention-2

FlashAttention-2

2025-03-0510:50

FlashAttention

FlashAttention

2025-03-0510:55

loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

FlashAttention

FlashAttention

AI-Talk