原文
[Submitted on 25 Feb 2025 (v1), last revised 26 Feb 2025 (this version, v2)]
View a PDF of the paper titled The FFT Strikes Back: An Efficient Alternative to Self-Attention, by Jacob Fein-Ashley
View PDF HTML (experimental)Abstract:Conventional self-attention mechanisms incur quadratic complexity, limiting their scalability on long sequences. We introduce FFTNet, an adaptive spectral filtering framework that leverages the Fast Fourier Transform (FFT) to achieve global token mixing in $\mathcal{O}(n\log n)$ time. By transforming inputs into the frequency domain, FFTNet exploits the orthogonality and energy preservation guaranteed by Parseval's theorem to capture long-range dependencies efficiently. A learnable spectral filter and modReLU activation dynamically emphasize salient frequency components, providing a rigorous and adaptive alternative to traditional self-attention. Experiments on the Long Range Arena and ImageNet benchmarks validate our theoretical insights and demonstrate superior performance over fixed Fourier and standard attention models.
From: Jacob Fein-Ashley [view email]
[v1] Tue, 25 Feb 2025 17:43:43 UTC (366 KB)
[v2] Wed, 26 Feb 2025 16:31:58 UTC (367 KB)
[v1] Tue, 25 Feb 2025 17:43:43 UTC (366 KB)
[v2] Wed, 26 Feb 2025 16:31:58 UTC (367 KB)