[NeurIPS 2025] Official code for "Tropical Attention: Neural Algorithmic Reasoning for Combinatorial Algorithms"
-
Updated
Oct 23, 2025 - Python
[NeurIPS 2025] Official code for "Tropical Attention: Neural Algorithmic Reasoning for Combinatorial Algorithms"
This work provides extensive empirical results on training LMs to count. We find that while traditional RNNs trivially achieve inductive counting, Transformers have to rely on positional embeddings to count out-of-domain. Modern RNNs (e.g. rwkv, mamba) also largely underperform traditional RNNs in generalizing counting inductively.
Mamba Modulation: On the Length Generalization of Mamba Models
Add a description, image, and links to the length-generalization topic page so that developers can more easily learn about it.
To associate your repository with the length-generalization topic, visit your repo's landing page and select "manage topics."