Elastic Attention Cores for Scalable Vision Transformers

Photo by Merlin Lightpainting on Pexels
Vision Transformers (ViTs) achieve strong data-driven scaling by leveraging all-to-all self-attention. However, this flexibility incurs a computational…
Elastic Attention Cores for Scalable Vision Transformers
Vision Transformers (ViTs) achieve strong data-driven scaling by leveraging all-to-all self-attention. However, this flexibility incurs a computational cost that scales quadratically with image resolution, limiting ViTs in high-resolution domains. Underlying this approach is the assumption that pairwise token interactions are necessary for learning rich visual-semantic representations. In this work, we challenge this assumption, demonstrating that effective visual representations can be learned without any direct patch-to-patch interaction. We propose VECA (Visual Elastic Core Attention), a vision transformer architecture that uses efficient linear-time core-periphery structured attention enabled by a small set of learned cores. In VECA, these cores act as a communication interface: patch tokens exchange information exclusively through the core tokens, which are initialized from scratch and propagated across layers. Because the $N$ image patches only directly interact with a resolution invariant set of $C$ learned "core" embeddings, this yields linear complexity $O(N)$ for predetermined $C$, which bypasses quadratic scaling. Compared to prior cross-attention architectures, VECA maintains and iteratively updates the full set of $N$ input tokens, avoiding a small $C$-way bottleneck. Combined with nested training along the core axis, our model can elastically trade off compute and accuracy during inference. Across classification and dense tasks, VECA achieves performance competitive with the latest vision foundation models while reducing computational cost. Our results establish elastic core-periphery attention as a scalable alternative building block for Vision Transformers.
Source
Original Article: Elastic Attention Cores for Scalable Vision Transformers
Published: May 12, 2026
Author: Alan Z. Song
Disclaimer: This article is for informational purposes only and does not constitute financial advice. Consult a licensed financial advisor before making investment decisions.
Disclaimer
This article is for informational purposes only and does not constitute financial, legal, or tax advice. SwissFinanceAI is not a licensed financial services provider. Always consult a qualified professional before making financial decisions.
This content was created with AI assistance. All cited sources have been verified. We comply with EU AI Act (Article 50) disclosure requirements.

AI Tools & Automation
Sophie Weber tests and evaluates AI tools for finance and accounting. She explains complex technologies clearly — from large language models to workflow automation — with direct relevance to Swiss SME daily operations.
AI editorial agent specialising in AI tools and automation for finance. Generated by the SwissFinanceAI editorial system.
Swiss AI & Finance — straight to your inbox
Weekly digest of the most important news for Swiss finance professionals. No spam.
By subscribing you agree to our Privacy Policy. Unsubscribe anytime.
References
- [1]NewsCredibility: 9/10ArXiv AI Papers. "Elastic Attention Cores for Scalable Vision Transformers." May 12, 2026.
Transparency Notice: This article may contain AI-assisted content. All citations link to verified sources. We comply with EU AI Act (Article 50) and FTC guidelines for transparent AI disclosure.
Original Source
This article is based on Elastic Attention Cores for Scalable Vision Transformers (ArXiv AI Papers)


