Close Menu
    Trending
    • “The success of an AI product depends on how intuitively users can interact with its capabilities”
    • How to Crack Machine Learning System-Design Interviews
    • Music, Lyrics, and Agentic AI: Building a Smart Song Explainer using Python and OpenAI
    • An Anthropic Merger, “Lying,” and a 52-Page Memo
    • Apple’s $1 Billion Bet on Google Gemini to Fix Siri
    • Critical Mistakes Companies Make When Integrating AI/ML into Their Processes
    • Nu kan du gruppchatta med ChatGPT – OpenAI testar ny funktion
    • OpenAI’s new LLM exposes the secrets of how AI really works
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » We Didn’t Invent Attention — We Just Rediscovered It
    Artificial Intelligence

    We Didn’t Invent Attention — We Just Rediscovered It

    ProfitlyAIBy ProfitlyAINovember 5, 2025No Comments11 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    , somebody claims they’ve invented a revolutionary AI structure. However whenever you see the identical mathematical sample — selective amplification + normalization — emerge independently from gradient descent, evolution, and chemical reactions, you understand we didn’t invent the eye mechanism with the Transformers structure. We rediscovered basic optimization rules that govern how any system processes info underneath power constraints. Understanding consideration as amplification slightly than choice suggests particular architectural enhancements and explains why present approaches work. Eight minutes right here offers you a psychological mannequin that might information higher system design for the subsequent decade.

    When Vaswani and colleagues revealed “Consideration Is All You Want” in 2017, they thought they have been proposing one thing revolutionary [1]. Their transformer structure deserted recurrent networks totally, relying as a substitute on consideration mechanisms to course of whole textual content sequences concurrently. The mathematical core was easy: compute compatibility scores between positions, convert them to weights, and use these for selective mixture of data.

    However this sample seems to emerge independently wherever info processing programs face useful resource constraints underneath complexity. Not as a result of there’s some common regulation of consideration, however as a result of sure mathematical buildings appear to characterize convergent options to basic optimization issues.

    We could also be taking a look at a kind of uncommon circumstances the place biology, chemistry, and AI have converged on comparable computational methods — not by way of shared mechanisms, however by way of shared mathematical constraints.

    The five hundred-Million-Yr Experiment

    The organic proof for attention-like mechanisms is remarkably deep. The optic tectum/superior colliculus system, which implements spatial consideration by way of aggressive inhibition, exhibits extraordinary evolutionary conservation throughout vertebrates [2]. From fish to people, this neural structure maintains structural and purposeful consistency throughout 500+ million years of evolution.

    However maybe extra intriguing is the convergent evolution.

    Impartial lineages developed attention-like selective processing a number of occasions: compound eye programs in bugs [3], digicam eyes in cephalopods [4], hierarchical visible processing in birds [5], and cortical consideration networks in mammals [2]. Regardless of vastly totally different neural architectures and evolutionary histories, these programs converged on comparable options for selective info processing.

    This raises a compelling query: Are we seeing proof of basic computational constraints that govern how complicated programs should course of info underneath useful resource limitations?

    Even easy organisms counsel this sample scales remarkably. C. elegans, with solely 302 neurons, demonstrates subtle attention-like behaviors in meals looking for and predator avoidance [6]. Vegetation exhibit attention-like selective useful resource allocation, directing progress responses towards related environmental stimuli whereas ignoring others [7].

    The evolutionary conservation is putting, however we must be cautious about direct equivalences. Organic consideration includes particular neural circuits formed by evolutionary pressures fairly totally different from the optimization landscapes that produce AI architectures.

    Consideration as Amplification: Reframing the Mechanism

    Current theoretical work has basically challenged how we perceive consideration mechanisms. Philosophers Peter Fazekas and Bence Nanay demonstrated that conventional “filter” and “highlight” metaphors basically mischaracterize what consideration truly does [8].

    They assert that focus doesn’t choose inputs — it amplifies presynaptic indicators in a non-stimulus-driven manner, interacting with built-in normalization mechanisms that create the looks of choice. The mathematical construction they determine is the next:

    1. Amplification: Enhance the energy of sure enter indicators
    2. Normalization: Constructed-in mechanisms (like divisive normalization) course of these amplified indicators
    3. Obvious Choice: The mixture creates what seems to be selective filtering
    Determine 1: Consideration doesn’t filter inputs — it amplifies sure indicators, then normalization creates obvious selectivity. Like an audio mixer with computerized achieve management, the consequence appears selective however the mechanism is amplification. Picture by creator.

    This framework explains seemingly contradictory findings in neuroscience. Results like elevated firing charges, receptive area discount, and encompass suppression all emerge from the identical underlying mechanism — amplification interacting with normalization computations that function independently of consideration.

    Fazekas and Nanay targeted particularly on organic neural programs. The query of whether or not this amplification framework extends to different domains stays open, however the mathematical parallels are suggestive.

    Chemical Computer systems and Molecular Amplification

    Maybe essentially the most shocking proof comes from chemical programs. Baltussen and colleagues demonstrated that the formose response — a community of autocatalytic reactions involving formaldehyde, dihydroxyacetone, and metallic catalysts — can carry out subtle computation [9].

    Determine 2. A Chemical Pc in Motion: Combine 5 easy chemical compounds in a stirred reactor, and one thing outstanding occurs — the chemical soup learns to acknowledge patterns, predict future modifications, and type info into classes. No programming, no coaching, no silicon chips. Simply molecules doing math. This formose response community processes info utilizing the identical selective amplification rules that energy ChatGPT’s consideration mechanism, but it surely advanced naturally by way of chemistry alone. Picture by creator.

    The system exhibits selective amplification throughout as much as 10⁶ totally different molecular species, reaching > 95% accuracy on nonlinear classification duties. Totally different molecular species reply differentially to enter patterns, creating what seems to be chemical consideration by way of selective amplification. Remarkably, the system operates on timescales (500 ms to 60 minutes) that overlap with organic and synthetic consideration mechanisms.

    However the chemical system lacks the hierarchical management mechanisms and studying dynamics that characterize organic consideration. But the mathematical construction — selective amplification creating obvious selectivity — seems strikingly comparable. Programmable autocatalytic networks present further proof. Steel ions like Nd³⁺ create biphasic management mechanisms, each accelerating and inhibiting reactions relying on focus [10]. This generates controllable selective amplification that implements Boolean logic features and polynomial mappings by way of purely chemical processes.

    Data-Theoretic Constraints and Common Optimization

    The convergence throughout these totally different domains could replicate deeper mathematical requirements. Data bottleneck concept gives a proper framework: any system with restricted processing capability should resolve the optimization drawback of minimizing info retention whereas preserving task-relevant particulars [11].

    Jan Karbowski’s work on info thermodynamics reveals common power constraints on info processing [12]. The elemental thermodynamic sure on computation creates choice stress for environment friendly selective processing mechanisms throughout all substrates able to computation:

    Data processing prices power, so environment friendly consideration mechanisms have a survival/efficiency benefit, the place σ represents entropy (S) manufacturing fee, and ΔI represents info processing capability.

    Each time any system — whether or not a mind, a pc, and even chemical reactions — processes info, it should dissipate power as waste warmth. The extra info you course of, the extra power you will need to waste. Since consideration mechanisms course of info (deciding what to deal with), they’re topic to this power tax.

    This creates common stress for environment friendly architectures — whether or not you’re evolution designing a mind, chemistry organizing reactions, or gradient descent coaching transformers.

    Neural networks working at criticality — the sting between order and chaos — maximize info processing capability whereas sustaining stability [13]. Empirical measurements present that aware consideration in people happens exactly at these vital transitions [14]. Transformer networks throughout coaching exhibit comparable section transitions, organizing consideration weights close to vital factors the place info processing is optimized [15].

    This means the chance that attention-like mechanisms could emerge wherever programs face the elemental trade-off between processing capability and power effectivity underneath useful resource constraints.

    Convergent Arithmetic, Not Common Mechanisms

    The proof factors towards a preliminary conclusion. Reasonably than discovering common mechanisms, we could also be witnessing convergent mathematical options to comparable optimization issues:

    The mathematical construction — selective amplification mixed with normalization — seems throughout these domains, however the underlying mechanisms and constraints differ considerably.

    For transformer architectures, this reframing suggests particular insights:

    • Q·Ok computation implements amplification.

    The dot product Q·Ok^T computes semantic compatibility between question and key representations, performing as a realized amplification operate the place excessive compatibility scores amplify sign pathways.The scaling issue √d_k prevents saturation in high-dimensional areas, sustaining gradient movement.

    • Softmax normalization creates winner-take-all dynamics

    Softmax implements aggressive normalization by way of divisive renormalization. The exponential time period amplifies variations (winner-take-all dynamics) whereas sum normalization ensures Σw_ij = 1. Mathematically this operate is equal to a divisive normalization.

    • Weighted V mixture produces obvious selectivity

    On this mixture there may be not specific choice operator, it’s mainly a linear mixture of worth vectors. The obvious selectivity emerges from the sparsity sample induced by softmax normalization. Excessive consideration weights create efficient gating with out specific gating mechanisms.

    The mixture softmax(amplification) induce a winner-take-all dynamics on the worth area.

    Implications for AI Improvement

    Understanding consideration as amplification + normalization slightly than choice gives a number of sensible insights for AI structure design:

    • Separating Amplification and Normalization: Present transformers conflate these mechanisms. We’d discover architectures that decouple them, permitting for extra versatile normalization methods past softmax [16].
    • Non-Content material-Primarily based Amplification: Organic consideration contains “not-stimulus-driven” amplification. Present transformer consideration is solely content-based (Q·Ok compatibility). We might examine realized positional biases, task-specific amplification patterns, or meta-learned amplification methods.
    • Native Normalization Swimming pools: Biology makes use of “swimming pools of surrounding neurons” for normalization slightly than international normalization. This means exploring native consideration neighborhoods, hierarchical normalization throughout layers, or dynamic normalization pool choice.
    • Crucial Dynamics: The proof for consideration working close to vital factors means that efficient consideration mechanisms ought to exhibit particular statistical signatures — power-law distributions, avalanche dynamics, and demanding fluctuations [17].

    Open Questions and Future Instructions

    A number of basic questions stay:

    1. How deep do the mathematical parallels prolong? Are we seeing true computational equivalence or superficial similarity?
    2. What can chemical reservoir computing train us about minimal consideration architectures? If easy chemical networks can obtain attention-like computation, what does this counsel in regards to the complexity necessities for AI consideration?
    3. Do information-theoretic constraints predict the evolution of consideration in scaling AI programs? As fashions turn out to be bigger and face extra complicated environments, will consideration mechanisms naturally evolve towards these common optimization rules?
    4. How can we combine organic insights about hierarchical management and adaptation into AI architectures? The hole between static transformer consideration and dynamic organic consideration stays substantial.

    Conclusion

    The story of consideration seems to be much less about invention and extra about rediscovery. Whether or not within the formose response’s chemical networks, the superior colliculus’s neural circuits, or transformer architectures’ realized weights, we see variations on a mathematical theme: selective amplification mixed with normalization to create obvious selectivity.

    This doesn’t lower the achievement of transformer architectures — if something, it suggests they characterize a basic computational perception that transcends their particular implementation. The mathematical constraints that govern environment friendly info processing underneath useful resource limitations seem to push totally different programs towards comparable options.

    As we proceed scaling AI programs, understanding these deeper mathematical rules could show extra invaluable than mimicking organic mechanisms straight. The convergent evolution of attention-like processing suggests we’re working with basic computational constraints, not engineering selections.

    Nature spent 500 million years exploring these optimization landscapes by way of evolution. We rediscovered comparable options by way of gradient descent in a number of years. The query now’s whether or not understanding these mathematical rules can information us towards even higher options that transcend each organic and present synthetic approaches.

    Closing observe

    The actual take a look at: if somebody reads this and designs a greater consideration mechanism because of this, we’ve created worth.


    Thanks for studying — and sharing!

    Javier Marin
    Utilized AI Advisor | Manufacturing AI Methods + Regulatory Compliance
    [email protected]


    References

    [1] Vaswani, A., et al. (2017). Consideration is all you want. Advances in Neural Data Processing Methods, 30, 5998–6008.

    [2] Knudsen, E. I. (2007). Basic parts of consideration. Annual Overview of Neuroscience, 30, 57–78.

    [3] Nityananda, V., et al. (2016). Consideration-like processes in bugs. Proceedings of the Royal Society B, 283(1842), 20161986.

    [4] Cartron, L., et al. (2013). Visible object recognition in cuttlefish. Animal Cognition, 16(3), 391–401.

    [5] Wylie, D. R., & Crowder, N. A. (2014). Avian fashions for 3D scene evaluation. Proceedings of the IEEE, 102(5), 704–717.

    [6] Jang, H., et al. (2012). Neuromodulatory state and intercourse specify various behaviors by way of antagonistic synaptic pathways in C. elegans. Neuron, 75(4), 585–592.

    [7] Trewavas, A. (2009). Plant behaviour and intelligence. Plant, Cell & Atmosphere, 32(6), 606–616.

    [8] Fazekas, P., & Nanay, B. (2021). Consideration is amplification, not choice. British Journal for the Philosophy of Science, 72(1), 299–324.

    [9] Baltussen, M. G., et al. (2024). Chemical reservoir computation in a self-organizing response community. Nature, 631(8021), 549–555.

    [10] Kriukov, D. V., et al. (2024). Exploring the programmability of autocatalytic chemical response networks. Nature Communications, 15(1), 8649.

    [11] Tishby, N., & Zaslavsky, N. (2015). Deep studying and the data bottleneck precept. arXiv preprint arXiv:1503.02406.

    [12] Karbowski, J. (2024). Data thermodynamics: From physics to neuroscience. Entropy, 26(9), 779.

    [13] Beggs, J. M., & Plenz, D. (2003). Neuronal avalanches in neocortical circuits. Journal of Neuroscience, 23(35), 11167–11177.

    [14] Freeman, W. J. (2008). Neurodynamics: An exploration in mesoscopic mind dynamics. Springer-Verlag.

    [15] Gao, J., et al. (2016). Common resilience patterns in complicated networks. Nature, 530(7590), 307–312.

    [16] Reynolds, J. H., & Heeger, D. J. (2009). The normalization mannequin of consideration. Neuron, 61(2), 168–185.

    [17] Shew, W. L., et al. (2009). Neuronal avalanches indicate most dynamic vary in cortical networks at criticality. Journal of Neuroscience, 29(49), 15595–15600.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleHow to Evaluate Retrieval Quality in RAG Pipelines (part 2): Mean Reciprocal Rank (MRR) and Average Precision (AP)
    Next Article AI Papers to Read in 2025
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    “The success of an AI product depends on how intuitively users can interact with its capabilities”

    November 14, 2025
    Artificial Intelligence

    How to Crack Machine Learning System-Design Interviews

    November 14, 2025
    Artificial Intelligence

    Music, Lyrics, and Agentic AI: Building a Smart Song Explainer using Python and OpenAI

    November 14, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Novel AI model inspired by neural dynamics from the brain | MIT News

    May 3, 2025

    Amazon CEO’s New Memo Signals a Brutal Truth: More AI, Fewer Humans

    June 24, 2025

    Exporting MLflow Experiments from Restricted HPC Systems

    April 24, 2025

    How to Use Gyroscope in Presentations, or Why Take a JoyCon to DPG2025

    April 21, 2025

    Guide: Använd Gemini som din personliga tränare

    June 20, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Why the Future Is Human + Machine

    October 21, 2025

    Enterprise AI Investments 2025: Top Use-Cases

    August 6, 2025

    Visual Pollen Classification Using CNNs and Vision Transformers

    October 1, 2025
    Our Picks

    “The success of an AI product depends on how intuitively users can interact with its capabilities”

    November 14, 2025

    How to Crack Machine Learning System-Design Interviews

    November 14, 2025

    Music, Lyrics, and Agentic AI: Building a Smart Song Explainer using Python and OpenAI

    November 14, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.