Close Menu
    Trending
    • Optimizing Data Transfer in Distributed AI/ML Training Workloads
    • Achieving 5x Agentic Coding Performance with Few-Shot Prompting
    • Why the Sophistication of Your Prompt Correlates Almost Perfectly with the Sophistication of the Response, as Research by Anthropic Found
    • From Transactions to Trends: Predict When a Customer Is About to Stop Buying
    • America’s coming war over AI regulation
    • “Dr. Google” had its issues. Can ChatGPT Health do better?
    • Evaluating Multi-Step LLM-Generated Content: Why Customer Journeys Require Structural Metrics
    • Why SaaS Product Management Is the Best Domain for Data-Driven Professionals in 2026
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » The Great Data Closure: Why Databricks and Snowflake Are Hitting Their Ceiling
    Artificial Intelligence

    The Great Data Closure: Why Databricks and Snowflake Are Hitting Their Ceiling

    ProfitlyAIBy ProfitlyAIJanuary 16, 2026No Comments14 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Introduction

    a knowledge firm actually develop?

    This week what would have been information a 12 months in the past was now not information. Snowflake invested in AtScale, a supplier of semantic layer companies in a strategic funding within the waning firm’s historical past. An odd transfer, given the dedication to the open semantic interchange or “OSI” (yet one more acronym or .yaa) which seems to be metricflow masquerading as one thing else.

    In the meantime, Databricks, the AI and Information firm, invested in AI-winner and all-round VC paramore Loveable — the quickly rising vibe-coding firm from Sweden.

    Beginning a enterprise arm is a tried-and-tested route for enterprises. Everyone from Walmart and Hitachi to banks like JPMorgan and Goldman Sachs, and naturally the hyperscalers — MSFT, GOOG — themselves have enterprise arms (although surprisingly not AWS).

    The advantages are clear. An funding right into a spherical can provide the best of first refusal. It presents each events affect round complementary roadmap options in addition to clear distribution benefits. “Synergy” is the phrase utilized in boardrooms, although it’s the much less insidious and pleasant youthful brother of central cost cutting so prevalent in PE moderately than venture-backed companies.

    It ought to due to this fact come as no shock to see that Databricks are branching out outdoors of Information. In spite of everything (and Ali has been very open about this), the workforce understands the best way to develop the corporate is thru new use instances, most notably AI. Whereas Dolly was a flop, the jury is out on the partnership with OpenAI. AI/BI, in addition to Databricks Functions, are promising initiatives designed to carry extra mates into the tent — outdoors of the core SYSADMIN cluster directors.

    Snowflake in the meantime could also be attempting the same tack however with differing ranges of success. Other than Streamlit, it’s not clear what worth its acquisitions are really bringing. Openflow, Neolithic Nifi under-the-hood, just isn’t properly obtained. Slightly, it’s the inside developments such because the embedding of dbt core into the Snowflake platform that look like gaining extra traction.

    On this article, we’ll dive into the various factors at play and make some predictions for 2026. Let’s get caught in!

    Progress by way of use instances

    Databricks has an issue. A giant drawback. And that’s fairness.

    Because the fourth-largest privately held firm on this planet, on the tender age of 12 its staff require liquidity. And liquidity is dear (see this excellent article).

    To make good on its inside commitments, Databricks wanted maybe $5bn+ when it did this raise. The quantity it wants per 12 months is critical. It’s due to this fact merely not an choice to stop elevating cash with out firing staff and slicing prices.

    The expansion is staggering. In the latest series L (!) the company cites 55% yearly period-on-period progress resulting in a valuation of over $130bn. The corporate should proceed to lift cash to pay its opex and fairness, however there may be one other constraint which is valuation. At this level Databricks’ potential to lift cash is virtually a bellwether for the business, and so there’s a vested curiosity for everybody concerned (the record is big) to maintain issues up.

    Supply: previous article

    The dream is to proceed rising the corporate as this may maintain the valuation — valuations are tied to income progress. Which brings us again to make use of instances.

    The clear use instances, as proven right here, are roughly:

    • Large information processing and spark
    • Inside this, Machine Studying workloads
    • AI workloads
    • Information warehousing
    • Ingestion or Lakeflow (Arcion we suspect was maybe a bit early)
    • Enterprise Intelligence
    • Functions

    It’s price noting these sectors are all forecasted to develop at round 15–30% all in, per the overwhelming majority of market experiences (an instance here). This displays the underlying demand for extra information, extra automation, and extra effectivity which I imagine is finally justified, particularly within the age of AI.

    Sources like Technavio, Mordor Intelligence, and infrequently simply plain previous Business Due Diligence Studies not often disagree with one another, with virtually the entire them placing the protected 15–30% vary in sectors individuals suppose must be invested in. The purpose is; this isn’t a stagnant or a rocketing market, and so they all agree. In distinction to, for instance, AI

    It will seem to point, due to this fact, that the underside or “flooring” for Databricks can be a couple of progress of 15–30%, and with it maybe a 40% haircut to valuation multiples (assuming linear correlation; sure, sure, assumptions, assumptions — some extra information here), barring after all any exogenous shocks to the system equivalent to OpenAI going out of business or war.

    That is hardly regarding as a bear-case, which makes me marvel — what’s the bull?

    The bull lies within the two A’s: AI use instances and Functions.

    AI as a means out

    If Databricks can efficiently associate with the mannequin suppliers and develop into the de-facto engine for internet hosting fashions and working the related workflows, it may very well be large.

    Handkerchief maths — the income is $4.8bn RR rising at 55%. Say we’re rising at 30% in regular state, we’re lacking 25%. 25% of $4.8 is $1.2bn. The place can this come from? Supposedly current AI merchandise and current warehousing is already over $2bn (see here). What occurs subsequent 12 months when Databricks is at $6bn and we have to develop 50% and due to this fact want $3bn? Is the enterprise going to double the AI half?

    Confluent is a benchmark. It’s the largest Kafka/stream processing firm, with a income of about $1.1bn annualised. It grows about 25% y-o-y however traded at about 8x income and bought to IBM for $11bn, so about 11x income. Even with its loyal fanbase and robust adoptions for AI use instances (see for instance marketecture from Sean Falconer.), it could nonetheless wrestle to place one other $250m of annual progress on yearly.

    Applications are another story. People who construct data-intensive functions usually are not people who usually construct internal-facing merchandise, a activity usually borne by in-house groups of software program engineers or consultants. These are groups that already know the way to do that, and know learn how to do it properly, with current expertise particularly designed for its function, specifically core engineering primitives like React, Postgres (self-hosted) and Quick API.

    Dashboards with a twist? Source

    An information engineer might log in to Loveable, spin up Neon-Postgres, a declarative spark ETL pipeline, and front-end in Databricks. They may. However will they wish to add this to their ever-increasing backlog? I’m not positive.

    The purpose is the core enterprise just isn’t rising quick sufficient to maintain the present valuation so extra traces of enterprise are required. Databricks is sort of a golden goose at the craps table, who continues to keep away from rolling the unutterable quantity. They will now proceed making an increasing number of bets, whereas all these across the desk proceed to learn.

    Databricks is topped out as a data-only firm.

    We’ve written before about methods they might have moved out of this. Spark-structured streaming was an apparent alternative, however the ship has sailed, and it’s corporations like Aiven and Veverica that are actually in pole place for the Flink race.

    📚 Learn: What not to miss in Real-time Data and AI in 2025 📚

    To develop into a model-serving firm or an ‘AI Cloud’ appears additionally a tall order. Coreweave, Lambda, and naturally Nebius are all on track to really challenge the hyperscalers here.

    An AI cloud is essentially pushed by a excessive availability of GPU-optimised compute. This doesn’t simply imply leasing EC2 situations from Jeff Bezos. It means sliding into Jensen Huang’s DMs and shopping for a ton of GPUs.

    Nebius has about 20,000, with another 30,000 on the best way — this Yahoo report thinks the numbers are higher. All of the AI Clouds lease area in information centres in addition to constructing their very own. Inference, not like spark, just isn’t a commodity due to the immense software program, {hardware}, and logistical challenges required.

    Allow us to not overlook that Nebius owns just over 25% of Clickhouse — each groups being very software engineering-led and Russian; the Yandex Alumni Club.

    If there may be one factor now we have discovered it’s that it’s simpler to go up the worth chain than down it. I wrote about this funnel maybe two years in the past now nevertheless it appears more true than ever.

    Who remembers this from my early running a blog days? Learn: Unstructured Data funnel

    Snowflake easily eats into dbt. Databricks has simply eaten into Snowflake’s warehouse income. Microsoft will eat into Databricks’. And in flip, with uncooked information centre energy, NVIDIA and Meta partnerships, and a military of the most effective builders within the enterprise, Nebius can eat into the hyperscalers.

    Information warehousing underneath assault

    With each passing day proprietary information warehousing platforms appear an increasing number of unlikely to be the technical finish for AI and Information infrastructure.

    Salesforce are increasing levies, databases are supporting cross-query capabilities, CDOs are running Duck DB in Snowflake itself.

    Even Bill Inmon acknowledges warehousing corporations missed the warehousing!

    Whereas handy, there’s a scale at which enterprises and even late stage start-ups are demanding higher openness, higher flexibility and cheaper compute.

    At Orchestra we’ve seen this first-hand. The businesses applied sciences equivalent to Iceberg are overwhelmingly large. From the biggest telecom suppliers to the Reserving.com’s of this world (who occur to make use of and love Snowflake; extra on this later), conventional information warehousing is unlikely to proceed dominating the share of price range it has achieved for the final decade.

    There are a number of methods Snowflake has additionally tried to broaden its core providing:

    • Assist for managed iceberg; open compute engine
    • Information cataloging (Select *)
    • Functions (streamlit)
    • Spark and different types of compute like containers
    • AI brokers for Analysts AKA snowflake intelligence
    • Transformation (i.e. dbt)

    Sarcastically for a proprietary engine supplier, it could seem that Iceberg is a big progress avenue, in addition to AI. See extra from TT here.

    Snowflake prospects adore it.

    Information Pangea

    I feel the definitions of the pioneers, early adopters, late adopters, and laggards are altering.

    Early Adopters now embrace a heavy real-time part and AI-first strategy to the stack. That is more likely to revert to Machine Studying as individuals realise AI just isn’t a hammer for each nail.

    These corporations wish to associate with a number of giant distributors, and have a excessive urge for food for constructing in addition to shopping for software program. They are going to have at the very least one vendor within the streaming/AI, question engine and analytics area. A superb instance is booking.com, or maybe Fresha, who makes use of Snowflake, Starrocks, and Kafka (I cherished the article beneath).

    📚 Learn: Exploring how modern streaming tools power the next generation of analytics with StarRocks. 📚

    Early Adopters could have the normal analytics stack after which one different space. They lack the dimensions to totally buy-in to an enterprise-wide information and AI technique, so concentrate on these use-cases they know work. Automation, Reporting.

    The previous “early adopters” would have had the Andreesen Horowitz data stack. That, I’m afraid, is now not cool, or in. That was the previous structure. The late adopters have the overall stack.

    The laggards? Who is aware of. They are going to in all probability go along with whoever their CTO is aware of essentially the most. Be it Informatica (see this incredible reddit post), Cloth, or maybe even GCP!

    The subsequent step: chaos for smaller distributors

    Quite a lot of corporations are altering tack. Secoda were acquired by Atlassian, Choose Star have been acquired by Snowflake. Arch.dev, the creators of Meltano, shut-down and passed the project to Matatika. From the massive corporations to the small, slowing income progress mixed with large strain from bloated VC rounds make constructing a “Trendy-Information Stack”-style firm an untenable strategy.

    📚 Learn: The Remaining Voyage of the Trendy Information Stack | Can the Context Layer for AI provide catalogs with the last chopper out of Saigon? 📚

    What would occur when the Databricks and Snowflake progress numbers lastly begin to sluggish, as we argue they need to right here?

    What would occur if there was a big exogenous market shock or OpenAI ran out of cash quicker than anticipated?

    What occurs as Salesforce enhance taxes and therefore instruments like Fivetran and dbt enhance in worth much more?

    An ideal storm for migrations and re-architecturing is brewing. Information infrastructure is extraordinarily sticky, which implies in tough occasions, corporations increase costs. EC2 spot situations have not likely modified a lot in worth through the years, and so neither too has information infra compute — and but even AWS are raising prices of GPUs.

    The marginal price of onboarding an extra instrument is changing into very excessive. We used to construct every little thing ourselves because it was the one means. However having one instrument for each drawback doesn’t work both.

    Picture the writer’s

    We should always not overlook that Parkinson’s law applies to IT budgets too. Regardless of the price range is, the price range will get spent. Think about in case you had a instrument that helped you automate extra issues with AI whereas lowering your wareouse invoice and lowering your BI Licenses (sometimes a big 25–50% P&L price range line) — what do you do?

    You don’t pat your self on the again — you spend it. You spend it on extra stuff, doing extra stuff. You’ll in all probability push your Databricks and Snowflake invoice again up. However you’ll have extra to point out for it.

    Consolidation is driving funds again into centre of gravities. These are Snowflake, Databricks, GCP, AWS and Microsoft (and to a lesser extent, palantir). This spells chaos for many smaller distributors.

    Conclusion — brace for less complicated structure

    The Salesforce Tax is a pivotal second in our business. Corporations like Salesforce, SAP, and ServiceNow all have an immense quantity of knowledge and sufficient clout to maintain it there.

    As Information Individuals, anybody who has achieved a migration from Salesforce to Netsuite is aware of that migrating these instruments might be the largest, most costly, and most painful transfer anybody faces of their skilled careers.

    Salesforce charging infrastructure service suppliers charges will increase costs, which in flip, mixed with the more and more precarious home of playing cards we see in AI and Information, all level in the direction of large consolidation.

    ServiceNow’s acquisition of Information.World, I feel, supplies some readability into why we’ll see information groups make extra use of current tooling, simplifying structure within the course of. Information.World is a supplier of data graphs and ontologies. By mapping the ServiceNow information schema to an ontology, a gargantuan activity, ServiceNow might find yourself with half-decent AI and brokers working inside ServiceNow.

    AgentForce and Data360 is Salesforce’s try, and supposedly already has $1.4bn in revenue, although we suspect it contains a variety of legacy in there too.

    These suppliers do not likely need information working round as AI use instances in Snowflake or Databricks. They need the Procurement Specialists, Finance Professionals, and Advertising and marketing Gurus staying in their platforms — and so they have the means to make them keep.

    This isn’t monetary recommendation and this isn’t a loopy prediction. To foretell that Snowflake and Databricks will find yourself rising extra alongside the analyst consensus is hardly difficult.

    However the concept that the largest information corporations’ progress is on the verge of slowing is difficult. It challenges the rhetoric. It challenges the AI maximalist discourse.

    We’re coming into the period of the Nice Information Closure. Whereas the AI maximalists dream of a borderless future, the truth is a heavy ceiling constructed by the incumbents’ gravity. On this new panorama, the winner isn’t the one with the most effective set of instruments, however the people who benefit from what they’ve.

    About Me

    I’m the CEO of Orchestra. We assist Information Individuals construct, run and monitor their pipelines simply.

    You’ll find me on Linkedin here.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWhen Shapley Values Break: A Guide to Robust Model Explainability
    Next Article From RGB to Lab: Addressing Color Artifacts in AI Image Compositing
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Optimizing Data Transfer in Distributed AI/ML Training Workloads

    January 23, 2026
    Artificial Intelligence

    Achieving 5x Agentic Coding Performance with Few-Shot Prompting

    January 23, 2026
    Artificial Intelligence

    Why the Sophistication of Your Prompt Correlates Almost Perfectly with the Sophistication of the Response, as Research by Anthropic Found

    January 23, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Med Claude Explains kan Claude nu skapa egna blogginlägg

    June 4, 2025

    “The success of an AI product depends on how intuitively users can interact with its capabilities”

    November 14, 2025

    Key Differences Explained with Examples

    November 13, 2025

    Open the pod bay doors, Claude

    August 26, 2025

    Ny forskning visar varför AI-bilder ser så konstiga ut

    October 21, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    The Hidden Trap of Fixed and Random Effects

    July 18, 2025

    Studenter kan Vibe koda med Cursor Pro helt gratis i ett helt år

    May 8, 2025

    A major AI training data set contains millions of examples of personal data

    July 18, 2025
    Our Picks

    Optimizing Data Transfer in Distributed AI/ML Training Workloads

    January 23, 2026

    Achieving 5x Agentic Coding Performance with Few-Shot Prompting

    January 23, 2026

    Why the Sophistication of Your Prompt Correlates Almost Perfectly with the Sophistication of the Response, as Research by Anthropic Found

    January 23, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.