Close Menu
    Trending
    • Escaping the SQL Jungle | Towards Data Science
    • A Gentle Introduction to Nonlinear Constrained Optimization with Piecewise Linear Approximations
    • The Math That’s Killing Your AI Agent
    • Building Robust Credit Scoring Models (Part 3)
    • How to Measure AI Value
    • What’s the right path for AI? | MIT News
    • MIT and Hasso Plattner Institute establish collaborative hub for AI and creativity | MIT News
    • OpenAI is throwing everything into building a fully automated researcher
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » Escaping the SQL Jungle | Towards Data Science
    Artificial Intelligence

    Escaping the SQL Jungle | Towards Data Science

    ProfitlyAIBy ProfitlyAIMarch 21, 2026No Comments14 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    don’t collapse in a single day. They develop slowly, question by question.

    “What breaks after I change a desk?”

    A dashboard wants a brand new metric, so somebody writes a fast SQL question. One other group wants a barely totally different model of the identical dataset, so that they copy the question and modify it. A scheduled job seems. A saved process is added. Somebody creates a derived desk straight within the warehouse.

    Months later, the system seems nothing like the easy set of transformations it as soon as was.

    Enterprise logic is scattered throughout scripts, dashboards, and scheduled queries. No one is completely positive which datasets rely on which transformations. Making even a small change feels dangerous. A handful of engineers turn into the one ones who actually perceive how the system works as a result of there is no such thing as a documentation.

    Many organizations finally discover themselves trapped in what can solely be described as a SQL jungle.

    On this article we discover how programs find yourself on this state, tips on how to acknowledge the warning indicators, and tips on how to deliver construction again to analytical transformations. We’ll take a look at the rules behind a well-managed transformation layer, the way it matches into a contemporary information platform, and customary anti-patterns to keep away from:

    1. How the SQL jungle got here to be
    2. Necessities of a metamorphosis layer
    3. The place the transformation layer matches in a knowledge platform
    4. Widespread anti-patterns
    5. How you can acknowledge when your group wants a metamorphosis framework

    1. How the SQL jungle got here to be

    To know the “SQL jungle” we first want to have a look at how fashionable information architectures developed.

    1.1 The shift from ETL to ELT

    Traditionally information engineers constructed pipelines that adopted an ETL construction:

    Extract --> Remodel --> Load

    Knowledge was extracted from operational programs, remodeled utilizing pipeline instruments, after which loaded into a knowledge warehouse. Transformations had been applied in instruments comparable to SSIS, Spark or Python pipelines.

    As a result of these pipelines had been advanced and infrastructure-heavy, analysts depended closely on information engineers to create new datasets or transformations.

    Fashionable architectures have largely flipped this mannequin

    Extract --> Load --> Remodel

    As a substitute of remodeling information earlier than loading it, organizations now load uncooked information straight into the warehouse, and transformations occur there. This structure dramatically simplifies ingestion and permits analysts to work straight with SQL within the warehouse.

    It additionally launched an unintended aspect impact.


    1.2 Penalties of ELT

    Within the ELT structure, analysts can rework information themselves. This unlocked a lot sooner iteration but in addition launched a brand new problem. The dependency on information engineers disappeared, however so did the construction that engineering pipelines offered.

    Transformations can now be created by anybody (analysts, information scientists, engineer) in anywhere (BI instruments, notebooks, warehouse tables, SQL jobs).

    Over time, enterprise logic grew organically contained in the warehouse. Transformations gathered as scripts, saved procedures, triggers and scheduled jobs. Earlier than lengthy, the system became a dense jungle of SQL logic and lots of handbook (re-)work.

    In abstract:

    ETL centralized transformation logic in engineering pipelines.

    ELT democratized transformations by transferring them into the warehouse.

    With out construction, transformations develop unmanaged, leading to a system that turns into undocumented, fragile and inconsistent. A system wherein totally different dashboards could compute the identical metric in numerous methods and enterprise logic turns into duplicated throughout queries, reviews, and tables.


    1.3 Bringing again construction with a metamorphosis layer

    On this article we use a metamorphosis layer to handle transformations contained in the warehouse successfully. This layer combines the engineering self-discipline of ETL pipelines whereas preserving the pace and suppleness of the ELT structure:

    The transformation layer brings engineering self-discipline to analytical transformations.

    When applied efficiently, the transformation layer turns into the one place the place enterprise logic is outlined and maintained. It acts because the semantic spine of the info platform, bridging the hole between uncooked operational information and business-facing analytical fashions.

    With out the transformation layer, organizations typically accumulate massive quantities of knowledge however have problem to show it into dependable info. The reason is that enterprise logic tends to unfold throughout the platform. Metrics get redefined in dashboards, notebooks, queries and so on.

    Over time this results in some of the widespread issues in analytics: a number of conflicting definitions of the identical metric.


    2. Necessities of a Transformation Layer

    If the core downside is unmanaged transformations, the following logical query is:

    What would well-managed transformations appear like?

    Analytical transformations ought to observe the identical engineering rules we anticipate in software program programs, going from ad-hoc scripts scattered throughout databases to “transformations as maintainable software program parts“.

    On this chapter, we focus on what necessities a metamorphosis layer should meet with the intention to correctly handle transformations and, doing so, tame the SQL jungle.


    2.1 From SQL scripts to modular parts

    As a substitute of enormous SQL scripts or saved procedures, transformations are damaged up into small, composable fashions.

    To be clear: a mannequin is simply an SQL question saved as a file. This question defines how one dataset is constructed from one other dataset.

    The examples under present how information transformation and modeling instrument <em>dbt</em> creates fashions. Every instrument has their very own manner, the precept of turning scripts into parts is extra necessary than the precise implementation.

    Examples:

    -- fashions/staging/stg_orders.sql
    choose
        order_id,
        customer_id,
        quantity,
        order_date
    from uncooked.orders

    When executed, this question materializes as a desk (staging.stg_orders) or view in your warehouse. Fashions can then construct on high of one another by referencing one another:

    -- fashions/intermediate/int_customer_orders.sql
    choose
        customer_id,
        sum(quantity) as total_spent
    from {{ ref('stg_orders') }}
    group by customer_id

    And:

    -- fashions/marts/customer_revenue.sql
    choose
        c.customer_id,
        c.title,
        o.total_spent
    from {{ ref('int_customer_orders') }} o
    be a part of {{ ref('stg_customers') }} c utilizing (customer_id)

    This creates a dependency graph:

    stg_orders
          ↓
    int_customer_orders
          ↓
    customer_revenue

    Every mannequin has a single duty and builds upon different fashions by referencing them (e.g. ref('stg_orders')). This strategy has has main benefits:

    • You possibly can see precisely the place information comes from
    • You understand what’s going to break if one thing modifications
    • You possibly can safely refactor transformations
    • You keep away from duplicating logic throughout queries

    This structured system of transformations makes transformation system simpler to learn, perceive, preserve and evolve.


    2.2 Transformations that stay in code

    A managed system shops transformations in version-controlled code repositories. Consider this as a challenge that comprises SQL recordsdata as a substitute of SQL being saved in a database. It’s much like how a software program challenge comprises supply code.

    This allows practices which might be fairly acquainted in software program engineering however traditionally uncommon in information pipelines:

    • pull requests
    • code opinions
    • model historical past
    • reproducible deployments

    As a substitute of enhancing SQL straight in manufacturing databases, engineers and analysts work in a managed growth workflow, even with the ability to experiment in branches.


    2.3 Knowledge High quality as a part of growth

    One other key functionality a managed transformation system ought to present is the flexibility to outline and run information assessments.

    Typical examples embrace:

    • making certain columns should not null
    • verifying uniqueness of main keys
    • validating relationships between tables
    • implementing accepted worth ranges

    These assessments validate assumptions concerning the information and assist catch points early. With out them, pipelines typically fail silently the place incorrect outcomes propagate downstream till somebody notices a damaged dashboard


    2.4 Clear lineage and documentation

    A managed transformation framework additionally supplies visibility into the info system itself.

    This usually contains:

    • automated lineage graphs (the place does the info come from?)
    • dataset documentation
    • descriptions of fashions and columns
    • dependency monitoring between transformations

    This dramatically reduces reliance on tribal data. New group members can discover the system somewhat than counting on a single one who “is aware of how the whole lot works.”


    2.5 Structured modeling layers

    One other widespread sample launched by managed transformation frameworks is the flexibility to separate transformation layers.

    For instance, you would possibly make the most of the next layers:

    uncooked
    staging
    intermediate
    marts

    These layers are sometimes applied as separate schemas within the warehouse.

    Every layer has a selected function:

    • uncooked: ingested information from supply programs
    • staging: cleaned and standardized tables
    • intermediate: reusable transformation logic
    • marts: business-facing datasets

    This layered strategy prevents analytical logic from turning into tightly coupled to uncooked ingestion tables.


    3. The place the Transformation Layer Matches in a Knowledge Platform

    With the earlier chapters, it turns into clear to see the place a managed transformation framework matches inside a broader information structure.

    A simplified fashionable information platform typically seems like this:

    Operational programs / APIs
               ↓
          1. Knowledge ingestion
               ↓
          2. Uncooked information
               ↓
      3. Transformation layer
               ↓
        4. Analytics layer

    Every layer has a definite duty.

    3.1 Ingestion layer

    Duty: transferring information into the warehouse with minimal transformation. Instruments usually embrace customized ingestion scripts, Kafka or Airbyte.

    3.2 Uncooked information layer

    Accountable for storing information as shut as attainable to the supply system. Prioritizes completeness, reproducibility and traceability of knowledge. Little or no transformation ought to occur right here.

    3.3 Transformation layer

    That is the place the essential modelling work occurs.

    This layer converts uncooked datasets into structured, reusable analytical fashions. Typical duties include cleansing and standardizing information, becoming a member of datasets, defining enterprise logic, creating aggregated tables and defining metrics.

    That is the layer the place frameworks like dbt or SQLMesh function. Their function is to make sure these transformations are

    • structured
    • model managed
    • testable
    • documented

    With out this layer, transformation logic tends to fragment throughout queries dashboards and scripts.

    3.4 Analytics layer

    This layer consumes the modeled datasets. Typical shoppers embrace BI instruments like Tableau or PowerBI, information science workflows, machine studying pipelines and inner information functions.

    These instruments can depend on constant definitions of enterprise metrics since transformations are centralized within the modelling layer.


    3.5 Transformation instruments

    A number of instruments try to deal with the problem of the transformation layer. Two well-known examples are dbt and SQLMesh. These instruments make it very accessible to only get began making use of construction to your transformations.

    Simply do not forget that these instruments should not the structure itself, they’re merely frameworks that assist implement the architectural layer that we’d like.


    4. Widespread Anti-Patterns

    Even when organizations undertake fashionable information warehouses, the identical issues typically reappear if transformations stay unmanaged.

    Beneath are widespread anti-patterns that, individually, could seem innocent, however collectively they create the situations for the SQL jungle. When enterprise logic is fragmented, pipelines are fragile and dependencies are undocumented, onboarding new engineers is gradual and programs turn into troublesome to take care of and evolve.

    4.1 Enterprise logic applied in BI instruments

    One of the widespread issues is enterprise logic transferring into the BI layer. Take into consideration “calculating income in a Tableau dashboard”.

    At first this appears handy since analysts can shortly construct calculations with out ready for engineering help. In the long term, nonetheless, this results in a number of points:

    • metrics turn into duplicated throughout dashboards
    • definitions diverge over time
    • problem debugging

    As a substitute of being centralized, enterprise logic turns into fragmented throughout visualization instruments. A wholesome structure retains enterprise logic within the transformation layer, not in dashboards.


    4.2 Big SQL queries

    One other widespread anti-pattern is writing extraordinarily massive SQL queries that carry out many transformations without delay. Take into consideration queries that:

    • be a part of dozens of tables
    • include deeply nested subqueries
    • implement a number of levels of transformation in a single file

    These queries shortly turn into troublesome to learn, debug, reuse and preserve. Every mannequin ought to ideally have a single duty. Break transformations into small, composable fashions to extend maintainability.


    4.3 Mixing transformation layers

    Keep away from mixing transformation tasks inside the similar fashions, like:

    • becoming a member of uncooked ingestion tables straight with enterprise logic
    • mixing information cleansing with metric definitions
    • creating aggregated datasets straight from uncooked information

    With out separation between layers, pipelines turn into tightly coupled to uncooked supply constructions. To treatment this, introduce clear layers comparable to the sooner mentioned uncooked, staging, intermediate or marts.

    This helps isolate tasks and retains transformations simpler to evolve.


    4.4 Lack of testing

    In lots of programs, information transformations run with none type of validation. Pipelines execute efficiently even when the ensuing information is wrong.

    Introducing automated information assessments helps detect points like duplicate main keys, sudden null values and damaged relationships between tables earlier than they propagate into reviews and dashboards.


    4.5 Enhancing transformations straight in manufacturing

    One of the fragile patterns is modifying SQL straight contained in the manufacturing warehouse. This causes many issues the place:

    • modifications are undocumented
    • errors instantly have an effect on downstream programs
    • rollbacks are troublesome

    In transformation layer, transformations are handled as version-controlled code, permitting modifications to be reviewed and examined earlier than deployment.


    5. How you can Acknowledge When Your Group Wants a Transformation Framework

    Not each information platform wants a completely structured transformation framework from day one. In small programs, a handful of SQL queries could also be completely manageable.

    Nonetheless, because the variety of datasets and transformations grows, unmanaged SQL logic tends to build up. In some unspecified time in the future the system turns into obscure, preserve, and evolve.

    There are a number of indicators that your group could also be reaching this level.

    1. The variety of transformation queries retains rising
      Consider dozens or a whole bunch of derived tables
    2. Enterprise metrics are outlined in a number of locations
      Instance: totally different definition of “lively customers” throughout groups
    3. Problem understanding the system
      Onboarding new engineers takes weeks or months. Tribal data required for questions on information origins, dependencies and lineage
    4. Small modifications have unpredictable penalties
      Renaming a column could break a number of downstream datasets or dashboards
    5. Knowledge points are found too late
      High quality points floor after a clients discovers incorrect numbers on a dashboard; the results of incorrect information propagating unchecked via a number of layers of transformations.

    When these signs start to look, it’s normally time to introduce a structured transformation layer. Frameworks like dbt or SQLMesh are designed to assist groups introduce this construction whereas preserving the pliability that fashionable information warehouses present.


    Conclusion

    Fashionable information warehouses have made working with information sooner and extra accessible by shifting from ETL to ELT. Analysts can now rework information straight within the warehouse utilizing SQL, which enormously improves iteration pace and reduces dependence on advanced engineering pipelines.

    However this flexibility comes with a danger. With out construction, transformations shortly turn into fragmented throughout scripts, dashboards, notebooks, and scheduled queries. Over time this results in duplicated enterprise logic, unclear dependencies, and programs which might be troublesome to take care of: the SQL jungle.

    The answer is to introduce engineering self-discipline into the transformation layer. By treating SQL transformations as maintainable software program parts — model managed, modular, examined, and documented — organizations can construct information platforms that stay comprehensible as they develop.

    Frameworks like dbt or SQLMesh can assist implement this construction, however an important change is adopting the underlying precept: managing analytical transformations with the identical self-discipline we apply to software program programs.

    With this we will create a knowledge platform the place enterprise logic is clear, metrics are constant, and the system stays comprehensible even because it grows. When that occurs, the SQL jungle turns into one thing way more worthwhile: a structured basis that your complete group can belief.


    I hope this text was as clear as I meant it to be but when this isn’t the case please let me know what I can do to make clear additional. Within the meantime, take a look at my other articles on all types of programming-related matters.

    Completely happy coding!

    — Mike



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleA Gentle Introduction to Nonlinear Constrained Optimization with Piecewise Linear Approximations
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    A Gentle Introduction to Nonlinear Constrained Optimization with Piecewise Linear Approximations

    March 21, 2026
    Artificial Intelligence

    The Math That’s Killing Your AI Agent

    March 20, 2026
    Artificial Intelligence

    Building Robust Credit Scoring Models (Part 3)

    March 20, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Firefly Boards – ett nytt AI-verktyg som låter användare generera idéer

    April 25, 2025

    OpenAI kommande sociala app – den ultimata TikTok-AI-slopmaskin

    October 3, 2025

    Expanding robot perception | MIT News

    April 7, 2025

    Svenska AI-reformen – miljoner svenskar får gratis AI-verktyg

    May 9, 2025

    How To Build a Benchmark for Your Models

    May 15, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    🚪🚪🐐 Lessons in Decision Making from the Monty Hall Problem

    May 15, 2025

    Should You Turn Your Executives Into AI Avatars?

    September 16, 2025

    Elon Musk i konflikt med Groks källhänvisning

    June 22, 2025
    Our Picks

    Escaping the SQL Jungle | Towards Data Science

    March 21, 2026

    A Gentle Introduction to Nonlinear Constrained Optimization with Piecewise Linear Approximations

    March 21, 2026

    The Math That’s Killing Your AI Agent

    March 20, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.