Close Menu
    Trending
    • “The success of an AI product depends on how intuitively users can interact with its capabilities”
    • How to Crack Machine Learning System-Design Interviews
    • Music, Lyrics, and Agentic AI: Building a Smart Song Explainer using Python and OpenAI
    • An Anthropic Merger, “Lying,” and a 52-Page Memo
    • Apple’s $1 Billion Bet on Google Gemini to Fix Siri
    • Critical Mistakes Companies Make When Integrating AI/ML into Their Processes
    • Nu kan du gruppchatta med ChatGPT – OpenAI testar ny funktion
    • OpenAI’s new LLM exposes the secrets of how AI really works
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » The Three Ages of Data Science: When to Use Traditional Machine Learning, Deep Learning, or an LLM (Explained with One Example)
    Artificial Intelligence

    The Three Ages of Data Science: When to Use Traditional Machine Learning, Deep Learning, or an LLM (Explained with One Example)

    ProfitlyAIBy ProfitlyAINovember 11, 2025No Comments11 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    of the universe (made by some of the iconic singers ever) says this:

    Want I might return
    And alter these years
    I’m going by means of modifications

    Black sabbath – Adjustments

    This tune is extremely highly effective and talks about how life can change proper in entrance of you so rapidly.

    That tune is a few damaged coronary heart and a love story. Nevertheless, it additionally jogs my memory a whole lot of the modifications that my job, as a Information Scientist, has undergone over the past 10 years of my profession:

    • Once I began finding out Physics, the one factor I considered when somebody stated “Transformer” was Optimus Prime. Machine Studying for me was all about Linear Regression, SVM, Random Forest and so forth… [2016]
    • Once I did my Grasp’s Diploma in Huge Information and Physics of Complicated Programs, I first heard of “BERT” and varied Deep Studying applied sciences that appeared very promising at the moment. The primary GPT fashions got here out, they usually regarded very attention-grabbing, though nobody anticipated them to be as highly effective as they’re right this moment. [2018-2020]
    • Quick ahead to my life now as a full-time Information Scientist. At present, in the event you don’t know what GPT stands for and have by no means learn “Consideration is All You Want” you could have only a few possibilities of passing a Information Science System Design interview. [2021 – today]

    When folks state that the instruments and the on a regular basis lifetime of an individual working with information are considerably totally different than 10 (and even 5) years in the past, I agree all the best way. What I don’t agree with is the concept the instruments used prior to now needs to be erased simply because every part now appears to be solvable with GPT, LLMs, or Agentic AI.

    The purpose of this text is to think about a single job, which is classifying the love/hate/impartial intent of a Tweet. Particularly, we are going to do it with conventional Machine Studying, Deep Studying, and Giant Language Fashions.

    We’ll do that hands-on, utilizing Python, and we are going to describe why and when to make use of every strategy. Hopefully, after this text, you’ll be taught:

    1. The instruments used within the early days ought to nonetheless be thought of, studied, and at instances adopted.
    2. Latency, Accuracy, and Value needs to be evaluated when selecting the very best algorithm in your use case
    3. Adjustments within the Information Scientist world are crucial and to be embraced with out concern 🙂

    Let’s get began!

    1. The Use Case

    The case we’re coping with is one thing that’s really very adopted in Information Science/AI functions: sentiment evaluation. Because of this, given a textual content, we wish to extrapolate the “feeling” behind the creator of that textual content. That is very helpful for instances the place you wish to collect the suggestions behind a given evaluate of an object, a film, an merchandise you’re recommending, and so forth…

    On this weblog publish, we’re utilizing a really “well-known” sentiment evaluation instance, which is classifying the sensation behind a tweet. As I wished extra management, we is not going to work with natural tweets scraped from the online (the place labels are unsure). As a substitute, we will probably be utilizing content material generated by Giant Language Fashions that we are able to management.

    This system additionally permits us to tune the issue and the number of the issue and to look at how totally different strategies react.

    • Simple case: the love tweets sound like postcards, the hate ones are blunt, and the impartial messages speak about climate and occasional. If a mannequin struggles right here, one thing else is off.
    • More durable case: nonetheless love, hate, impartial, however now we inject sarcasm, combined tones, and refined hints that demand consideration to context. We even have much less information, to have a smaller dataset to coach with.
    • Further Exhausting case: we transfer to 5 feelings: love, hate, anger, disgust, envy, so the mannequin has to parse richer, extra layered sentences. Furthermore, we’ve 0 entries to coach the info: we cannot do any coaching.

    I’ve generated the info and put every of the information in a particular folder of the general public GitHub Folder I’ve created for this venture [data].

    Our purpose is to construct a sensible classification system that can have the ability to effectively grasp the sentiment behind the tweets. However how we could do it? Let’s determine it out.

    2. System Design

    An image that’s all the time extraordinarily useful to think about is the next:

    Picture made by creator

    Accuracy, price, and scale in a Machine Studying system type a triangle. You’ll be able to solely absolutely optimize two on the identical time.

    You’ll be able to have a really correct mannequin that scales very nicely with thousands and thousands of entries, however it received’t be fast. You’ll be able to have a fast mannequin that scales with thousands and thousands of entries, however it received’t be that correct. You’ll be able to have an correct and fast mannequin, however it received’t scale very nicely.

    These concerns are abstracted from the particular drawback, however they assist information which ML System Design to construct. We’ll come again to this.

    Additionally, the facility of our mannequin needs to be proportional to the scale of our coaching set. Usually, we attempt to keep away from the coaching set error to lower at the price of a rise within the check set (the well-known overfitting).

    Picture made by creator

    We don’t wish to be within the Underfitting or Overfitting space. Let me clarify why.

    In easy phrases, underfitting occurs when your mannequin is simply too easy to be taught the actual sample in your information. It’s like attempting to attract a straight line by means of a spiral. Overfitting is the alternative. The mannequin learns the coaching information too nicely, together with all of the noise, so it performs nice on what it has already seen however poorly on new information. The candy spot is the center floor, the place your mannequin understands the construction with out memorizing it.

    We’ll come again to this one as nicely.

    3. Simple Case: Conventional Machine Studying

    We open with the friendliest state of affairs: a extremely structured dataset of 1,000 tweets that we generated and labelled. The three courses (constructive, impartial, unfavourable) are balanced on goal, the language could be very express, and each row lives in a clear CSV.

    Let’s begin with a easy import block of code.

    Let’s see what the dataset appears to be like like:

    Picture made by creator

    Now, we anticipate that this received’t scale for thousands and thousands of rows (as a result of the dataset is simply too structured to be various). Nevertheless, we are able to construct a really fast and correct technique for this tiny and particular use case. Let’s begin with the modeling. Three details to think about:

    1. We’re doing prepare/check break up with 20% of the dataset within the check set.
    2. We’re going to use a TF-IDF strategy to get the embeddings of the phrases. TF-IDF stands for Time period Frequency–Inverse Doc Frequency. It’s a basic method that transforms textual content into numbers by giving every phrase a weight primarily based on how essential it’s in a doc in comparison with the entire dataset.
    3. We’ll mix this system with two ML fashions: Logistic Regression and Assist Vector Machines, from scikit-learn. Logistic Regression is straightforward and interpretable, usually used as a powerful baseline for textual content classification. Assist Vector Machines give attention to discovering the very best boundary between courses and often carry out very nicely when the info will not be too noisy.

    And the efficiency is basically excellent for each fashions.

    Picture made by creator

    For this quite simple case, the place we’ve a constant dataset of 1,000 rows, a conventional strategy will get the job finished. No want for billions of parameter fashions like GPT.

    4. Exhausting Case: Deep Studying

    The second dataset remains to be artificial, however it’s designed to be annoying on goal. Labels stay love, hate, and impartial, but the tweets lean on sarcasm, combined tone, and backhanded compliments. On prime of that, the coaching pool is smaller whereas the validation slice stays massive, so the fashions work with much less proof and extra ambiguity.

    Now that we’ve this ambiguity, we have to take out the larger weapons. There are Deep Studying embedding fashions that keep sturdy accuracy and nonetheless scale nicely in these instances (bear in mind the triangle and the error versus complexity plot!). Particularly, Deep Studying embedding fashions be taught the that means of phrases from their context as a substitute of treating them as remoted tokens.

    For this weblog publish, we are going to use BERT, which is likely one of the most well-known embedding fashions on the market. Let’s first import some libraries:

    … and a few helpers.

    Thanks to those features, we are able to rapidly consider our embedding mannequin vs the TF-IDF strategy.

    Picture made by creator

    As we are able to see, the TF-IDF mannequin is extraordinarily underperforming within the constructive labels, whereas it preserves excessive accuracy when utilizing the embedding mannequin (BERT).

    5. Further Exhausting case: LLM Agent

    Okay, now let’s make issues VERY arduous:

    1. We solely have 100 rows.
    2. We assume we have no idea the labels, that means we can’t prepare any machine studying mannequin.
    3. We’ve got 5 labels: envy, hate, love, disgust, anger.

    As we cannot prepare something, however we nonetheless wish to carry out our classification, we should undertake a way that in some way already has the classifications inside. Giant Language Fashions are the best instance of such a way.

    Word that if we used LLMs for the opposite two instances, it might be like capturing a fly with a cannon. However right here, it makes excellent sense: the duty is difficult, and we’ve no method to do something sensible, as a result of we cannot prepare our mannequin (we don’t have the coaching set).

    On this case, we’ve accuracy at a big scale. Nevertheless, the API takes a while, so we’ve to attend a second or two earlier than the response comes again (bear in mind the triangle!).

    Let’s import some libraries:

    And that is the classification API name:

    And we are able to see that the LLM does an incredible classification job:

    6. Conclusions

    Over the previous decade, the function of the Information Scientist has modified as dramatically because the know-how itself. This would possibly result in the concept of simply utilizing probably the most highly effective instruments on the market, however that’s NOT the very best route for a lot of instances.

    As a substitute of reaching for the largest mannequin first, we examined one drawback by means of a easy lens: accuracy, latency, and value.

    Particularly, here’s what we did, step-by-step:

    • We outlined our use case as tweet sentiment classification, aiming to detect love, hate, or impartial intent. We designed three datasets of accelerating problem: a clear one, a sarcastic one, and a zero-training one.
    • We tackled the straightforward case utilizing TF-IDF with Logistic Regression and SVM. The tweets have been clear and direct, and each fashions carried out nearly completely.
    • We moved to the arduous case, the place sarcasm, combined tone, and refined context made the duty extra advanced. We used BERT embeddings to seize that means past particular person phrases.
    • Lastly, for the additional arduous case with no coaching information, we used a Giant Language Mannequin to categorise feelings instantly by means of zero-shot studying.

    Every step confirmed how the proper device depends upon the issue. Conventional ML is quick and dependable when the info is structured. Deep Studying fashions assist when that means hides between the traces. LLMs are highly effective when you don’t have any labels or want broad generalization.

    7. Earlier than you head out!

    Thanks once more in your time. It means loads ❤️

    My identify is Piero Paialunga, and I’m this man right here:

    Picture made by creator

    I’m initially from Italy, maintain a Ph.D. from the College of Cincinnati, and work as a Information Scientist at The Commerce Desk in New York Metropolis. I write about AI, Machine Studying, and the evolving function of knowledge scientists each right here on TDS and on LinkedIn. For those who favored the article and wish to know extra about machine studying and observe my research, you may:

    A. Comply with me on Linkedin, the place I publish all my tales
    B. Comply with me on GitHub, the place you may see all my code
    C. For questions, you may ship me an e-mail at piero.paialunga@hotmail



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleIs OpenAI Already Too Big to Fail?
    Next Article Do You Really Need GraphRAG? A Practitioner’s Guide Beyond the Hype
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    “The success of an AI product depends on how intuitively users can interact with its capabilities”

    November 14, 2025
    Artificial Intelligence

    How to Crack Machine Learning System-Design Interviews

    November 14, 2025
    Artificial Intelligence

    Music, Lyrics, and Agentic AI: Building a Smart Song Explainer using Python and OpenAI

    November 14, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Boosting Your Anomaly Detection With LLMs

    September 4, 2025

    At the core of problem-solving | MIT News

    April 4, 2025

    Danmark planerar ny lag mot deepfakes

    June 28, 2025

    Apple arbetar på nya chip för AI-servrar, Mac-datorer och smarta glasögon

    May 13, 2025

    Character AI AvatarFX kan nu göra foton och karaktärer levande i videoform

    April 26, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    A Developer’s Guide to Building Scalable AI: Workflows vs Agents

    June 27, 2025

    The agent workforce: Redefining how work gets done 

    November 3, 2025

    Layoffs, Shorter Work Weeks, and the Race to Automate Work

    September 23, 2025
    Our Picks

    “The success of an AI product depends on how intuitively users can interact with its capabilities”

    November 14, 2025

    How to Crack Machine Learning System-Design Interviews

    November 14, 2025

    Music, Lyrics, and Agentic AI: Building a Smart Song Explainer using Python and OpenAI

    November 14, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.