Close Menu
    Trending
    • DataRobot + Nebius: An enterprise-ready AI Factory optimized for agents
    • Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes
    • The New Experience of Coding with AI
    • Why You Should Stop Worrying About AI Taking Data Science Jobs
    • Why Chatbots Are Coming for Your Medical Records
    • One Model to Rule Them All? SAP-RPT-1 and the Future of Tabular Foundation Models
    • Sustaining diplomacy amid competition in US-China relations | MIT News
    • The Pentagon is making plans for AI companies to train on classified data, defense official says
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes
    Artificial Intelligence

    Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes

    ProfitlyAIBy ProfitlyAIMarch 18, 2026No Comments21 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    Introduction

    , we frequently encounter prediction issues the place the end result has an uncommon distribution: a big mass of zeros mixed with a steady or rely distribution for optimistic values. For those who’ve labored in any customer-facing area, you’ve nearly actually run into this. Take into consideration predicting buyer spending. In any given week, the overwhelming majority of customers in your platform don’t buy something in any respect, however the ones who do would possibly spend wherever from $5 to $5,000. Insurance coverage claims observe an identical sample: most policyholders don’t file something in a given quarter, however the claims that do are available differ enormously in measurement. You see the identical construction in mortgage prepayments, worker turnover timing, advert click on income, and numerous different enterprise outcomes.

    The intuition for many groups is to succeed in for the standard regression mannequin and attempt to make it work. I’ve seen this play out a number of occasions. Somebody matches an OLS mannequin, will get unfavorable predictions for half the shopper base, provides a flooring at zero, and calls it a day. Or they fight a log-transform, run into the $log(0)$ downside, tack on a $+1$ offset, and hope for the very best. These workarounds would possibly work, however they gloss over a basic situation: the zeros and the optimistic values in your information are sometimes generated by utterly completely different processes. A buyer who won’t ever purchase your product is essentially completely different from a buyer who buys often however occurred to not this week. Treating them the identical method in a single mannequin forces the algorithm to compromise on each teams, and it often does a poor job on every.

    The two-stage hurdle mannequin gives a extra principled answer by decomposing the issue into two distinct questions.
    First, will the end result be zero or optimistic?
    And second, provided that it’s optimistic, what’s going to the worth be?
    By separating the “if” from the “how a lot,” we will use the proper instruments on every sub-problem independently with completely different algorithms, completely different options, and completely different assumptions, then mix the outcomes right into a single prediction.

    On this article, I’ll stroll by way of the idea behind hurdle fashions, present a working Python implementation, and talk about the sensible concerns that matter when deploying these fashions in manufacturing.
    readers who’re already accustomed to the motivation can skip straight to the implementation part.

    The Drawback with Commonplace Approaches

    Why Not Simply Use Linear Regression? To make this concrete, think about predicting buyer spend.
    If 80% of consumers spend zero and the remaining 20% spend between 10 and 1000 {dollars}, a linear regression mannequin instantly runs into hassle.
    The mannequin can (and can) predict unfavorable spend for some prospects, which is nonsensical since you may’t spend unfavorable {dollars}.
    It is going to additionally battle on the boundary: the huge spike at zero pulls the regression line down, inflicting the mannequin to underpredict zeros and overpredict small optimistic values concurrently.
    The variance construction can also be fallacious.
    Prospects who spend nothing have zero variance by definition, whereas prospects who do spend have excessive variance.
    Whereas you should utilize heteroskedasticity-robust customary errors to get legitimate inference regardless of non-constant variance, that solely fixes the usual errors and doesn’t repair the predictions themselves.
    The fitted values are nonetheless coming from a linear mannequin that’s making an attempt to common over a spike at zero and a right-skewed optimistic distribution, which is a poor match no matter the way you compute the arrogance intervals.

    Why Not Log-Rework? The following factor most individuals strive is a log-transform: $log(y + 1)$ or $log(y + epsilon)$.
    This compresses the proper tail and makes the optimistic values look extra regular, nevertheless it introduces its personal set of issues.
    The selection of offset ($1$ or $epsilon$) is bigoted, and your predictions will change relying on what you decide.
    If you back-transform through $exp(hat{y}) – 1$, you introduce a scientific bias resulting from Jensen’s inequality, because the anticipated worth of the exponentiated prediction just isn’t the identical because the exponentiation of the anticipated prediction.
    Extra essentially, the mannequin nonetheless doesn’t distinguish between a buyer who by no means spends and one who typically spends however occurred to be zero this era.
    Each get mapped to $log(0 + 1) = 0$, and the mannequin treats them identically though they signify very completely different buyer behaviors.

    What This Means for Forecasting. The deeper situation with forcing a single mannequin onto zero-inflated information goes past poor level estimates.
    If you ask one mannequin to explain two essentially completely different behaviors (not partaking in any respect vs. partaking at various intensities), you find yourself with a mannequin that conflates the drivers of every.
    The options that predict whether or not a buyer will buy in any respect are sometimes fairly completely different from the options that predict how a lot they’ll spend given a purchase order.
    Recency and engagement frequency would possibly dominate the “will they purchase” query, whereas earnings and product class preferences matter extra for the “how a lot” query.
    A single regression mixes these alerts collectively, making it troublesome to disentangle what’s really driving the forecast.

    This additionally has sensible implications for the way you act on the mannequin.
    In case your forecast is low for a specific buyer, is it as a result of they’re unlikely to buy, or as a result of they’re more likely to buy however at a small quantity?
    The optimum enterprise response to every state of affairs is completely different.
    You would possibly ship a re-engagement marketing campaign for the primary case and an upsell supply for the second.
    A single mannequin provides you one quantity, however there is no such thing as a option to inform which lever to tug.

    The Two-Stage Hurdle Mannequin

    Conceptual Framework. The core concept behind hurdle fashions is surprisingly intuitive.
    Zeros and positives usually come up from completely different data-generating processes, so we must always mannequin them individually.
    Consider it as two sequential questions your mannequin must reply.
    First, does this buyer cross the “hurdle” and have interaction in any respect?
    And second, provided that they’ve engaged, how a lot do they spend?
    Formally, we will write the distribution of the end result $Y$ conditional on options $X$ as:

    $$ P(Y = y | X) = start{instances} 1 – pi(X) & textual content{if } y = 0 pi(X) cdot f(y | X, y > 0) & textual content{if } y > 0 finish{instances} $$

    Right here, $pi(X)$ is the likelihood of crossing the hurdle (having a optimistic consequence), and $f(y | X, y > 0)$ is the conditional distribution of $y$ provided that it’s optimistic.
    The great thing about this formulation is that these two parts might be modeled independently.
    You should use a gradient boosting classifier for the primary stage and a gamma regression for the second, or logistic regression paired with a neural community, or some other mixture that fits your information.
    Every stage will get its personal characteristic set, its personal hyperparameters, and its personal analysis metrics.
    This modularity is what makes hurdle fashions so sensible in manufacturing settings.

    Stage 1: The Classification Mannequin. The primary stage is an easy binary classification downside: predict whether or not $y > 0$.
    You’re coaching on the complete dataset, with each statement labeled as both zero or optimistic.
    It is a downside that the ML group has a long time of tooling for.
    Logistic regression provides you an interpretable and quick baseline.
    Gradient boosting strategies like XGBoost or LightGBM deal with non-linearities and have interactions effectively.
    Neural networks work when you might have high-dimensional or unstructured options.
    The output from this stage is $hat{pi}(X) = P(Y > 0 | X)$, a calibrated likelihood that the end result will likely be optimistic.

    The necessary factor to get proper right here is calibration.
    Since we’re going to multiply this likelihood by the conditional quantity within the subsequent stage, we want $hat{pi}(X)$ to be a real likelihood, not only a rating that ranks effectively.
    In case your classifier outputs possibilities which might be systematically too excessive or too low, the mixed prediction will inherit that bias.
    Platt scaling may help in case your base classifier isn’t well-calibrated out of the field.

    Stage 2: The Conditional Regression Mannequin. The second stage predicts the worth of $y$ conditional on $y > 0$.
    That is the place the hurdle mannequin shines in comparison with customary approaches since you’re coaching a regression mannequin completely on the optimistic subset of your information, so the mannequin by no means has to take care of the spike at zero.
    This implies you should utilize the complete vary of regression strategies with out worrying about how they deal with zeros.

    The selection of mannequin for this stage relies upon closely on the form of your optimistic outcomes.
    If $log(y | y > 0)$ is roughly regular, you should utilize OLS on the log-transformed goal (with acceptable bias correction on back-transformation, which we’ll cowl beneath).
    For right-skewed optimistic steady outcomes, a GLM with a gamma household is a pure selection.
    For those who’re coping with overdispersed rely information, unfavorable binomial regression works effectively.
    A straightforward technique is simply to make use of Autogluon because the ensemble mannequin and never have to fret concerning the distribution of your information.
    The output is $hat{mu}(X) = E[Y | X, Y > 0]$, the anticipated worth conditional on the end result being optimistic.

    Mixed Prediction. The ultimate prediction combines each phases multiplicatively:

    $$ hat{E}[Y | X] = hat{pi}(X) cdot hat{mu}(X) $$

    This offers the unconditional anticipated worth of $Y$, accounting for each the likelihood that the end result is optimistic and the anticipated magnitude given positivity.
    If a buyer has a 30% likelihood of buying and their anticipated spend given a purchase order is 100 {dollars}, then their unconditional anticipated spend is 30 {dollars}.
    This decomposition additionally makes enterprise interpretation simple.
    You possibly can individually receive characteristic significance on each the likelihood of engagement versus what drives the depth of engagement to see what must be addressed.

    Implementation

    Coaching Pipeline. The coaching pipeline is easy.
    We prepare Stage 1 on the complete dataset with a binary goal, then prepare Stage 2 on solely the optimistic observations with the unique steady goal.
    At prediction time, we get a likelihood from Stage 1 and a conditional imply from Stage 2, then multiply them collectively.

    We will implement this in Python utilizing scikit-learn as a place to begin.
    The next class wraps each phases right into a single estimator that follows the scikit-learn API, making it simple to drop into present pipelines and use with instruments like cross-validation and grid search.

    import numpy as np
    from sklearn.linear_model import LogisticRegression
    from sklearn.ensemble import GradientBoostingRegressor
    from sklearn.base import BaseEstimator, RegressorMixin
    
    class HurdleModel(BaseEstimator, RegressorMixin):
        """
        Two-stage hurdle mannequin for zero-inflated steady outcomes.
    
        Stage 1: Binary classifier for P(Y > 0)
        Stage 2: Regressor for E[Y | Y > 0]
        """
    
        def __init__(self, classifier=None, regressor=None):
            self.classifier = classifier or LogisticRegression()
            self.regressor = regressor or GradientBoostingRegressor()
    
        def match(self, X, y):
            # Stage 1: Practice classifier on all information
            y_binary = (y > 0).astype(int)
            self.classifier.match(X, y_binary)
    
            # Stage 2: Practice regressor on optimistic outcomes solely
            positive_mask = y > 0
            if positive_mask.sum() > 0:
                X_positive = X[positive_mask]
                y_positive = y[positive_mask]
                self.regressor.match(X_positive, y_positive)
    
            return self
    
        def predict(self, X):
            # P(Y > 0)
            prob_positive = self.classifier.predict_proba(X)[:, 1]
    
            # E[Y | Y > 0]
            conditional_mean = self.regressor.predict(X)
    
            # E[Y] = P(Y > 0) * E[Y | Y > 0]
            return prob_positive * conditional_mean
    
        def predict_proba_positive(self, X):
            """Return likelihood of optimistic consequence."""
            return self.classifier.predict_proba(X)[:, 1]
    
        def predict_conditional(self, X):
            """Return anticipated worth given optimistic consequence."""
            return self.regressor.predict(X)

    Sensible Issues

    Characteristic Engineering. One of many good properties of this framework is that the 2 phases can use fully completely different characteristic units.
    In my expertise, the options that predict whether or not somebody engages in any respect are sometimes fairly completely different from the options that predict how a lot they interact.
    For Stage 1, behavioral alerts are likely to dominate: previous exercise, recency, frequency, whether or not the shopper has ever bought earlier than.
    Demographic indicators and contextual elements like time of 12 months or day of week additionally assist separate the “will interact” group from the “received’t interact” group.
    For Stage 2, depth alerts matter extra: historic buy quantities, spending velocity, capability indicators like earnings or credit score restrict, and product or class preferences.
    These options assist distinguish the 50 greenback spender from the five hundred greenback spender, conditional on each of them making a purchase order.
    Moreover, we will use characteristic boosting by feeding within the output of the stage 1 mannequin into the stage 2 mannequin as a further characteristic.
    This enables the stage 2 mannequin to learn the way the likelihood of engagement interacts with the depth alerts, which improves efficiency.

    Dealing with Class Imbalance. If zeros dominate your dataset, say 95% of observations are zero, then Stage 1 faces a category imbalance downside.
    That is frequent in purposes like advert clicks or insurance coverage claims.
    The usual toolkit applies right here: you may tune the classification threshold to optimize in your particular enterprise goal moderately than utilizing the default 0.5 cutoff, upweight the minority class throughout coaching by way of pattern weights, or apply undersampling to resolve this.
    The secret’s to consider carefully about what you’re optimizing for.
    In lots of enterprise settings, you care extra about precision on the high of the ranked listing than you do about total accuracy, and tuning your threshold accordingly could make an enormous distinction.

    Mannequin Calibration. For the reason that mixed prediction $hat{pi}(X) cdot hat{mu}(X)$ is a product of two fashions, each must be well-calibrated for the ultimate output to be dependable.
    If Stage 1’s possibilities are systematically inflated by 10%, your mixed predictions will likely be inflated by 10% throughout the board, no matter how good Stage 2 is.
    For Stage 1, test calibration curves and apply Platt scaling if the uncooked possibilities are off.
    For Stage 2, confirm that the predictions are unbiased on the optimistic subset, which means the imply of your predictions ought to roughly match the imply of the actuals when evaluated on holdout information the place $y > 0$.
    I’ve discovered that calibration points in Stage 1 are the extra frequent supply of issues in observe, particularly when extending the classifier to a discrete-time hazard mannequin.

    Analysis Metrics. Evaluating a two-stage mannequin requires fascinated by every stage individually after which wanting on the mixed output.
    For Stage 1, customary classification metrics apply: AUC-ROC and AUC-PR for rating high quality, precision and recall at your chosen threshold for operational efficiency, and the Brier rating for calibration.
    For Stage 2, you need to consider solely on the optimistic subset since that’s what the mannequin was educated on.
    RMSE and MAE offer you a way of absolute error, MAPE tells you about share errors (which issues when your outcomes span a number of orders of magnitude), and quantile protection tells you whether or not your prediction intervals are sincere.

    For the mixed mannequin, have a look at total RMSE and MAE on the complete check set, but additionally break it down by whether or not the true consequence was zero or optimistic.
    A mannequin that appears nice on combination could be horrible at one finish of the distribution.
    Elevate charts by predicted decile are additionally helpful for speaking mannequin efficiency to stakeholders who don’t assume by way of RMSE.

    When to Use Hurdle vs. Zero-Inflated Fashions. It is a distinction value getting proper, as a result of hurdle fashions and zero-inflated fashions (like ZIP or ZINB) make completely different assumptions about the place the zeros come from.
    Hurdle fashions assume that every one zeros come up from a single course of, the “non-participation” course of.
    When you cross the hurdle, you’re within the optimistic regime, and the zeros are totally defined by Stage 1.
    Zero-inflated fashions, then again, assume that zeros can come from two sources: some are “structural” zeros (prospects who might by no means be optimistic, like somebody who doesn’t personal a automobile being requested about auto insurance coverage claims), and others are “sampling” zeros (prospects who might have been optimistic however simply weren’t this time).

    To make this concrete with a retail instance: a hurdle mannequin says a buyer both decides to buy or doesn’t, and in the event that they store, they spend some optimistic quantity.
    A zero-inflated mannequin says some prospects by no means store at this retailer (structural zeros), whereas others do store right here often however simply didn’t at this time (sampling zeros).
    In case your zeros genuinely come from two distinct populations, a zero-inflated mannequin is extra acceptable.
    However in lots of sensible settings, the hurdle framing is each easier and enough, and I’d advocate beginning there except you might have a transparent motive to imagine in two sorts of zeros.

    Extensions and Variations

    Multi-Class Hurdle. Generally the binary cut up between zero and optimistic isn’t granular sufficient.
    In case your consequence has a number of significant states (say none, small, and enormous), you may lengthen the hurdle framework right into a multi-class model.
    The primary stage turns into a multinomial classifier that assigns every statement to one in every of $Ok$ buckets, after which separate regression fashions deal with every bucket’s conditional distribution.
    Formally, this seems to be like:

    $$ P(Y) = start{instances} pi_0 & textual content{if } Y = 0 pi_1 cdot f_{textual content{small}}(Y) & textual content{if } 0 < Y leq tau pi_2 cdot f_{textual content{massive}}(Y) & textual content{if } Y > tau finish{instances} $$

    That is notably helpful when the optimistic outcomes themselves have distinct sub-populations.
    As an illustration, in modeling insurance coverage claims, there’s usually a transparent separation between small routine claims and enormous catastrophic ones, and making an attempt to suit a single distribution to each results in poor tail estimates.
    The edge $tau$ might be set based mostly on area data or estimated from the information utilizing combination mannequin strategies.

    Generalizing the Levels. One factor value emphasizing is that neither stage must be a particular sort of mannequin.
    All through this text, I’ve introduced Stage 1 as a binary classifier, however that’s simply the best model.
    If the timing of the occasion issues, you possibly can substitute Stage 1 with a discrete-choice survival mannequin that predicts not simply whether or not a buyer will buy, however when.
    That is particularly helpful for subscription or retention contexts the place the “hurdle” has a temporal dimension.
    Equally, Stage 2 doesn’t should be a single hand-tuned regression.
    You would use an AutoML framework like AutoGluon to ensemble over a big set of candidate fashions (gradient boosting, neural networks, linear fashions) and let it discover the very best mixture for predicting the conditional quantity.
    The hurdle framework is agnostic to what sits inside every stage, so you need to be happy to swap in no matter modeling method most closely fits your information and use case.


    Frequent Pitfalls

    These are errors I’ve both made myself or seen others make when deploying hurdle fashions.
    None of them are apparent till you’ve been bitten, so that they’re value studying by way of even if you happen to’re already comfy with the framework.

    1. Leaking Stage 2 Data into Stage 1. For those who engineer options from the goal, one thing like “common historic spend” or “whole lifetime worth,” that you must watch out about how that data flows into every stage.
    A characteristic that summarizes previous spend implicitly incorporates details about whether or not the shopper has ever spent something, which implies Stage 1 could be getting a free sign that wouldn’t be out there at prediction time for brand new prospects.
    The repair is to consider carefully concerning the temporal construction of your options and ensure each phases solely see data that will be out there on the time of prediction.

    2. Ignoring the Conditional Nature of Stage 2. This one is delicate however necessary.
    Stage 2 is educated solely on observations the place $y > 0$, so it needs to be evaluated solely on that subset too.
    I’ve seen folks compute RMSE throughout the complete check set (together with zeros) and conclude that Stage 2 is horrible.
    So whenever you’re reporting metrics for Stage 2, at all times filter to the optimistic subset first.
    Equally, when diagnosing points with the mixed mannequin, ensure you decompose the error into its Stage 1 and Stage 2 parts.
    A excessive total error could be pushed fully by poor classification in Stage 1, even when Stage 2 is doing nice on the optimistic observations.

    4. Misaligned Practice/Take a look at Splits. Each phases want to make use of the identical prepare/check splits.
    This sounds apparent, nevertheless it’s simple to mess up in observe, particularly if you happen to’re coaching the 2 phases in separate notebooks or pipelines.
    If Stage 1 sees a buyer in coaching however Stage 2 sees the identical buyer in its check set (since you re-split the positive-only information independently), you’ve launched information leakage.
    The best repair is to do your prepare/check cut up as soon as originally on the complete dataset, after which derive the Stage 2 coaching information by filtering the coaching fold to optimistic observations.
    For those who’re doing cross-validation, the fold assignments have to be constant throughout each phases.

    5.
    Assuming Independence Between Levels.
    Whereas we mannequin the 2 phases individually, the underlying options and outcomes are sometimes correlated in ways in which matter.
    Prospects with excessive $hat{pi}(X)$ (more likely to interact) usually even have excessive $hat{mu}(X)$ (possible to spend so much once they do).
    This implies the multiplicative mixture $hat{pi}(X) cdot hat{mu}(X)$ can amplify errors in methods you wouldn’t see if the phases had been really impartial.
    Maintain this in thoughts when deciphering characteristic significance.
    A characteristic that reveals up as necessary in each phases is doing double obligation, and its whole contribution to the mixed prediction is bigger than both stage’s significance rating suggests.

    Remaining Remarks

    Alternate Makes use of: Past the examples lined on this article, hurdle fashions present up in a shocking number of enterprise contexts.
    In advertising, they’re a pure match for modeling buyer lifetime worth, the place many purchasers churn earlier than making a second buy, making a mass of zeros, whereas retained prospects generate broadly various quantities of income.
    In healthcare analytics, affected person value modeling follows the identical sample: most sufferers have zero claims in a given interval, however the claims that do are available vary from routine workplace visits to main surgical procedures.
    For demand forecasting with intermittent demand patterns (spare elements, luxurious items, B2B transactions), the two-stage decomposition naturally captures the sporadic nature of purchases and avoids the smoothing artifacts that plague conventional time collection strategies.
    In credit score threat, anticipated loss calculations are inherently a hurdle downside: what’s the likelihood of default (Stage 1), and what’s the loss given default (Stage 2)?
    For those who’re working with any consequence the place zeros have a essentially completely different which means than “only a small worth,” hurdle fashions are value contemplating as a primary method.

    Two-stage hurdle fashions present a principled method to predicting zero-inflated outcomes by decomposing the issue into two conceptually distinct elements: whether or not an occasion happens and what magnitude it takes conditional on prevalence.
    This decomposition provides flexibility, since every stage can use completely different algorithms, options, and tuning methods.
    It provides interpretability, as a result of you may individually analyze and current what drives participation versus what drives depth, which is usually precisely the breakdown that product managers and executives wish to see.
    And it usually delivers higher predictive efficiency than a single mannequin making an attempt to deal with each the spike at zero and the continual optimistic distribution concurrently.
    The important thing perception is recognizing that zeros and optimistic values usually come up from completely different mechanisms, and modeling them individually respects that construction moderately than preventing in opposition to it.

    Whereas this text covers the core framework, we haven’t touched on a number of different necessary extensions that deserve their very own therapy.
    Bayesian formulations of hurdle fashions can incorporate prior data and supply pure uncertainty quantification, which might tie in properly with our hierarchical Bayesian series.
    Think about estimating product-level hurdle fashions the place merchandise with sparse information borrow power from their class.
    Deep studying approaches open up the opportunity of utilizing unstructured options (textual content, pictures) in both stage.
    If in case you have the chance to use hurdle fashions in your individual work, I’d love to listen to about it!
    Please don’t hesitate to succeed in out with questions, insights, or tales by way of my email or LinkedIn.
    If in case you have any suggestions on this text, or wish to request one other matter in causal inference/machine studying, please additionally be happy to succeed in out.
    Thanks for studying!



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleThe New Experience of Coding with AI
    Next Article DataRobot + Nebius: An enterprise-ready AI Factory optimized for agents
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    The New Experience of Coding with AI

    March 18, 2026
    Artificial Intelligence

    Why You Should Stop Worrying About AI Taking Data Science Jobs

    March 18, 2026
    Artificial Intelligence

    One Model to Rule Them All? SAP-RPT-1 and the Future of Tabular Foundation Models

    March 18, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    YouTube lanserar Lens för Shorts: AI-sökning direkt i videon

    June 2, 2025

    MobileNetV3 Paper Walkthrough: The Tiny Giant Getting Even Smarter

    November 2, 2025

    Phase two of military AI has arrived

    April 15, 2025

    Top Scholarships To Study Artificial Intelligence Abroad In 2025 » Ofemwire

    April 4, 2025

    From Connections to Meaning: Why Heterogeneous Graph Transformers (HGT) Change Demand Forecasting

    January 27, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    There and Back Again: An AI Career Journey

    July 14, 2025

    Kan Googles nya Nano Banana verkligen ersätta Photoshop

    August 29, 2025

    The Shape‑First Tune‑Up Provides Organizations with a Means to Reduce MongoDB Expenses by 79%

    May 2, 2025
    Our Picks

    DataRobot + Nebius: An enterprise-ready AI Factory optimized for agents

    March 18, 2026

    Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes

    March 18, 2026

    The New Experience of Coding with AI

    March 18, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.