normally begins the identical method. In a management assembly, somebody says: “Let’s use AI!” Heads nod, enthusiasm builds, and earlier than you realize it, the room lands on the default conclusion: “Positive — we’ll construct a chatbot.” That intuition is comprehensible. Massive language fashions are highly effective, ubiquitous, and interesting. They promise intuitive entry to common information and performance.
The crew walks away and begins constructing. Quickly, demo time comes round. A cultured chat interface seems, accompanied by assured arguments about why this time, will probably be totally different. At that time, nonetheless, it normally hasn’t reached actual customers in actual conditions, and analysis is biased and optimistic. Somebody within the viewers inevitably comes up with a customized query, irritating the bot. The builders promise to repair “it”, however usually, the underlying problem is systemic.
As soon as the chatbot hits the bottom, preliminary optimism is commonly matched by person frustration. Right here, issues get a bit private as a result of over the previous weeks, I used to be pressured to spend a while speaking to totally different chatbots. I are inclined to delay interactions with service suppliers till the scenario turns into unsustainable, and a few these circumstances had piled up. Smiling chatbot widgets grew to become my final hope earlier than an everlasting hotline name, however:
- After logging in to my automotive insurer’s website, I requested to clarify an unannounced value improve, solely to understand the chatbot had no entry to my pricing information. All it may provide was the hotline quantity. Ouch.
- After a flight was canceled on the final minute, I requested the airline’s chatbot for the explanation. It politely apologized that, for the reason that departure time was already previously, it couldn’t assist me. It was open to debate all different matters, although.
- On a telco website, I requested why my cellular plan had abruptly expired. The chatbot confidently replied that it couldn’t touch upon contractual issues and referred me to the FAQs. As anticipated, these had been lengthy however irrelevant.
These interactions didn’t carry me nearer to an answer and left me on the reverse finish of enjoyment. The chatbots felt like overseas our bodies. Sitting there, they consumed actual property, latency, and a spotlight, however didn’t add worth.
Let’s skip the controversy on whether or not these are intentional darkish patterns. The actual fact is, legacy techniques because the above carry a heavy burden of entropy. They arrive with tons of distinctive information, information, and context. The second you attempt to combine them with a general-purpose LLM, you make two worlds conflict. The mannequin must ingest the context of your product so it could purpose meaningfully about your area. Correct context engineering requires talent and time for relentless analysis and iteration. And earlier than you even get to that time, your information must be prepared, however in most organizations, information is noisy, fragmented, or simply lacking.
On this put up, I’ll recap insights from my e book The Art of AI Product Development and my current speak on the Google Web AI Summit and share a extra natural, incremental strategy to integrating AI into present merchandise.
Utilizing smaller fashions for low-risk, incremental AI integration
“When implementing AI, I see extra organizations fail by beginning too huge than beginning too small.” ( Andrew Ng).
AI integration wants time:
- Your technical crew wants to arrange the information and be taught the out there methods and instruments.
- You have to prototype and iterate to seek out the candy spots of AI worth in your product and market.
- Customers have to calibrate their belief when shifting to new probabilistic experiences.
To adapt to those studying curves, you shouldn’t rush to reveal AI — particularly open-ended chat performance — to your customers. AI introduces uncertainty and errors into the expertise, which most individuals don’t like.
One efficient strategy to tempo your AI journey within the brownfield context is through the use of small language fashions (SLMs), which generally vary from a couple of hundred million to a couple billion parameters. They’ll combine flexibly along with your product’s present information and infrastructure, somewhat than including extra technological overhead.
How SLMs are skilled
Most SLMs are derived from bigger fashions by knowledge distillation. On this setup, a big mannequin acts because the trainer and a smaller one as the coed. For instance, Google’s Gemini served because the trainer for Gemma 2 and Gemma 3 , whereas Meta’s Llama Behemoth skilled its herd of smaller Llama 4 models. Simply as a human trainer condenses years of research into clear explanations and structured classes, the massive mannequin distills its huge parameter area right into a smaller, denser illustration that the coed can take up. The result’s a compact mannequin that retains a lot of the trainer’s competence however operates with far fewer parameters and dramatically decrease computational price.
Utilizing SLMs
One of many key benefits of SLMs is their deployment flexibility. Not like LLMs which are principally used by exterior APIs, smaller fashions will be run domestically, both in your group’s infrastructure or instantly on the person’s gadget:
- Native deployment: You may host SLMs by yourself servers or inside your cloud atmosphere, retaining full management over information, latency, and compliance. This setup is good for enterprise functions the place delicate info or regulatory constraints make third-party APIs impractical.
📈 Native deployment additionally affords you versatile fine-tuning alternatives as you gather extra information and want to reply to rising person expectations.
- On-device deployment by way of the browser: Trendy browsers have built-in AI capabilities you could depend on. As an illustration, Chrome integrates Gemini Nano by way of the built-in AI APIs, whereas Microsoft Edge consists of Phi-4 (see Prompt API documentation). Working fashions instantly within the browser allows low-latency, privacy-preserving use circumstances resembling good textual content ideas, kind autofill, or contextual assist.
If you want to be taught extra in regards to the technicalities of SLMs, listed here are a few helpful sources:
Let’s now transfer on and see what you’ll be able to construct with SLMs to offer person worth and make regular progress in your AI integration.
Product alternatives for SLMs
SLMs shine in targeted, well-defined duties the place the context and information are already identified — the sorts of use circumstances that dwell deep inside present merchandise. You may consider them as specialised, embedded intelligence somewhat than general-purpose assistants. Let’s stroll by the principle buckets of alternative they unlock within the brownfield, as illustrated within the following alternative tree.

1. Higher product analytics
Earlier than exposing AI options to customers, search for methods to enhance your product from the within. Most merchandise already generate a steady stream of unstructured textual content — assist chats, assist requests, in-app suggestions. SLMs can analyze this information in actual time and floor insights that inform each product choices and instant person expertise. Listed below are some examples:
- Tag and route assist chats as they occur, directing technical points to the correct groups.
- Flag churn alerts throughout a session, prompting well timed interventions.
- Counsel related content material or actions primarily based on the person’s present context.
- Detect repeated friction factors whereas the person remains to be within the move, not weeks later in a retrospective.
These inside enablers preserve danger low whereas including worth and giving your crew time to be taught. They strengthen your information basis and put together you for extra seen, user-facing AI options down the street.
2. Take away friction
Subsequent, take a step again and audit UX debt that’s already there. Within the brownfield, most merchandise aren’t precisely a designer’s dream. They had been designed below the technical and architectural constraints of their time. With AI, we now have a chance to carry a few of these constraints, lowering friction and creating quicker, extra intuitive experiences.
instance is the good filters on search-based web sites like Reserving.com. Historically, these pages use lengthy lists of checkboxes and classes that attempt to cowl each potential person choice. They’re cumbersome to design and use, and in the long run, many customers can’t discover the setting that issues to them.
Language-based filtering adjustments this. As an alternative of navigating a posh taxonomy, customers merely sort what they need (for instance “pet-friendly inns close to the seashore”), and the mannequin interprets it right into a structured question behind the scenes.

Extra broadly, search for areas in your product the place customers want to use your inside logic — your classes, constructions, or terminology — and change that with pure language interplay. Every time customers can categorical intent instantly, you take away a layer of cognitive friction and make the product smarter and friendlier.
3. Increase
Together with your person expertise decluttered, it’s time to consider augmentation — including small, helpful AI capabilities to your product. As an alternative of reinventing the core expertise, have a look at what customers are already doing round your product — the aspect duties, workarounds, or exterior instruments they depend on to achieve their aim. Can targeted AI fashions assist them do it quicker or smarter?
For instance, a journey app may combine a contextual journey observe generator that summarizes itinerary particulars or drafts messages for co-travelers. A productiveness device may embody a gathering recap generator that summarizes discussions or motion gadgets from textual content notes, with out sending information to the cloud.
These options develop organically from actual person habits and lengthen your product’s context as an alternative of redefining it.
4. Personalize
Profitable personalization is the holy grail of AI. It flips the normal dynamic: as an alternative of asking customers to be taught and adapt to your product, your product now adapts to them like a well-fitting glove.
Whenever you begin, preserve ambition at bay — you don’t want a completely adaptive assistant. Reasonably, introduce small, low-risk changes in what customers see, how info is phrased, or which choices seem first. On the content material stage, AI can adapt tone and elegance, like utilizing concise wording for specialists and extra explanatory phrasing for newcomers. On the expertise stage, it could create adaptive interfaces. As an illustration, a project-management device may floor probably the most related actions (“create process,” “share replace,” “generate abstract”) primarily based on the person’s previous workflows.
⚠️ When personalization goes flawed, it rapidly erodes belief. Customers sense that they’ve traded private information for an expertise that doesn’t really feel higher. Thus, introduce personalization solely as soon as your information is able to assist it.
Why “small” wins over time
Every profitable AI characteristic — be it an analytics enchancment, a frictionless UX touchpoint, or a personalised step in a bigger move — strengthens your information basis and builds your crew’s iteration muscle and AI literacy. It additionally lays the groundwork for bigger, extra complicated functions later. When your “small” options work reliably, they change into reusable elements in larger workflows or modular agent techniques (cf. Nvidia’s paper Small Language Models are the Future of Agentic AI).
To summarize:
✅ Begin small — favor gradual enchancment over disruption.
✅ Experiment quick — smaller fashions imply decrease price and quicker suggestions loops.
✅ Be cautious — begin internally; introduce user-facing AI when you’ve validated it.
✅ Construct your iteration muscle — regular, compound progress beats headline initiatives.
Initially revealed at https://jannalipenkova.substack.com.
