Close Menu
    Trending
    • Dispatch: Partying at one of Africa’s largest AI gatherings
    • Topp 10 AI-filmer genom tiderna
    • OpenAIs nya webbläsare ChatGPT Atlas
    • Creating AI that matters | MIT News
    • Scaling Recommender Transformers to a Billion Parameters
    • Hidden Gems in NumPy: 7 Functions Every Data Scientist Should Know
    • Is RAG Dead? The Rise of Context Engineering and Semantic Layers for Agentic AI
    • ChatGPT Gets More Personal. Is Society Ready for It?
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » Building Research Agents for Tech Insights
    Artificial Intelligence

    Building Research Agents for Tech Insights

    ProfitlyAIBy ProfitlyAISeptember 13, 2025No Comments11 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    ChatGPT one thing like: “Please scout all of tech for me and summarize developments and patterns primarily based on what you suppose I might be serious about,” you realize that you just’d get one thing generic, the place it searches a couple of web sites and information sources and arms you these.

    It’s because ChatGPT is constructed for normal use circumstances. It applies regular search strategies to fetch data, usually limiting itself to a couple net pages.

    This text will present you how one can construct a distinct segment agent that may scout all of tech, combination tens of millions of texts, filter information primarily based on a persona, and discover patterns and themes you possibly can act on.

    The purpose of this workflow is to keep away from sitting and scrolling by boards and social media by yourself. The agent ought to do it for you, grabbing no matter is beneficial.


    We’ll be capable of pull this off utilizing a novel information supply, a managed workflow, and a few immediate chaining strategies.

    The three completely different processes, the API, fetching/filtering information, summarizing | Picture by creator

    By caching information, we are able to preserve the associated fee down to a couple cents per report.

    If you wish to strive the bot with out booting it up your self, you possibly can be part of this Discord channel. You’ll discover the repository here if you wish to construct it by yourself.

    This text focuses on the final structure and how one can construct it, not the smaller coding particulars as you will discover these in Github.

    Notes on constructing

    For those who’re new to constructing with brokers, you would possibly really feel like this one isn’t groundbreaking sufficient.

    Nonetheless, if you wish to construct one thing that works, you’ll need to use numerous software program engineering to your AI functions. Even when LLMs can now act on their very own, they nonetheless want steering and guardrails.

    For workflows like this, the place there’s a clear path the system ought to take, you must construct extra structured “workflow-like” programs. If in case you have a human within the loop, you possibly can work with one thing extra dynamic.

    The explanation this workflow works so nicely is as a result of I’ve an excellent information supply behind it. With out this information moat, the workflow wouldn’t be capable of do higher than ChatGPT.

    Making ready and caching information

    Earlier than we are able to construct an agent, we have to put together an information supply it might probably faucet into.

    One thing I feel lots of people get improper after they work with LLM programs is the idea that AI can course of and combination information fully by itself.

    In some unspecified time in the future, we would be capable of give them sufficient instruments to construct on their very own, however we’re not there but when it comes to reliability.

    So after we construct programs like this, we want information pipelines to be simply as clear as for another system.

    The system I’ve constructed right here makes use of an information supply I already had out there, which implies I perceive how one can train the LLM to faucet into it.

    It ingests 1000’s of texts from tech boards and web sites per day and makes use of small NLP fashions to interrupt down the principle key phrases, categorize them, and analyze sentiment.

    This lets us see which key phrases are trending inside completely different classes over a particular time interval.


    To construct this agent, I added one other endpoint that collects “details” for every of those key phrases.

    This endpoint receives a key phrase and a time interval, and the system kinds feedback and posts by engagement. Then it course of the texts in chunks with smaller fashions that may determine which “details” to maintain.

    The “facts” extracting process for each keyword | Image by author
    We apply a final LLM to summarize which details are most vital, conserving the supply citations intact.


    This can be a type of immediate chaining course of, and I constructed it to imitate LlamaIndex’s quotation engine.

    The primary time the endpoint known as for a key phrase, it might probably take as much as half a minute to finish. However because the system caches the end result, any repeat request takes just some milliseconds.

    So long as the fashions are sufficiently small, the price of operating this on a couple of hundred key phrases per day is minimal. Later, we are able to have the system run a number of key phrases in parallel.

    You may in all probability think about now that we are able to construct a system to fetch these key phrases and details to construct completely different experiences with LLMs.

    When to work with small vs bigger fashions

    Earlier than transferring on, let’s simply point out that choosing the proper mannequin measurement issues.

    I feel that is on everybody’s thoughts proper now.

    There are fairly superior fashions you should use for any workflow, however as we begin to apply increasingly LLMs to those functions, the variety of calls per run provides up rapidly and this will get costly.

    So, when you possibly can, use smaller fashions.

    You noticed that I used smaller fashions to quote and group sources in chunks. Different duties which can be nice for small fashions embrace routing and parsing pure language into structured information.

    For those who discover that the mannequin is faltering, you possibly can break the duty down into smaller issues and use immediate chaining, first do one factor, then use that end result to do the following, and so forth.

    You continue to need to use bigger LLMs when it is advisable to discover patterns in very massive texts, or whenever you’re speaking with people.

    On this workflow, the associated fee is minimal as a result of the info is cached, we use smaller fashions for many duties, and the one distinctive massive LLM calls are the ultimate ones.

    How this agent works

    Let’s undergo how the agent works beneath the hood. I constructed the agent to run inside Discord, however that’s not the main target right here. We’ll give attention to the agent structure.

    I cut up the method into two components: one setup, and one information. The primary course of asks the person to arrange their profile.


    Since I already know how one can work with the info supply, I’ve constructed a reasonably intensive system immediate that helps the LLM translate these inputs into one thing we are able to fetch information with later.

    PROMPT_PROFILE_NOTES = """
    You're tasked with defining a person persona primarily based on the person's profile abstract.
    Your job is to:
    1. Decide a brief persona description for the person.
    2. Choose probably the most related classes (main and minor).
    3. Select key phrases the person ought to monitor, strictly following the foundations under (max 6).
    4. Resolve on time interval (primarily based solely on what the person asks for).
    5. Resolve whether or not the person prefers concise or detailed summaries.
    Step 1. Character
    - Write a brief description of how we should always take into consideration the person.
    - Examples:
    - CMO for non-technical product → "non-technical, skip jargon, give attention to product key phrases."
    - CEO → "solely embrace extremely related key phrases, no technical overload, straight to the purpose."
    - Developer → "technical, serious about detailed developer dialog and technical phrases."
    [...]
    """
    

    I’ve additionally outlined a schema for the outputs I want:

    class ProfileNotesResponse(BaseModel):
     persona: str
     major_categories: Checklist[str]
     minor_categories: Checklist[str]
     key phrases: Checklist[str]
     time_period: str
     concise_summaries: bool

    With out having area information of the API and the way it works, it’s unlikely that an LLM would determine how to do that by itself.

    You may strive constructing a extra intensive system the place the LLM first tries to be taught the API or the programs it’s supposed to make use of, however that might make the workflow extra unpredictable and dear.

    For duties like this, I attempt to all the time use structured outputs in JSON format. That method we are able to validate the end result, and if validation fails, we re-run it.

    That is the best solution to work with LLMs in a system, particularly when there’s no human within the loop to examine what the mannequin returns.

    As soon as the LLM has translated the person profile into the properties we outlined within the schema, we retailer the profile someplace. I used MongoDB, however that’s optionally available.

    Storing the persona isn’t strictly required, however you do must translate what the person says right into a kind that permits you to generate information.

    Producing the experiences

    Let’s have a look at what occurs within the second step when the person triggers the report.

    When the person hits the /information command, with or with out a time interval set, we first fetch the person profile information we’ve saved.

    This provides the system the context it must fetch related information, utilizing each classes and key phrases tied to the profile. The default time interval is weekly.

    From this, we get a listing of prime and trending key phrases for the chosen time interval that could be fascinating to the person.

    Example of trending keywords that can come up from the system in two different categories | Image by author
    With out this information supply, constructing one thing like this could have been tough. The information must be ready prematurely for the LLM to work with it correctly.

    After fetching key phrases, it might make sense so as to add an LLM step that filters out key phrases irrelevant to the person. I didn’t do this right here.

    The extra pointless data an LLM is handed, the more durable it turns into for it to give attention to what actually issues. Your job is to be sure that no matter you feed it’s related to the person’s precise query.

    Subsequent, we use the endpoint ready earlier, which comprises cached “details” for every key phrase. This provides us already vetted and sorted data for every one.

    We run key phrase calls in parallel to hurry issues up, however the first individual to request a brand new key phrase nonetheless has to attend a bit longer.

    As soon as the outcomes are in, we mix the info, take away duplicates, and parse the citations so every reality hyperlinks again to a particular supply through a key phrase quantity.

    We then run the info by a prompt-chaining course of. The primary LLM finds 5 to 7 themes and ranks them by relevance, primarily based on the person profile. It additionally pulls out the important thing factors.

    Short chain of prompting, breaking the task into smaller ones | Image by author
    The second LLM move makes use of each the themes and unique information to generate two completely different abstract lengths, together with a title.

    We are able to do that to ensure to scale back cognitive load on the mannequin.
    This final step to construct the report takes probably the most time, since I selected to make use of a reasoning mannequin like GPT-5.

    You may swap it for one thing quicker, however I discover superior fashions are higher at this final stuff.

    The total course of takes a couple of minutes, relying on how a lot has already been cached that day.

    Take a look at the completed end result under.

    How the tech scounting bot works in Discord | Image by author
    If you wish to have a look at the code and construct this bot your self, you will discover it here. For those who simply need to generate a report, you possibly can be part of this channel.

    I’ve some plans to enhance it, however I’m completely happy to listen to suggestions in the event you discover it helpful.

    And in order for you a problem, you possibly can rebuild it into one thing else, like a content material generator.

    Notes on constructing brokers

    Each agent you construct can be completely different, so that is on no account a blueprint for constructing with LLMs. However you possibly can see the extent of software program engineering this calls for.

    LLMs, at the very least for now, don’t take away the necessity for good software program and information engineers.

    For this workflow, I’m largely utilizing LLMs to translate pure language into JSON after which transfer that by the system programmatically. It’s the best solution to management the agent course of, but additionally not what folks often think about after they consider AI functions.

    There are conditions the place utilizing a extra free-moving agent is good, particularly when there’s a human within the loop.

    Nonetheless, hopefully you discovered one thing, or received inspiration to construct one thing by yourself.

    If you wish to comply with my writing, comply with me right here, my website, Substack, or LinkedIn.

    ❤



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleDocling: The Document Alchemist | Towards Data Science
    Next Article No Peeking Ahead: Time-Aware Graph Fraud Detection
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Creating AI that matters | MIT News

    October 21, 2025
    Artificial Intelligence

    Scaling Recommender Transformers to a Billion Parameters

    October 21, 2025
    Artificial Intelligence

    Hidden Gems in NumPy: 7 Functions Every Data Scientist Should Know

    October 21, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    The Crucial Role of Color Theory in Data Analysis and Visualization

    September 11, 2025

    From Reporting to Reasoning: How AI Is Rewriting the Rules of Data App Development

    July 1, 2025

    Mistral Le Chat blir en riktig ChatGPT-utmanare – med nya minnessystemet

    September 8, 2025

    CLIP Model Overview :  Unlocking the Power of Multimodal AI

    July 14, 2025

    Mastering NLP with spaCy – Part 3

    August 20, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Building connected data ecosystems for AI at scale

    October 10, 2025

    This tool strips away anti-AI protections from digital art

    July 10, 2025

    Generalists Can Also Dig Deep

    September 12, 2025
    Our Picks

    Dispatch: Partying at one of Africa’s largest AI gatherings

    October 22, 2025

    Topp 10 AI-filmer genom tiderna

    October 22, 2025

    OpenAIs nya webbläsare ChatGPT Atlas

    October 22, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.