Close Menu
    Trending
    • Linear Regression Is Actually a Projection Problem, Part 1: The Geometric Intuition
    • Vibe Coding with AI: Best Practices for Human-AI Collaboration in Software Development
    • Generative AI improves a wireless vision system that sees through obstructions | MIT News
    • A better method for identifying overconfident large language models | MIT News
    • DataRobot + Nebius: An enterprise-ready AI Factory optimized for agents
    • Two-Stage Hurdle Models: Predicting Zero-Inflated Outcomes
    • The New Experience of Coding with AI
    • Why You Should Stop Worrying About AI Taking Data Science Jobs
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » Vibe Coding with AI: Best Practices for Human-AI Collaboration in Software Development
    Artificial Intelligence

    Vibe Coding with AI: Best Practices for Human-AI Collaboration in Software Development

    ProfitlyAIBy ProfitlyAIMarch 19, 2026No Comments17 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    — collaborating with an agentic AI-powered IDE to construct software program — is quickly turning into a mainstream growth method. Duties that when required weeks of engineering effort can now typically be accomplished in hours or days. Fashionable AI-assisted growth environments can generate structured, modular code throughout a number of languages, design architectures, write assessments, and even debug points with minimal human enter.

    A rising ecosystem of such instruments has emerged, many constructed on high of acquainted growth environments equivalent to VS Code. Whereas these platforms supply related capabilities, they’re evolving so quickly that any differentiating characteristic in a single software usually seems in competing instruments inside a brief time frame. Because of this, the precise software a company chooses is usually much less essential than how successfully builders be taught to work with these AI techniques to maximise productiveness whereas controlling value and complexity.

    So the pertinent query is that if AI can generate high-quality code quicker than most builders can write it manually, what function stays for the developer?

    The problem is now not merely writing code. As a substitute, builders should discover ways to collaborate successfully with AI coding brokers:

    • How ought to builders construction directions and prompts to information the system towards the specified consequence?
    • The place ought to people intervene within the growth course of?
    • How can groups validate AI-generated code to make sure it’s dependable, maintainable, and production-ready?

    On this article, we discover sensible ideas for working with AI-enhanced growth environments. We are going to define key dangers related to Vibe coding instruments and have a look at methods to mitigate them. Moderately than specializing in any particular software, we are going to look at the broader human-AI collaboration mannequin that permits groups to extract probably the most worth from these techniques.

    As an instance these concepts, we are going to stroll by means of a easy however lifelike use case: constructing an clever search system utilizing Retrieval Augmented Technology (RAG) on a dataset of reports articles. Whereas the issue could seem easy, it reveals a number of refined methods wherein AI-generated architectures and code can drift towards pointless complexity with out cautious human oversight.

    By this instance, we are going to look at each the strengths and limitations of AI-assisted growth, and spotlight the function that builders nonetheless play in guiding, validating, and refining the output of those highly effective instruments.

    The Use Case

    Whereas the ideas mentioned right here apply to any kind of software program growth, let’s illustrate them with a sensible instance: constructing an clever AI-powered search system (RAG) over a dataset of news articles (CC0). The dataset incorporates enterprise and sports activities information articles revealed over 2015 and 2016, together with the title.

    The vibe coder used right here is Google Antigravity however as talked about earlier, this isn’t essential as different instruments additionally perform in a really related manner.

    Dangers related to Vibe Coding

    As with every highly effective know-how, vibe coding introduces a brand new set of dangers which can be straightforward to miss—exactly due to how briskly and succesful the system seems.

    On this instance, as I labored by means of constructing a easy RAG system over information articles, three patterns turned instantly obvious.

    First, the traditional garbage-in-garbage-out precept nonetheless applies. The AI generates code rapidly and confidently—however when the prompts have been even barely ambiguous, the output drifts away from what is definitely wanted. Pace doesn’t assure correctness.

    Second, prompting stays a core talent, regardless that the interface has modified. As a substitute of writing LLM system prompts immediately, we at the moment are prompting the IDE. However the accountability stays the identical: clear, exact directions. In truth, poor prompting has a really tangible value — builders rapidly burn by means of Professional mannequin limits with out getting nearer to a usable answer.

    Third, and extra subtly, over-engineering is an actual threat. As a result of the system can generate complicated architectures effortlessly and at little value, it typically does. Left unchecked, this will result in designs which can be much more complicated than the issue requires —introducing pointless parts that may be tough to take care of later.

    These dangers usually are not theoretical—they immediately affect how the system evolves. The query then turns into: how will we management them?

    What can groups do about them

    To handle these dangers, listed below are a number of core ideas that ought to kind the inspiration of AI-powered SDLC:

    Begin With Clear Necessities

    Earlier than asking the AI to generate structure or code, you will need to set up a minimum of a minimal definition of the issue. In ultimate eventualities, this will come from an present enterprise necessities doc. Nonetheless, in lots of AI initiatives the one requirement the shopper could present is to level to a doc repository and specify a loosely outlined purpose equivalent to “Customers ought to be capable of ask questions in regards to the information articles and obtain contextual responses.” Whereas this will look like an inexpensive place to begin to a human, it’s truly an especially open-ended scope for an AI system to interpret and code and qualifies as a garbage-in immediate. It’s much like working an LLM with none guardrails — there’s a good likelihood the output won’t be what you anticipate. A sensible solution to constrain the scope is to outline a set of consultant take a look at queries that customers are more likely to ask. These queries present the AI with an preliminary scope boundary and scale back the danger of pointless complexity within the ensuing system.

    Generate the Structure Earlier than Writing Code

    Except you’re constructing a trivially easy prototype, it’s prudent to at all times ask to create an structure doc first and optionally, a duties plan to see the sequence wherein it’ll execute the important thing steps equivalent to information ingestion, agent construct, take a look at case execution and outcomes validation. Use a big considering mannequin (equivalent to Gemini-3-Professional in Planning mode) for this step. Even in case you have an structure in thoughts, reasonably than offering that upfront and making a bias, ask the AI to design the structure from a clear slate. After which use your personal design to problem, refine and make sure the design. The developer’s function then turns into one among important analysis — asking questions equivalent to — What if we simplified this element? or What occurs if the info measurement turned 10X?. By this iterative dialogue, the structure step by step converges towards a design that balances functionality, value, and maintainability.

    As an instance the above concepts for our information articles use case, I began with the next preliminary immediate to set fundamental boundaries for the Vibe coder. I particularly requested it to offer an structure doc solely with out starting to code.

    We are going to start on a brand new venture within the C:antigravityVibeCode folder. All artifacts equivalent to supply code, outputs, structure docs must be saved in applicable folders inside this. The Dataset folder has a file referred to as Articles.xlsx. Which include many information articles in enterprise and sports activities together with heading and date revealed. Have a look into it if you would like. There should be a RAG constructed which may reply a various vary of consumer questions from particular, straightforward to retrieve ones to broad ones that will require scanning all of the articles, which is perhaps impractical. I want a technique and structure how you'll do it. Some take a look at queries might be:
    What do the articles say about oil costs in 2015 ?
    Inform me enterprise highlights in 2016
    What sport are many of the articles about ?
    In sports activities, are there point out of information being created ?
    The LLM used will probably be gemini 2.5 flash, however for now lets have the structure doc solely

    An in depth structure doc RAG_Architecture.md was produced and a rationale was offered as to why an agentic dual-engine RAG is required.

    To efficiently reply your mixture of particular questions and broad area queries (like counting subjects) with out scanning all articles each time, a regular vector search RAG shouldn't be going to work. As a substitute, you want an Agentic dual-engine RAG structure.
    
    I've drafted an in depth technique and structure doc that addresses this and saves to the venture folder as requested: 
    
    …VibeCodeArchitectureRAG_Architecture.md
    .
    
    Right here is the high-level abstract of the technique:
    
    1. The Challenges with Your Dataset
    Your dataset has restricted structured columns (Date, Heading, NewsType). Which means that questions like "What sport are many of the articles about?" are almost unattainable to reply natively utilizing both pure SQL or pure Vector Search with out extracting the precise "sport" entity first.
    
    2. Twin-Engine RAG Setup
    Our structure depends on a Information Enrichment Pipeline throughout ingestion and an Clever Routing Agent at question time:
    
    Ingestion (LLM Pre-processing): Through the ingestion part, we move the textual content of every article to Gemini 2.5 Flash to robotically extract subjects (e.g., cricket, soccer, oil) and entities, and optionally summarize the textual content. These new metadata fields are saved in a daily relational database (or Pandas dataframe) and connected as metadata tags to the textual content embeddings in your vector database.
    Clever Router Agent (Gemini 2.5 Flash): An agent sits in entrance of the consumer's queries to find out what software to make use of to retrieve the data:
    The SQL/Information Analytics Engine Instrument: If the question is aggregational or broad (e.g., "Inform me enterprise highlights in 2016" or "What sport are many of the articles about?"), the LLM generates a SQL/Pandas question, counts/filters the extracted metadata, and synthesizes the precise solutions with out touching the vector engine.
    The Vector Semantic Search Instrument: If the question is very qualitative and particular (e.g., "What do the articles say about oil costs in 2015?" or "In sports activities, are there point out of information being created?"), the Router makes use of vector search, however strictly makes use of the structured fields (like Date=2015 or NewsType='sports activities') to filter the embeddings previous to looking.

    Validate the Design

    It’s worthwhile spending time to learn the reasons offered for the architectural decisions made, and the responses to your ideas and challenges. The AI will present detailed explanations of the professionals and cons of every facet of the design, however understanding and assuring oneself that it’s the proper steadiness between performance, complexity and maintainability continues to be the function of the developer, or architect within the group, because the case could also be.

    Strengthen the Design by means of Edge Instances

    As soon as we perceive the structure and rationale, its time to consider Edge take a look at circumstances that may break the structure. As an example, right here I requested "How about if the consumer asks to summarize the damaging articles in your complete database. Which can be tens of 1000's of articles, How will you reply"

    In response, an modification to the structure is proposed. Which is so as to add a sentiment extraction throughout ingestion, hierarchical summarization (generate one sentence abstract of articles to stop context window overflow of the LLM), and a strategic sampling with SQL fallback. This refinement step could also be carried out iteratively for different edge circumstances that come to thoughts.

    Have the AI problem itself

    Attempt asking the AI to consider edge circumstances that break its personal structure. You can too choose a special mannequin, equivalent to Claude Opus for this step. A caveat: AI fashions typically are usually overly important of one other mannequin’s work, so the human’s judgment is crucial in deciding which ideas are worthwhile.

    Once I requested for this to be carried out, the mannequin generated a number of refined take a look at queries, which the structure could not be capable of deal with. Akin to complicated cross doc comparability, precise phrase/key phrase matching, deep relational and community queries, contradiction and declare verification and many others. For every, the reasoning was offered, and the recommended addition to the structure, from Map-Cut back, BM25, Information Graph, Timeline synthesizer software, Declare Evaluation module and many others.

    Listed here are the queries it generated and recommended amendments:

    1. Complicated Cross-Doc Comparability (The "Map-Cut back" Drawback)
        • Question: "Evaluate the first causes given for oil worth drops in 2015 versus the explanations given in 2016. What are the important thing variations?"
        • Potential Addition: A Map-Cut back or Multi-Agent workflow the place one agent summarizes 2015, one other summarizes 2016, and a last agent compares the 2 summaries.
    2. Actual Phrase / Lexical Key phrase Matching (The "BM25" Drawback)
        • Question: "Discover the precise article that used the precise phrase 'unprecedented fiscal cliff' or mentions the obscure ticker image 'ZYXW'."
        • Potential Addition: Hybrid Search. Including a standard lexical search engine (like TF-IDF or BM25) and mixing its scores with the vector search scores (Alpha mixing or Reciprocal Rank Fusion).
    3. Deep Relational & Community Queries (The "Information Graph" Drawback)
        • Question: "Which firms are most often talked about in the identical articles collectively when discussing the tech sector?"
        • Potential Addition: A Information Graph database (like Neo4j) constructed throughout ingestion that maps Entity -> Co-occurs_with -> Entity.
    4. Sequential and Causal Timelines (The "Temporal Reasoning" Drawback)
        • Question: "Hint the chronological sequence of occasions that led to the resignation of the cricket group captain, week by week."
        • Potential Addition: A devoted Timeline Synthesizer Instrument that forces strict chronological retrieval and chains the narrative sequentially.
    5. Contradiction and Declare Verification (The "Truth-Checking" Drawback)
        • Question: "Are there any articles that contradict one another concerning the success of the latest enterprise merger between Firm A and Firm B?"
        • Potential Addition: A Declare Evaluation Module that particularly evaluates retrieved chunks for arguing viewpoints earlier than passing them to the consumer.

    Detect Over-Engineering

    Evaluation critically if the extra complexity is warranted for the extra performance it would allow. On this case, for a easy downside like querying information articles, these modifications appeared extreme. Let’s query the reasoning for the complexity. “Will customers ask for an actual key phrase search to advantage the latency of including a BM25 layer?“. Maybe, a named entity extraction (NER) for metadata will probably be enough. “Are deep relational and community queries essential for information articles, that are by definition standalone artifacts, equivalent to to advantage the complexity and price of a information graph?“

    Problem the Complexity

    It’s completely cheap (and must be a finest follow), to push again on AI generated structure or code, when it appears overly complicated. In such a case, ask for reasoning, workarounds and less complicated alternate options. Right here, I switched the mannequin to Claude and requested to offer a Self-Critique. The end result, as connected right here, is an acknowledgment that the sting queries are unlikely or might be dealt with with less complicated modifications equivalent to higher software utilization by the agent. And on this case, the AI concluded that not one of the complicated architectural modifications are wanted and due to this fact, have been truly rolled again.

    Right here I’m noting the outcomes of the self-critique:

    1. Cross-Doc Comparability (oil 2015 vs 2016): Sure, very possible. Comparative temporal questions are pure for a information corpus. However truthfully, the present structure can deal with this — the agent simply calls the vector software twice with totally different date filters and compares. It would not want a brand new "Map-Cut back module." I over-complicated it.
    2. Actual Phrase / BM25: Unlikely. Customers of a information Q&A system not often seek for precise phrases or ticker symbols. That is extra of a search engine use case. I would drop this as a priority.
    3. Information Graph / Co-occurrence: **Unlikely at this scale.** With 2,700 articles, co-occurrence evaluation is a distinct segment analytical question. A easy SQL JOIN on the `Key_Entities` column may deal with the uncommon case. A Neo4j graph is very large overkill right here.
    4. Temporal/Causal Timelines: **Reasonably possible.** "What occurred main as much as X?" is a pure query for information. However the present structure can deal with it: vector search to search out the subject + SQL kind by date. No new module wanted — simply good agent prompting.
    5. Contradiction Detection: **Unlikely.** Customers not often ask "do articles contradict one another?" That is an educational NLP activity, not a typical consumer question.

    The structure is now prepared for code technology, with the required checks and balances integrated. The ensuing code will probably be correct and explainable.

    Key Perception

    The above narrative is supposed to determine a sample of collaboration between human and AI. At each part, the due diligence and oversight offered by an professional ensures the standard of the following stage’s consequence. As an instance additional, I instructed the vibe coder to publish detailed execution logs of take a look at queries (together with the sting circumstances) exhibiting the software calls and sequence, moreover the ultimate output. I reviewed the code and outcomes manually to grasp the logic, even when the AI assisted in producing explanations for components of it. This challenge-and-refine sample is repeated throughout phases. These ideas ensured that the generated software is maintainable, traceable, and aligned with requirements.

    Right here is an instance of one of many question execution logs:

    USER QUERY: Evaluate the first causes given for oil worth drops in 2015 versus the explanations given in 2016. What are the important thing variations?
    [TOOL PATH] -> search_articles_vector referred to as with question='causes for oil worth drops', news_type='', specific_topic='Oil', yr=2015
    [TOOL PATH] -> search_articles_vector referred to as with question='causes for oil worth drops', news_type='', specific_topic='Oil', yr=2016
    
    FINAL AGENT OUTPUT:
    In 2015, the first causes for oil worth drops included a **surging greenback**, **plentiful world crude provides/oversupply**, **rising US crude inventories**, and **weak demand**. The Worldwide Vitality Company additionally issued a bearish report contributing to the decline.
    
    In 2016, whereas **rising stockpiles**, **big storage overhangs**, **near-record manufacturing**, and **slowing demand** continued to be elements, new and vital causes emerged. The **row between Saudi Arabia and Iran** made cooperation amongst main exporters to chop output extremely unlikely, immediately impacting supply-side expectations. Moreover, **issues over China's financial system** and its **devaluation of the yuan** have been explicitly cited as contributing to the worth drops.
    
    The important thing variations lie within the emergence of **geopolitical tensions between Saudi Arabia and Iran** immediately affecting OPEC's means to handle provide, and the extra pronounced influence of **China's financial insurance policies** on world demand and market sentiment in 2016 in comparison with 2015.

    Takeaways and Conclusion

    The aim of this text is not to function a tutorial on vibe coding — quite a few assets already cowl that. As a substitute, the main focus is on how people ought to work together with AI-powered coding brokers, which have the potential to overwhelm even skilled builders with the velocity and complexity of the options they generate.

    The central precept is evident: AI is a software, not a alternative. Throughout each stage of growth — from design and coding to unit testing, integration, and safety validation — the beneficial workflow is:

    Immediate (Human) -> Generate (AI) -> Evaluation (Human + AI) -> Suggestions Immediate (Human) -> Iterate.

    On this loop, the human stays the ultimate arbiter. Whereas the AI has visibility into necessities, structure, code, and assessments, solely people can assess the broader context: consumer expectations, enterprise priorities, value and latency constraints, reliability, maintainability, and explainability. These elements finally decide whether or not a system succeeds in manufacturing and is extensively adopted by customers.

    Key Takeaways:

    • AI accelerates, people validate: Pace doesn’t exchange judgment.
    • Begin with structure and clear necessities: Outline boundaries and take a look at circumstances earlier than coding.
    • Watch out for over-engineering: Not each AI suggestion is critical; simplicity is a strategic selection.
    • Iterate by means of overview and suggestions: Keep a human-in-the-loop method at each stage.
    • Closing accountability lies with people: Solely people can weigh trade-offs, guarantee maintainability, and determine if the answer is match for manufacturing.

    By following these ideas, builders can harness the total potential of vibe coding whereas sustaining management, guaranteeing techniques are efficient, comprehensible, and finally adopted by the customers they’re constructed for.

    Join with me and share your feedback at www.linkedin.com/in/partha-sarkar-lets-talk-AI

    Reference

    News Articles — Dataset (CC0: Public Domain)

    Photos used on this article are generated utilizing Google Gemini. Code created by me.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleGenerative AI improves a wireless vision system that sees through obstructions | MIT News
    Next Article Linear Regression Is Actually a Projection Problem, Part 1: The Geometric Intuition
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Linear Regression Is Actually a Projection Problem, Part 1: The Geometric Intuition

    March 19, 2026
    Artificial Intelligence

    Generative AI improves a wireless vision system that sees through obstructions | MIT News

    March 19, 2026
    Artificial Intelligence

    A better method for identifying overconfident large language models | MIT News

    March 19, 2026
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Practical Eigenvectors | Towards Data Science

    May 2, 2025

    The Ultimate Guide to Power BI Aggregations

    November 12, 2025

    Should You Turn Your Executives Into AI Avatars?

    September 16, 2025

    Building AI-Powered Low-Code Workflows with n8n

    June 23, 2025

    Understanding Ethical AI: The Importance of Fairness and How to Avoid Common Biases in AI Systems

    April 9, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    Why Your AI Search Evaluation Is Probably Wrong (And How to Fix It)

    March 9, 2026

    OpenAI’s new LLM exposes the secrets of how AI really works

    November 13, 2025

    MIT scientists investigate memorization risk in the age of clinical AI | MIT News

    January 5, 2026
    Our Picks

    Linear Regression Is Actually a Projection Problem, Part 1: The Geometric Intuition

    March 19, 2026

    Vibe Coding with AI: Best Practices for Human-AI Collaboration in Software Development

    March 19, 2026

    Generative AI improves a wireless vision system that sees through obstructions | MIT News

    March 19, 2026
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.