Close Menu
    Trending
    • OpenAIs nya webbläsare ChatGPT Atlas
    • Creating AI that matters | MIT News
    • Scaling Recommender Transformers to a Billion Parameters
    • Hidden Gems in NumPy: 7 Functions Every Data Scientist Should Know
    • Is RAG Dead? The Rise of Context Engineering and Semantic Layers for Agentic AI
    • ChatGPT Gets More Personal. Is Society Ready for It?
    • Why the Future Is Human + Machine
    • Why AI Is Widening the Gap Between Top Talent and Everyone Else
    ProfitlyAI
    • Home
    • Latest News
    • AI Technology
    • Latest AI Innovations
    • AI Tools & Technologies
    • Artificial Intelligence
    ProfitlyAI
    Home » A Brief History of GPT Through Papers
    Artificial Intelligence

    A Brief History of GPT Through Papers

    ProfitlyAIBy ProfitlyAIAugust 27, 2025No Comments16 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email


    first in a sequence on language fashions and covers the advances main as much as the discharge of Chat-GPT.

    0) Prologue: The Turing take a look at

    In October 1950, Alan Turing proposed a take a look at. Was it attainable to have a dialog with a machine and never be capable to inform it other than a human. He known as this “the imitation recreation”. It was launched within the paper “Computing Equipment and Intelligence”. He was intending to make use of this take a look at as a proxy to the deeper and vaguer query, “can a machine assume”.

    Seventy years later, in 2020, a number of giant language fashions like ChatGPT from OpenAI handed trendy, rigorous variants of this take a look at.

    In 2022, OpenAI launched ChatGPT publicly and it instantly captured the world’s “consideration”.

    When you didn’t get that dad joke, you’ll by the top of this text. Picture by midjourney

    It was the primary chatbot you would have an prolonged dialog with on nearly any subject (the primary apparent utility of a Turing take a look at breaker).

    And since then, we all know how disruptive this know-how has been, with firms like OpenAI and Anthropic coaching and internet hosting these fashions changing into the quickest rising in historical past.

    Whereas it would seem to be it on the floor, such progress doesn’t occur in a vacuum and in a single day. Beneath the quilt, there are gradual advances that ultimately culminate in such an occasion. And certainly, there was a flurry of exercise (when it comes to papers), main as much as the 2020 breakthrough. And since then, a bunch of different necessary developments as these fashions proceed to get new capabilities and enhance.

    For the reason that panorama is beginning to stabilize, its an excellent time to evaluate among the key papers main as much as this breakthrough.

    Within the chart under is a timeline of the papers we’ll be masking on this chapter (14 on the axis means the 12 months 2014 and so forth).

    Historical past of key papers. Picture by creator.

    The important thing structure that brought about a quantum leap to materialize was known as the Transformer. So, what was the deep perception behind it?

    I) Transformers: subtracting, not including

    The structure that lastly heralded the breakthrough that led to conversational chatbots one can have an prolonged dialog with was the Transformer. Picture by midjourney.

    There was a single deep studying structure (known as the transformer) that pushed the boundaries of pure language fashions to new heights just a few brief months after it was launched. It was launched within the well-known 2017 paper, “Consideration is all you want”.

    So what was the important thing advance that facilitated this? What was the “lacking ingredient” that was not there within the earlier state-of-the-art that the transformer launched?

    What’s actually attention-grabbing is that for those who think about the delta between the architectural parts within the earlier state-of-the-art earlier than the Transformer and the Transformer itself, nothing new was added. As an alternative, a selected ingredient (recurrence) was subtracted. That is mirrored within the title of the paper — “Consideration is all you want”, which means you are able to do away with “different stuff” that’s not “consideration”. But when this well-known paper didn’t invent consideration, which one did?

    II) Translation is the place it began

    The use-case pushing the frontiers of AI fashions within the area of language was machine translation. Picture by midjourney.

    Though ChatGPT is a chatbot (a prerequisite for passing the Turing take a look at), the use case that was driving all of the early advances in direction of the transformer structure was language translation. In different phrases, translating from one human language to a different.

    So, there’s a “supply assertion” in language-1 (ex: English) and the purpose is to transform it to a “goal assertion” in language-2 (ex: Spanish).

    That is primarily a “sequence to sequence” activity. Given an enter sequence, return an output sequence. And there are a lot of different issues apart from translation that may be framed as sequence to sequence duties. As an example, a chatbot can also be a sequence to sequence activity. It takes the enter sequence from the consumer and returns the sequence that’s the response from the chatbot.

    Generally, progress in these sorts of fashions occurs iteratively. There’s some structure which is the present state-of-the-art on some activity. Researchers perceive its weaknesses, issues that make it exhausting to work with. A brand new structure is proposed that addresses these weaknesses. It’s run by the benchmarks and turns into the brand new dominant structure if it succeeds. That is additionally how the Transformer got here to be.

    The primary neural community primarily based language translation fashions operated in three steps (at a excessive degree). An encoder would embed the “supply assertion” right into a vector house, leading to a “supply vector” (the encoder). Then, the supply vector can be mapped to a “goal vector” by a neural community (some non-linear mapping) and eventually a decoder would map the ensuing vector to the “goal assertion”.

    Folks rapidly realized that the vector that was imagined to encode the supply assertion had an excessive amount of duty. The supply assertion could possibly be arbitrarily lengthy. So, as a substitute of a single vector for your complete assertion, let’s convert every phrase right into a vector after which have an intermediate ingredient that may pick the particular phrases that the decoder ought to focus extra on. This intermediate architectural ingredient was dubbed “the eye mechanism”.

    It so occurred then that the intermediate mechanism that was accountable for serving to the decoder pick the phrases to concentrate to had very fascinating scaling traits.

    The following concept was to make it the center piece of your complete structure. And that is what led to the present state-of-the-art mannequin, the transformer.

    Let’s have a look at the important thing papers in language translation main as much as the transformer.

    II-A) Consideration is born

    Since “Consideration is all we want” apparently, (see part I), let’s first perceive what consideration even is. We’ve got to go to the paper that launched it.

    2014: “Neural machine translation by collectively studying to align and translate” https://arxiv.org/abs/1409.0473

    This paper first launched the “Consideration mechanism”. Its a manner for various components of a supply sentence to “attend to” what goes right into a sure place within the goal assertion throughout translation.

    Listed below are the important thing factors:
    1) They began with the encoder-decoder mechanism for translating between languages as described above. The important thing limitation known as out was the encoder step (taking a supply assertion and encoding it to a vector in a excessive dimensional house). If the supply assertion was very lengthy (particularly longer than the standard lengths noticed within the coaching information), the efficiency of straightforward encoder-decoder fashions would deteriorate, as a result of a variety of duty was positioned on the goal vector to encode the total context of the supply assertion.
    2) Quoting from the paper on their new strategy: “An important distinguishing function of this strategy from the essential encoder–decoder is that it doesn’t try to encode a complete enter sentence right into a single fixed-length vector. As an alternative, it encodes the enter sentence right into a sequence of vectors and chooses a subset of those vectors adaptively whereas decoding the interpretation. This frees a neural translation mannequin from having to squash all the knowledge of a supply sentence, no matter its size, right into a fixed-length vector. We present this permits a mannequin to manage higher with lengthy sentences.”. In different phrases, they moved away from encoding your complete enter sentence as a vector in direction of encoding the person phrases of the enter sentence as vectors.
    3) On the decoder, in section-3 they are saying: “Intuitively, this implements a mechanism of consideration within the decoder. The decoder decides components of the supply sentence to concentrate to. By letting the decoder have an consideration mechanism, we relieve the encoder from the burden of getting to encode all info within the supply sentence right into a fixed-length vector. With this new strategy, the knowledge may be unfold all through the sequence of annotations, which may be selectively retrieved by the decoder accordingly.”. And that is the primary point out of the eye mechanism. The decoder decides what components of the enter sentence to “concentrate” to because it generates the output sequence.

    The mechanism by which the phrases have been transformed to vectors was primarily based on recurrent neural networks (RNNs). Particulars of this may be obtained from the paper itself. These recurrent neural networks relied on hidden states to encode the previous info of the sequence. Whereas it’s handy to have all that info encoded right into a single vector, it’s not good for parallelizability since that vector turns into a bottleneck and have to be computed earlier than the remainder of the sentence may be processed. And this limits the extent to which the facility of GPUs may be dropped at bear on coaching these fashions.

    II-B) And now its all you want, apparently

    We now get to essentially the most well-known paper that really launched the brand new Transformer structure that may later go on to beat the Turing take a look at.

    2017: “Consideration is all you want” https://arxiv.org/abs/1706.03762

    This one originated in Google deep mind.

    From the title, you possibly can infer that the authors are speaking of consideration prefer it’s already a factor. It was 3 years outdated on the time. So in the event that they didn’t invent “consideration”, then what was their novel contribution? Just like the title suggests, they simplified the structure to “simply consideration”, casting off recurrence fully. Effectively, they did mix consideration with easy feed-forward networks, so the title is a little bit of a lie. Actually, many of the parameters dwell within the feed-forward layers. However they removed the recurrent layers fully. Simply consideration and feed-forward and repeat. In parallel (“multi-head” and likewise in sequence).

    Since consideration had the good property of being parallelizable, they might scale to bigger architectures and prepare them in a extra parallelizable method leveraging the facility of GPUs.

    With this new, less complicated structure, they crossed a brand new state-of-the-art on the key translation datasets.

    Type of wild given their core contribution was eradicating a key element from current fashions and simplifying the structure. This might have simply been simply an ablation research within the earlier paper, that launched consideration.

    As to why this might need occurred to them, one can think about them being annoyed with the hardships the recurrent layers have been inflicting whereas the eye layers would have been very easy to coach. This might need led them to marvel “if the recurrent layers are so problematic, why not eliminate them?”.

    With this new attention-only structure, they crossed the state-of-the-art in language translation duties.

    III) Past translation

    OpenAI began taking these fashions from language translation and increasing them to various language duties. Picture by midjourney.

    That is the place OpenAI first enters this scene. In contrast to different analysis labs inside giant firms, they will chase a mandate of “basic intelligence” on language duties.

    III-A) Generative Pre-training

    On this paper, they launched the primary GPT (Generalized Pre-Educated) mannequin, GPT-1. The mannequin was imagined to be a general-purpose toolkit able to performing any language activity. It had about 117 million parameters.

    2018: “Bettering Language Understanding by Generative Pre-Coaching” https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf

    In contrast to Vaswani et.al. from the earlier paper who have been centered on language translation, the authors of this paper have been considering constructing a basic agent able to excelling at a number of language duties. This is sensible for a analysis group like OpenAI was at that time. The large concept on this paper is: don’t prepare fashions for each activity from scratch.

    First prepare a mannequin that’s typically good at language in an unsupervised method on a big corpus of textual content.

    Word that this step, coaching a basic mannequin on a big corpus of textual content was the topic of a landmark copyright lawsuit (between Anthropic, one of many firms that trains these fashions and the publishers of books it skilled its fashions on), extraordinarily consequential in the way forward for such AI fashions.

    On June 23, 2025, U.S. District Choose William Alsup dominated that Anthropic’s use of lawfully acquired (bought and scanned) copyrighted books to coach its AI fashions constituted “honest use” beneath U.S. copyright legislation. He described the coaching as “quintessentially transformative,” likening it to how “any reader aspiring to be a author” learns and synthesizes content material in their very own phrases. 

    Then, tune it additional in a supervised method on activity particular information. For the reason that transformer is a sequence-to-sequence mannequin, all we have now to do is body the duty as a sequence to sequence downside. For instance, if the duty is sentiment evaluation, the enter turns into the supply sentence whose sentiment must be deciphered. The goal output turns into “POSITIVE” or “NEGATIVE”.

    That is much like how a human first learns basic language abilities, after which makes a speciality of a selected discipline like legislation.

    Take Bobby Fisher, the chess grandmaster who first learnt Russian (since all good chess books of the time have been in that language) after which learn them to get good at Chess.

    III-B) Few shot studying

    2020, Language fashions are few shot learners https://arxiv.org/abs/2005.14165

    That is the paper that first launched the well-known mannequin, GPT-3. A couple of years after this, in November 2022, OpenAI launched ChatGPT to the general public. The mannequin underlying the chatbot was the identical because the one on this paper. This mannequin had 175 billion parameters.

    The authors spend a variety of time marveling at how good people are at typically studying to do novel language duties with only a few illustrative examples. They then dream about AI fashions exhibiting the identical sort of generalizability with out having to re-train the mannequin for each single activity. They argue that scaling the fashions to increasingly more parameters can take us in direction of this purpose.

    Quoting: “In recent times the capability of transformer language fashions has elevated considerably, from 100 million parameters, to 300 million parameters, to 1.5 billion parameters, to eight billion parameters, 11 billion parameters, and eventually 17 billion parameters. Every improve has introduced enhancements in textual content synthesis and/or downstream NLP duties, and there’s proof suggesting that log loss, which correlates effectively with many downstream duties, follows a clean development of enchancment with scale. Since in-context studying includes absorbing many abilities and duties throughout the parameters of the mannequin, it’s believable that in-context studying skills would possibly present equally sturdy positive aspects with scale.”

    The concept is to present the mannequin demonstrative examples at inference time versus utilizing them to coach its parameters. If no such examples are offered in-context, it’s known as “zero shot”. If one instance is offered, “one shot” and if just a few are offered, “few shot”.

    The graph under, taken from the paper, reveals not solely how the efficiency improves because the variety of mannequin parameters goes up, but in addition how the fashions are capable of benefit from the one or few examples proven to them. The efficiency of the one and few shot circumstances begins to tear away from zero shot because the variety of parameters is elevated.

    The efficiency of fashions will get higher because the variety of parameters will increase. But in addition, their potential to leverage few-shot examples begins tearing away. Picture from the paper referenced on this part.

    A captivating experiment was evaluating the fashions efficiency on easy arithmetic duties like two digit addition and subtraction. Quoting: “On addition and subtraction, GPT-3 shows sturdy proficiency when the variety of digits is small, attaining 100% accuracy on 2 digit addition, 98.9% at 2 digit subtraction, 80.2% at 3 digit addition, and 94.2% at 3-digit subtraction.”. Fashions with even 13 billion parameters failed miserably even on two digit addition.

    And this paragraph will need to have made the authors really feel like proud mother and father:

    “To identify-check whether or not the mannequin is just memorizing particular arithmetic issues, we took the 3-digit arithmetic issues in our take a look at set and looked for them in our coaching information in each the varieties “<NUM1> + <NUM2> =” and “<NUM1> plus <NUM2>”. Out of two,000 addition issues we discovered solely 17 matches (0.8%) and out of two,000 subtraction issues we discovered solely 2 matches (0.1%), suggesting that solely a trivial fraction of the proper solutions might have been memorized. As well as, inspection of incorrect solutions reveals that the mannequin typically makes errors equivalent to not carrying a “1”, suggesting it’s truly trying to carry out the related computation fairly than memorizing a desk.”.

    Wanting forward and Conclusion

    These have been among the key papers main as much as the GPT-3 mannequin that was launched to the general public as a chatbot (ChatGPT) in late 2022. Hopefully they supply a glimpse into the iterative evolution that led to the breaking of the Turing take a look at.

    There have been many notable papers since which have eliminated the constraints and improved additional the capabilities of those fashions.

    First, there was a must align the responses of the fashions to human preferences. To forestall the fashions from being poisonous, unhelpful, and many others. That is the place the idea of RLHF (Reinforcement Studying from Human Suggestions) was put into play. It used a method beforehand used to show fashions to play video video games, tailored to tuning the parameters of language fashions. The OpenAI paper was titled: “Coaching language fashions to comply with directions” and got here out in November 2022.

    When you have been an early adopter of those fashions, you would possibly do not forget that for those who requested it about present information it will say: “I’m a language mannequin skilled on a snapshot of the web earlier than 2022” or comparable and was unable to reply questions on occasions since that snapshot. Additional, as we noticed in part III-B, these fashions wouldn’t obtain excellent scores on easy arithmetic. Why depend on the generative course of for these sorts of issues when we have now specialised instruments. As an alternative of merely saying it wasn’t skilled on present affairs, the mannequin might merely name a information API and retrieve the knowledge it wanted. Equally, as a substitute of attempting to do arithmetic by its generative course of, it might merely name a calculator API. That is the place the toolformers paper (https://arxiv.org/abs/2302.04761) by Metas AI lab (FAIR on the time) taught these fashions to make use of instruments like information APIs and calculators.

    This text coated the advances as much as the discharge of ChatGPT, which may be pretty known as a pivotal second in AI fashions. Subsequent up, within the sequence, I’ll be masking follow-up advances like those talked about on this part which have continued to push the boundaries. Keep tuned.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleThe AI Hype Index: AI-designed antibiotics show promise
    Next Article Time Series Forecasting Made Simple (Part 4.1): Understanding Stationarity in a Time Series
    ProfitlyAI
    • Website

    Related Posts

    Artificial Intelligence

    Creating AI that matters | MIT News

    October 21, 2025
    Artificial Intelligence

    Scaling Recommender Transformers to a Billion Parameters

    October 21, 2025
    Artificial Intelligence

    Hidden Gems in NumPy: 7 Functions Every Data Scientist Should Know

    October 21, 2025
    Add A Comment
    Leave A Reply Cancel Reply

    Top Posts

    Hybrid AI model crafts smooth, high-quality videos in seconds | MIT News

    May 6, 2025

    Why AI Is Widening the Gap Between Top Talent and Everyone Else

    October 21, 2025

    Rise of “AI-First” Companies, AI Job Disruption, GPT-4o Update Gets Rolled Back, How Big Consulting Firms Use AI, and Meta AI App

    May 6, 2025

    Skapa olika ljudeffekter med ElevenLabs SB1 soundbord

    May 19, 2025

    51% av all internettrafik består nu av botar

    October 21, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    Most Popular

    AI-Powered Solutions for Enhanced Location Tracking • AI Parabellum

    April 7, 2025

    DeepMinds 145-sidiga rapport om säkerhetsstrategi för AGI

    April 4, 2025

    How to Build an MCQ App

    May 31, 2025
    Our Picks

    OpenAIs nya webbläsare ChatGPT Atlas

    October 22, 2025

    Creating AI that matters | MIT News

    October 21, 2025

    Scaling Recommender Transformers to a Billion Parameters

    October 21, 2025
    Categories
    • AI Technology
    • AI Tools & Technologies
    • Artificial Intelligence
    • Latest AI Innovations
    • Latest News
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us
    Copyright © 2025 ProfitlyAI All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.