https://www.techtarget.com/searchenterpriseai/definition/GPT-3
GPT-3, or the third-generation Generative Pre-trained Transformer, is a neural network machine learning (ML) model trained using internet data to generate any type of text. Developed by OpenAI, it requires a small amount of input text to generate large volumes of relevant and sophisticated machine-generated text.
GPT-3's deep learning neural network is a model with more than 17 billion ML parameters. To put things into perspective, the largest trained language model before GPT-3 was Microsoft's Turing Natural Language Generation (NLG) model, which had 17 billion parameters. As of early 2021, GPT-3 is the largest neural network ever produced. As a result, GPT-3 is better than any prior model for producing text that seems like a human could have written it.
GPT-3 and similar language processing models are commonly referred to as large language models (LLMs). Industry experts criticized GPT-3's developer OpenAI and former CEO Sam Altman for switching from an open source to a closed source approach in 2019. Other LLM developers include Google DeepMind, Meta AI, Microsoft, Nvidia and X.
GPT-3 processes input text to perform a variety of natural language tasks. It uses both NLG and natural language processing to understand and generate natural human language text. Generating content understandable to humans has historically been a challenge for machines that don't know the complexities and nuances of language. GPT-3 has been employed to create articles, poetry, stories, news reports and dialogue, using a small amount of input text to produce large amounts of copy.
GPT-3 can create anything with a text structure -- not just human language text. A key GPT-3 capability is understanding and generating coherent and contextually relevant responses to a wide range of prompts. It's highly versatile in tasks such as writing essays and stories, answering questions, summarizing text, composing poetry and generating programming code.
GPT-3's large size lets it capture complex patterns in text data and generate fluent and contextually appropriate output. This makes it valuable for automating content creation and enhancing natural language understanding tasks. GPT-3's ability to understand and generate humanlike text opens up applications in customer service, content creation, language translation and education.
One notable GPT-3 use case is OpenAI's ChatGPT language model. ChatGPT is a variant of the GPT-3 model, optimized for human dialogue, that can ask follow-up questions, admit mistakes it has made and challenge incorrect premises. ChatGPT was made free to the public during its research preview to collect user feedback. It was designed in part to reduce the possibility of harmful or deceitful responses.
Another common example is OpenAI's Dall-E, an AI image-generating neural network built on a 12 billion-parameter version of GPT-3. Dall-E was trained on a data set of text-image pairs and can generate images from user-submitted text prompts.
Using only a few snippets of example code text, GPT-3 can also create workable code that can be run without error, as programming code is a form of text. Using a bit of suggested text, one developer has combined the user interface prototyping tool Figma with GPT-3 to create websites by describing them in a sentence or two. GPT-3 has even been used to clone websites by providing a URL as suggested text. Developers are using GPT-3 in several ways, including generating code snippets, regular expressions, plots and charts from text descriptions, Excel functions and other development applications.
GPT-3 is starting to be used in healthcare. One 2022 study explored GPT-3's ability to aid in the diagnoses of neurodegenerative diseases such as dementia. It detects common symptoms, such as language impairment in patient speech, as part of the diagnosis process.
AI tools based on GPT-3 are also being used for the following applications:
GPT-3 is a language prediction model. This means that it has a neural network ML model that can take input text and transform it into what it predicts the most useful result will be. These systems are trained using a vast body of internet text to spot patterns in a process called generative pre-training. GPT-3 was trained on several data sets, each with different weights, including Common Crawl, WebText2 and Wikipedia.
GPT-3 is first trained through a supervised testing phase and then a reinforcement phase. When training ChatGPT, a team of trainers asks the language model a question with a correct output in mind. If the model answers incorrectly, the trainers tweak the model to teach it the right answer. The model can also give several answers that trainers rank from best to worst.
GPT-3 has more than 175 billion ML parameters and is significantly larger than its predecessors, including previous LLMs such as Bidirectional Encoder Representations from Transformers (BERT). Parameters are the parts of an LLM that define its skill on a problem, such as generating text. LLM performance generally scales as more data and parameters are added to the model.
When a user provides text input, the system analyzes the language and uses a text predictor based on its training to create the most likely output. The model can be fine-tuned, but even without much additional tuning or training, the model generates high-quality output text that feels similar to what humans would produce.
When a user provides text input, the system analyzes the language and uses a text predictor based on its training to create the most likely output. The model can be fine-tuned, but even without much additional tuning or training, the model generates high-quality output text that feels similar to what humans would produce.
GPT-3 advantages include the following:
While GPT-3 is remarkably large and powerful, it has several limitations and risks associated with its use.
OpenAI, the original developer of GPT-3, has several GPT-3 models. The algorithms of each GPT-3 AI model were developed using different training data and are designed for specific tasks. The most important include the following:
GPT-3 is used by a range of industries such as the following:
Formed in 2015 as a nonprofit, OpenAI developed GPT-3 as one of its research projects. It aimed to tackle the large goals of promoting and developing "friendly AI" in a way that benefits humanity as a whole.
The first version of GPT was released in 2018 and contained 117 million parameters. The second version of the model, GPT-2, was released in 2019 with around 1.5 billion parameters. GPT-3 jumped over GPT-2 by a huge margin with more than 175 billion parameters -- more than 100 times its predecessor and 10 times more than comparable programs.
Earlier pre-trained models, such as BERT, demonstrated the viability of the text generator method and showed the power that neural networks have to generate long strings of text that previously seemed unachievable.
OpenAI released access to GPT-3 incrementally to see how it would be used and to avoid potential problems. The model was released during a beta period that required users apply to use the model, initially at no cost. However, the beta period ended in October 2020, and the company released a pricing model based on a tiered credit-based system that ranges from a free access level for 100,000 credits or three months of access to hundreds of dollars per month for larger-scale access. In 2020, Microsoft invested $1 billion in OpenAI to become the exclusive licensee of the GPT-3 model. This means that Microsoft has sole access to GPT-3's underlying model.
ChatGPT launched in November 2022 and was free for public use during its research phase. This brought GPT-3 more mainstream attention than it previously had, giving many nontechnical users an opportunity to try the technology. GPT-4 was released in March of 2023 and is estimated to have 1.76 trillion parameters. OpenAI hasn't publicly stated the exact number of parameters in GPT-4, however.
There are many open source efforts in play to provide a free and non-licensed model as a counterweight to Microsoft's exclusive licensee status for GPT-3. New language models are published frequently on Hugging Face's platform.
It's unclear exactly how GPT-3 will develop in the future, but it's likely it will continue to find real-world uses and be embedded in various generative AI applications. Many applications already use GPT-3, including Apple's Siri virtual assistant. Where possible, GPT-4 is being integrated where GPT-3 has been in use.
GPT-3 is a generative AI model. Learn the difference between generative AI and predictive AI.
27 Jan 2025