When was llm introduced
Content on WhatAnswers is provided "as is" for informational purposes. While we strive for accuracy, we make no guarantees. Content is AI-assisted and should not be used as professional advice.
Last updated: April 17, 2026
Key Facts
- The term 'large language model' gained prominence in 2018 with the release of BERT and GPT.
- Google's BERT was introduced in October 2018 and used bidirectional training for context understanding.
- OpenAI's GPT was released in June 2018, featuring a 12-layer transformer decoder.
- The transformer architecture, introduced in 2017 by Vaswani et al., enabled the development of LLMs.
- By 2020, models like GPT-3 had 175 billion parameters, vastly increasing performance and capabilities.
Overview
Large Language Models (LLMs) emerged as a transformative force in artificial intelligence starting in 2018. These models leverage deep learning and the transformer architecture to process and generate human-like text at scale.
LLMs differ from earlier NLP systems by using self-attention mechanisms and massive datasets, enabling them to understand context, nuance, and complex language patterns. Their introduction marked a turning point in AI-driven language applications.
- 2018 is widely recognized as the breakthrough year for LLMs, with both BERT and GPT released within months of each other.
- BERT (Bidirectional Encoder Representations from Transformers), developed by Google, was introduced in October 2018 and revolutionized how models understand context in sentences.
- GPT (Generative Pre-trained Transformer) by OpenAI launched in June 2018, using a decoder-only transformer to generate coherent text.
- The foundational transformer architecture was introduced in the 2017 paper 'Attention Is All You Need' by Vaswani et al., enabling efficient parallel processing of text.
- These models were trained on gigabytes of text data from books, websites, and encyclopedias, allowing them to generalize across tasks without task-specific training.
How It Works
LLMs function by predicting the next word in a sequence using deep neural networks trained on vast text corpora. They rely on the transformer architecture to process input efficiently and generate human-like responses.
- Transformer Architecture: Introduced in 2017, this design uses self-attention layers to weigh the importance of words in a sentence, improving context understanding.
- Pre-training: Models are first trained on unlabeled text to learn grammar, facts, and reasoning patterns using a masked language modeling objective.
- Parameters: GPT-3, released in 2020, has 175 billion parameters, making it one of the largest models at the time of its launch.
- Fine-tuning: After pre-training, models can be fine-tuned on specific tasks like translation, summarization, or question answering with minimal labeled data.
- Tokenization: Text is broken into subword units called tokens; for example, GPT-3 uses a 50,000-token vocabulary to handle diverse inputs.
- Scaling Laws: Research shows that model performance improves predictably with increased data, parameters, and compute, guiding LLM development strategies.
Comparison at a Glance
Below is a comparison of key LLMs by release year, developer, and technical specifications.
| Model | Year | Developer | Parameters | Architecture |
|---|---|---|---|---|
| GPT | 2018 | OpenAI | 117 million | Decoder-only |
| BERT-Base | 2018 | 110 million | Encoder-only | |
| GPT-2 | 2019 | OpenAI | 1.5 billion | Decoder-only |
| GPT-3 | 2020 | OpenAI | 175 billion | Decoder-only |
| BERT-Large | 2018 | 340 million | Encoder-only |
The table highlights how rapidly LLMs scaled in size and capability between 2018 and 2020. While BERT excelled at understanding tasks like question answering, GPT variants advanced generative abilities, powering chatbots and content creation tools.
Why It Matters
The rise of LLMs has reshaped industries from customer service to education, enabling automation and intelligent interfaces. Their ability to understand and generate language has made them foundational tools in modern AI systems.
- Customer support: Companies use LLMs to power chatbots that handle inquiries, reducing response time from hours to seconds.
- Content creation: Tools like Jasper and Copy.ai use LLMs to generate marketing copy, blog posts, and social media content.
- Education: LLMs support tutoring systems by providing explanations, practice questions, and personalized feedback.
- Accessibility: These models enable real-time translation and text-to-speech systems, improving access for non-native speakers and people with disabilities.
- Research: Scientists use LLMs to extract insights from academic papers, accelerating discovery in medicine and technology.
- Ethical concerns: Issues like bias, misinformation, and job displacement have prompted calls for regulation and responsible AI development.
As LLMs continue to evolve, they are becoming integrated into everyday applications, from search engines to virtual assistants, marking a new era in human-computer interaction.
More When Was in Daily Life
Also in Daily Life
More "When Was" Questions
Trending on WhatAnswers
Browse by Topic
Browse by Question Type
Sources
- WikipediaCC-BY-SA-4.0
Missing an answer?
Suggest a question and we'll generate an answer for it.