Skip to content

Latest commit

 

History

History

history

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

History leading to SOTA LLM

Timeline from 1990-2023 ...with key breakthroughs, papers, and popular applications

Statistical Language Models (SLMs) Era: 1990-2010 (with Summaries)

SLMs analyzed natural language using probabilistic methods, calculating sentence probabilities as products of conditional probabilities.

Neural Language Models (NLMs) Era: 2011-2017 (future summaries)

NLMs leveraged neural networks to predict word sequences, introducing word vectors and overcoming SLM limitations.

Pre-trained Language Models (PLMs) Era: 2018-2020 (future summaries)

PLMs introduced the "pre-training and fine-tuning" paradigm, training on large volumes of text before task-specific fine-tuning.

Large Language Models (LLMs) Era: 2020-Present (future summaries)

LLMs are trained on massive text corpora with billions of parameters, approximating human-level performance in various tasks.

  • 2020: GPT-3 [OpenAI, 175B] [Few-shot learning across various NLP tasks]
  • 2020: GShard [Google, 600B] [Multilingual translation]
  • 2021: Switch Transformer [Google, 1.6T] [Efficient language modeling]
  • 2021: Megatron-Turing NLG [Microsoft & NVIDIA, 530B] [Natural language generation]
  • 2022: PaLM [Google, 540B] [Reasoning and problem-solving]
  • 2022: BLOOM [BigScience, 176B] [Open-source multilingual language model]
  • 2023: GPT-4 [OpenAI, undisclosed] [Advanced language understanding and generation]