A Large Language Model (LLM) is an artificial intelligence model designed to understand, generate, and manipulate human language. These models are trained on massive datasets and use deep learning techniques, specifically the transformer architecture, to predict the next word in a sequence.
The term has evolved alongside the increase in model size and data volume, notably gaining prominence with models like BERT and GPT-2 around 2018-2019.
LLMs have become the foundation of modern AI, powering charbots, coding assistants, and content generation tools used by millions daily.