Large Language Model (LLM)
Core ConceptsA type of AI model trained on vast amounts of text data that can generate, summarize, translate, and reason about language.
Full Explanation
LLMs are neural networks with billions of parameters, trained using self-supervised learning on internet-scale text. They predict the next token in a sequence, which — at sufficient scale — produces models capable of complex reasoning, coding, writing, and conversation. Examples include GPT-4o, Claude, Gemini, and LLaMA.
GPT-4o, Claude Opus 4, and Gemini 2.5 Pro are all large language models.
Related Terms
The basic unit of text that AI language models process — roughly equivalent to 3/4 of a word in English.
The neural network architecture that underpins all modern large language models, introduced by Google in 2017.
Further training a pre-trained AI model on a smaller, task-specific dataset to specialize its behavior.
A large AI model trained on broad data at scale that can be adapted for many different downstream tasks.