Plain-language definitions for the acronyms and terms shaping AI today. Each entry is tagged by the AI literacy level at which it most naturally lives.
An AI system that can plan, take actions, and use tools autonomously to accomplish a goal, often across multiple steps.
AI systems designed to operate with autonomy — making decisions and taking actions in pursuit of objectives rather than just answering questions.
A hypothetical AI with broad, human-level cognitive abilities across most domains, as opposed to narrow AI focused on a single task.
The broad field of building computer systems that perform tasks normally requiring human intelligence.
The research area focused on ensuring AI systems behave in accordance with human values and intentions.
A computing system loosely modeled on the human brain, made of interconnected layers of nodes (neurons) that learn patterns from data.
A defined way for software to talk to other software. AI APIs let developers send prompts to a model and receive responses programmatically.
A neural network technique that lets a model weigh the importance of different parts of an input when producing each part of the output — the foundation of modern transformer models.
The core training algorithm for neural networks, which adjusts model weights by propagating errors backward through the layers.
The number of training examples processed before the model's parameters are updated in a single training step.
A standardized test used to compare AI model performance, such as MMLU, HumanEval, or GSM8K.
Systematic errors or unfair outcomes in model behavior, often stemming from training data or design choices.
A prompting and reasoning technique where the model explicitly works through intermediate steps before producing a final answer.
A conversational interface, often powered by an LLM, that interacts with users through natural language.
A machine learning task where the model assigns inputs to one of a set of predefined categories.
A neural network architecture especially well suited to image and spatial data, using convolutional layers to detect local patterns.
The text or output that a model produces in response to a prompt.
The processing power (typically GPU/TPU hours) consumed by training or running AI models.
The maximum amount of text (measured in tokens) that a language model can consider at once when producing a response.
A large, structured collection of text used as training or evaluation data.
A collection of examples used to train, validate, or test a machine learning model.
A subfield of machine learning that uses neural networks with many layers to learn complex patterns from large datasets.
A generative model that learns to create images (or other data) by gradually denoising random noise — the technology behind tools like Stable Diffusion and DALL·E.
A technique where a smaller "student" model is trained to mimic the outputs of a larger "teacher" model, producing a faster, cheaper version with similar quality.
A numerical (vector) representation of text, images, or other data that captures semantic meaning, enabling similarity search and retrieval.
One complete pass of the training algorithm over the entire training dataset.
The process of measuring a model's performance on specific tasks, often using benchmarks or custom test sets.
A prompting technique where a model is given a few examples of the desired task in the prompt to guide its response.
The process of further training a pre-trained model on a smaller, specialized dataset to adapt it to a specific task or domain.
A large, general-purpose model trained on broad data that can be adapted to many downstream tasks.
A model architecture where two neural networks (a generator and a discriminator) compete, producing increasingly realistic synthetic data.
AI systems that create new content — text, images, audio, video, code — rather than just analyzing or classifying existing data.
A family of large language models developed by OpenAI, based on the transformer architecture.
Specialized hardware originally for graphics, now the workhorse for training and running large AI models due to massive parallelism.
The optimization algorithm used to train neural networks by iteratively adjusting weights to minimize a loss function.
Connecting a model's output to verifiable, external sources of truth (such as documents or databases) to reduce hallucination.
Rules, filters, or constraints applied to AI systems to keep their behavior within safe and appropriate bounds.
When a generative AI model produces output that is fluent and confident but factually incorrect or fabricated.
A workflow design where a human reviews, corrects, or approves AI outputs at key points.
A configuration value (such as learning rate or batch size) set before training, as opposed to weights learned during training.
The process of using a trained model to generate predictions or outputs from new inputs (as opposed to training).
A fine-tuning technique that trains a model to follow natural-language instructions across many task types.
The date after which a model's training data ends, beyond which it has no built-in knowledge of world events.
The time it takes for a model to produce a response after receiving a request.
A group of neurons in a neural network that processes inputs from the previous layer and passes outputs to the next.
A hyperparameter that controls how much model weights change in response to error during training.
A neural network trained on vast amounts of text to predict and generate human-like language.
The practices and tooling for deploying, monitoring, and maintaining LLM-powered applications in production.
A parameter-efficient fine-tuning method that adds small, trainable matrices to a frozen base model, dramatically reducing training cost.
A mathematical function that measures how far a model's predictions are from the correct answers, used to guide training.
A subfield of AI in which systems learn patterns from data rather than following explicitly programmed rules.
An open protocol that standardizes how AI applications connect to external data sources, tools, and services.
The trained mathematical artifact (weights and architecture) that performs a task such as generating text or classifying images.
A standardized document describing a model's intended use, training data, performance, limitations, and risks.
A model architecture in which many specialized sub-networks ("experts") share the work, with a router selecting which experts process each input.
Describes models that can process and/or generate more than one type of data — for example, text, images, audio, and video together.
AI designed and trained for a single, specific task (e.g., spam filtering), as contrasted with AGI.
The branch of AI focused on enabling computers to understand, interpret, and generate human language.
A model composed of layers of interconnected nodes that learn to map inputs to outputs through training.
An AI model whose weights (and often training code) are publicly released, allowing inspection, modification, and self-hosting.
When a model learns training data too closely, including its noise, and consequently performs poorly on new, unseen data.
A learned weight or value inside a model. Modern LLMs have billions to trillions of parameters.
A family of fine-tuning techniques (including LoRA) that update only a small fraction of a model's parameters.
The initial phase of training a foundation model on a large, broad dataset before any task-specific fine-tuning.
The input text or instructions given to a generative model to elicit a desired response.
The practice of designing and refining prompts to reliably produce high-quality outputs from a model.
A security attack where malicious instructions hidden in input data trick a model into ignoring its original instructions.
A technique that combines a generative model with a retrieval system, letting the model pull in relevant external documents at query time.
A model designed to spend additional compute "thinking" through a problem before answering, often improving accuracy on complex tasks.
A metric that measures the fraction of relevant items that a system successfully retrieved or identified.
A machine learning task that predicts a continuous numerical value (e.g., a price) rather than a category.
A training paradigm where an agent learns by taking actions in an environment and receiving rewards or penalties.
A training method that uses human preferences to shape a model's behavior, widely used to make LLMs more helpful and safe.
A neural network architecture designed for sequential data, processing inputs one step at a time while maintaining a hidden state.
The discipline of designing AI systems that behave reliably, avoid harm, and remain under meaningful human control.
A training approach where the model generates its own labels from unlabeled data — for example, predicting the next word in a sentence.
Search that matches meaning rather than exact keywords, typically powered by embeddings.
A language model with relatively few parameters, optimized for speed, cost, or on-device deployment.
The current best published performance on a given task or benchmark.
A training approach where the model learns from labeled examples (input/output pairs).
A foundational instruction given to a model that defines its role, behavior, and constraints across a conversation.
A sampling parameter that controls randomness in a model's output — lower values are more deterministic, higher values more creative.
A chunk of text (a word, subword, or character) that a language model processes as a single unit.
The component that splits raw text into tokens (and converts tokens back to text) for a language model.
The capability of an AI model to call external functions, APIs, or services to gather information or take actions.
Sampling strategies that limit the model's word choices to the most probable options, controlling output variety.
Custom AI accelerator hardware developed by Google, optimized for tensor operations used in deep learning.
The process of teaching a model by exposing it to data and adjusting its parameters to minimize errors.
Reusing a model trained on one task as the starting point for training on a related task.
The neural network architecture, introduced in 2017, that underpins modern LLMs and most state-of-the-art generative models.
A training approach where the model finds structure in unlabeled data (e.g., clustering similar items).
A database optimized for storing and searching embeddings, enabling fast similarity search at scale.
The numerical parameters inside a neural network that are adjusted during training and determine the model's behavior.
Techniques and practices for making AI model decisions understandable to humans.
A model's ability to perform a task without having seen any examples of it during training or in the prompt.