AITutoro
🇬🇧

AI glossary: 52 key terms explained, from tokens to benchmarks

AI moves fast. The vocabulary moves faster — and most glossaries hand you a definition without telling you why the term matters or how it connects to everything else. This AI glossary works differently. Each term builds on the last, so by the end you won't only know what a transformer is — you'll understand why it changed everything.

We organized this glossary by concept clusters, not alphabetically. Each section builds on the one before it, so reading front to back gives you a layered understanding of how AI systems work. Every term includes a plain-language definition, why it matters in practice, and explicit connections to related concepts. That said, feel free to jump to any section — each term stands on its own. Bookmark this page and come back when a term trips you up.

Core concepts

Artificial intelligence (AI)

AI is any system that performs tasks normally requiring human intelligence — recognizing images, translating languages, making decisions. But there's a critical split in how AI systems work.

Traditional AI follows pre-programmed rules. A spam filter checking for banned keywords is traditional AI: reactive, deterministic, limited to what engineers explicitly coded. Generative AI learns patterns from massive datasets and creates new content — text, images, code, audio — from those patterns. It's proactive rather than reactive, producing outputs its creators never explicitly programmed.

When people say "AI" in 2026, they almost always mean generative AI. The rest of this glossary follows that thread.

Why it matters

Understanding the split between traditional AI and generative AI is fundamental. When people say "AI" in 2026, they almost always mean generative AI — knowing this distinction prevents confusion in every conversation about the technology.

Generative AI

Generative AI creates new content by learning patterns from training data, then applying those patterns to produce original outputs. Ask it to write an email, and it generates one word at a time based on statistical patterns it learned during training.

The tools you've likely encountered — ChatGPT, Claude, Gemini, DALL-E — are all generative AI. Each uses a different model architecture, but they share the same fundamental approach: learn patterns, then generate.

Why it matters

Generative AI is the category behind every major AI tool in 2026. Understanding that it creates by predicting patterns — not by understanding — sets realistic expectations for what these tools can and cannot do.

Machine learning (ML)

Machine learning is the subset of AI where systems learn from data instead of following hard-coded rules. Three flavors exist:

  • Supervised learning trains on labeled data. Show the model thousands of photos tagged "cat" and "dog," and it learns to tell them apart.
  • Unsupervised learning discovers patterns in unlabeled data. Hand it customer purchase histories with no categories, and it finds natural groupings on its own.
  • Reinforcement learning learns through trial and error, receiving rewards for good outcomes and penalties for bad ones — the same approach behind game-playing AI.

All generative AI is built on machine learning. The next term narrows the lens further.

Why it matters

Machine learning is the engine underneath all generative AI. Knowing the three flavors — supervised, unsupervised, and reinforcement learning — helps you understand why models behave the way they do and which approach fits which problem.

Deep learning

Deep learning is a subset of machine learning that uses neural networks with many layers — at least four, often hundreds or thousands. These layers let the model learn increasingly abstract representations of data: early layers might detect edges in an image, while deeper layers recognize faces.

Deep learning is what made modern generative AI possible.

Why it matters

Deep learning is the breakthrough that made modern generative AI possible. Without deep multi-layer neural networks, we wouldn't have transformers, LLMs, or any of the tools that define AI today.

How AI models are built

Neural network

A neural network stacks simple processing units called neurons in layers. Each neuron computes a weighted sum of its inputs, adds a bias term, then passes the result through a nonlinear activation function. The network learns by adjusting those weights and biases until its outputs match the expected results. For a deeper technical explanation, see IBM's guide to neural networks.

Think of it like a series of filters. Raw data enters the first layer. Each subsequent layer refines the signal, extracting increasingly useful patterns. The final layer produces the output — a classification, a prediction, or a generated token.

Why it matters

Neural networks are the foundational architecture for all modern AI. Understanding how layers of neurons process and refine data explains why AI systems can learn patterns too complex for traditional programming.

Transformer

The transformer is the neural network architecture behind every major AI model today. Introduced in the 2017 paper Attention is All You Need (Vaswani et al., 2017), it solved a fundamental problem: earlier architectures (RNNs) processed sequences one element at a time, which was slow and made it hard to capture relationships between distant words.

Transformers use a mechanism called self-attention that evaluates all parts of an input simultaneously, determining which elements are most relevant to each other. This parallel processing made transformers faster to train and better at capturing long-range dependencies in text.

Two key subtypes exist. Encoder-only models (like BERT) excel at understanding text — classification, search, sentiment analysis. Decoder-only models (like GPT) excel at generating text — writing, coding, conversation. Most modern chatbots use decoder-only transformers.

Why it matters

The transformer architecture changed everything. Its self-attention mechanism enabled parallel processing that made training on massive datasets feasible — directly leading to the LLMs and AI tools we use today.

Large language model (LLM)

An LLM is a transformer-based model pre-trained on vast amounts of text data — sources like Common Crawl (50+ billion web pages) and Wikipedia (60+ million pages across all languages). LLMs contain hundreds of billions to trillions of parameters and can generate, summarize, translate, and reason over text.

The "large" in LLM refers to both the training data and the parameter count.

Why it matters

LLMs are the models behind ChatGPT, Claude, Gemini, and every major AI chatbot. Understanding their scale — trained on billions of web pages with trillions of parameters — explains both their capabilities and their limitations.

Parameters

Parameters are the internal variables — weights, biases, and embeddings — that a model adjusts during training to improve its predictions. Weights control how strongly each input influences the model's decisions. Biases shift decision thresholds.

Scale comparisons put the numbers in context: GPT-3 has 175 billion parameters. GPT-4 is estimated at 1.76 trillion. DeepSeek R1 has 671 billion.

But more parameters don't automatically mean a better model. Architecture, training data quality, and training techniques matter as much or more. DeepSeek R1, with 671 billion parameters, outperforms some models with higher counts on specific benchmarks.

Why it matters

Parameter count is the most commonly cited model spec, but it's also the most misunderstood. Knowing that architecture and training quality matter as much as raw parameter count prevents you from equating bigger with better.

Embeddings

Embeddings are numerical representations that capture meaning. They convert words, sentences, or entire documents into multi-dimensional vectors — arrays of numbers where semantically similar items cluster close together in vector space.

The word "king" and "queen" would sit near each other. So would "Paris" and "France." This spatial relationship lets AI systems understand similarity, analogy, and context without being explicitly taught those connections.

Key algorithms include Word2Vec (captures word associations) and BERT (captures contextual word meanings — "bank" near "river" vs. "bank" near "money"). Embeddings power everything from search to recommendation systems.

The architecture pipeline is now clear: neural networks provide the learning framework, transformers process sequences in parallel, LLMs scale transformers with massive data, parameters define the model's learned knowledge, and embeddings represent meaning as vectors.

Why it matters

Embeddings are how AI turns language into math. They power search, recommendations, and RAG systems — understanding them explains why AI can find related concepts even when exact keywords don't match.

How AI reads and responds

Token

A token is the smallest unit of text that language models process. One token equals roughly four characters or 0.75 words. The sentence "How are you doing today?" becomes roughly six tokens.

Every interaction with an AI model — input and output — is measured in tokens. This matters because tokens are what you pay for. API pricing is per million tokens processed.

Tokens aren't limited to text. Images use approximately 258 tokens per tile, video uses 263 tokens per second, and audio uses 32 tokens per second. Understanding tokens is essential for managing costs and knowing why some responses get cut off.

Why it matters

Tokens are what you pay for. Every AI API prices by tokens processed, and every context window is measured in tokens. Understanding token economics is essential for managing costs and output quality.

Context window

The context window is the model's working memory — the total number of tokens it can consider at once. Think of tokens as the unit; the context window is the capacity.

Current sizes vary dramatically: Claude offers a 200K-token window (with 1 million in beta), Gemini 2.0 Flash handles roughly 1 million tokens, and GPT-5.2 provides up to 400K tokens. A larger context window means the model can process longer documents and maintain coherence across extended conversations.

The trade-off: larger windows increase accuracy and reduce hallucinations, but they require quadratically more computing power. Doubling the window doesn't double the cost — it roughly quadruples it.

Why it matters

The context window determines how much information a model can work with at once. It directly affects whether you can process long documents, maintain conversation coherence, and reduce hallucinations — and it has major cost implications.

Knowledge cutoff

The knowledge cutoff is the date after which a model's training data contains no information. Events, publications, and developments after this date are unknown to the model — unless provided through web search grounding, RAG, or explicit context injection.

Current cutoffs (February 2026): GPT-5.2 has a cutoff of August 31, 2025. Claude Opus 4.6 has a cutoff of early 2025. Gemini 3 Pro has a cutoff of September 2025.

Critically, models don't know their own cutoff with precision and may respond about post-cutoff events by hallucinating plausible-sounding information. The mitigation is web search grounding: ChatGPT Search (Bing), Gemini's Google integration, and Copilot's Bing access retrieve real-time information to supplement training memory.

Why it matters

The knowledge cutoff explains why AI tools confidently give wrong answers about recent events — they fill in gaps with statistical likelihood rather than facts. Always use web search grounding for current-events questions and verify time-sensitive claims from any AI response.

Prompt engineering

Prompt engineering is the practice of structuring your inputs so the model produces better outputs. The same question, framed differently, can yield dramatically different results.

Key techniques include:

  • Zero-shot prompting: Ask directly with no examples. "Translate this to French."
  • Few-shot prompting: Provide examples of the pattern you want. Show three translations, then ask for a fourth.
  • Chain-of-thought prompting: Ask the model to reason step by step, which improves accuracy on complex problems.
  • Role prompting: Assign a persona. "You are a senior data analyst. Review this dataset."
  • Prompt chaining: Link multiple prompts for complex tasks — first summarize, then analyze, then recommend.

Mastering these techniques is the fastest way to get more value from any AI tool. See our AI prompts collection for ready-to-use examples.

Why it matters

Prompt engineering is the single fastest way to improve your AI results. The same model can produce mediocre or excellent output depending entirely on how you structure your input.

Chain-of-thought prompting

Chain-of-thought (CoT) prompting instructs the model to work through a problem step by step before stating its final answer. Instead of jumping to a conclusion, the model shows its reasoning.

The simplest implementation: append "Think step by step" or "Show your reasoning" to the prompt. Few-shot CoT goes further: show two or three worked examples of step-by-step reasoning, then present the target problem.

Why it works mechanically: autoregressive models generate each token conditioned on prior tokens. When forced to generate intermediate reasoning steps, the model's attention over those visible steps improves the quality of the final answer — the written reasoning acts as a scaffold that the completion must be consistent with.

CoT adds tokens (and therefore cost and latency) to every request. Extended thinking modes in Claude, ChatGPT, and Gemini automate CoT internally, so manual CoT prompting matters most when using base models or APIs without built-in thinking modes.

Why it matters

Chain-of-thought prompting is the single highest-impact prompting technique for complex problems. The visible reasoning improves accuracy — not just transparency — which is why extended thinking modes essentially automate this technique at the model level.

System prompt

A system prompt is a set of instructions given to an AI model before any user interaction begins. It defines the model's persona, tone, task scope, constraints, and persistent context — without the user needing to repeat these in every message.

Examples of effective system prompt content:

  • Persona: "You are a senior financial analyst. Always cite data sources."
  • Constraints: "Only discuss topics related to our product. Redirect off-topic questions politely."
  • Format rules: "Always respond in bullet points. Use British English."
  • Context: "The user's company is Acme Corp. Their primary market is healthcare IT."

System prompts are the primary mechanism for customizing AI behavior in production applications. API users set them programmatically per session. Consumer products like ChatGPT and Claude expose them through Custom Instructions settings.

Why it matters

System prompts are how you encode requirements into AI behavior once instead of repeating them every conversation. For developers building AI products, the system prompt is the primary control surface. For power users, mastering Custom Instructions unlocks consistent, personalized behavior.

Temperature

Temperature controls the randomness of a model's output during inference. It scales the probability distribution over next-token predictions via the softmax function.

Low temperature (0.0–0.3) produces focused, deterministic outputs — ideal for factual tasks, code generation, and data extraction. The model picks the highest-probability token almost every time.

High temperature (above 1.0) flattens the probability curve, giving lower-probability tokens a better chance of being selected. This increases variety but can also produce nonsensical output.

A common misconception: temperature doesn't control "creativity." The model doesn't become more intelligent at higher settings. It becomes more random. The quality ceiling stays the same; the floor drops.

Why it matters

Knowing how temperature works lets you tune AI output for your specific task. Low for factual precision, moderate for balanced writing, and understanding that high temperature means more randomness — not more creativity.

Structured outputs

Structured outputs is an AI capability that constrains the model to return data in a specified format — typically JSON, XML, or a custom schema — rather than free-form text. You define the expected structure; the model guarantees it produces data that conforms to it.

Example: instead of asking "What are the key dates in this contract?" (free-text response), you provide a JSON schema with fields {party_name, effective_date, termination_date, notice_period} — and the model fills each field reliably.

Structured outputs differ from asking the model to "please format as JSON." That approach produces valid JSON most of the time. True structured outputs use constrained decoding to guarantee format compliance — the model physically cannot produce tokens that would break the schema.

Supported by: GPT-5.2 API (strict mode), Gemini API (JSON mode with schema), Claude API (tool use as structured output mechanism).

Why it matters

Structured outputs transform AI from a text generator into a reliable data processor. When downstream code depends on parsing AI output, you need format guarantees, not probabilities — structured outputs are what make AI integrations production-grade.

Inference

Inference is what happens every time you send a prompt — the model generating output from your input. If training is school, inference is the job.

Three phases occur during inference: prefill (processing all input tokens simultaneously), decode (generating output tokens one at a time), and output conversion (turning tokens into readable text).

Inference must be fast because it happens in real time. And every token generated costs money. Training happens once and takes days or weeks. Inference happens millions of times a day and must complete in seconds.

Why it matters

Inference is where all the cost accumulates. Training happens once; inference happens every single time someone sends a prompt. Understanding this distinction explains why API pricing and response speed matter so much.

Latency

Latency in AI measures the time between sending a request and receiving a response. Two figures matter in practice:

Time to first token (TTFT): How long until the model begins streaming output. Critical for interactive interfaces — 500ms TTFT feels sluggish; under 200ms feels instant. ChatGPT's Advanced Voice Mode targets a 232ms average TTFT.

End-to-end latency: Total time to complete the full response. Scales with output length because autoregressive generation produces tokens sequentially — a 1,000-token response always takes longer than a 100-token one.

Three levers drive latency: model size (larger models are slower on equivalent hardware), hardware (specialized inference chips reduce TTFT significantly), and output length (unavoidably linear). Extended thinking modes add 2–30 seconds depending on reasoning depth.

Why it matters

Latency determines whether an AI system feels responsive. For real-time applications like voice and live coding, TTFT is the critical metric. For batch processing, end-to-end time matters. Knowing which levers control latency helps you trade off speed against output quality.

Training, fine-tuning, and keeping models current

Training data

Training data is the information a model learns from. It can be labeled (each data point tagged with the correct answer, used in supervised learning) or unlabeled (raw data the model finds patterns in on its own).

Quality matters more than quantity. A model trained on carefully curated, diverse, well-structured data outperforms one trained on a larger but noisy dataset. Preparation involves collection, cleaning, transformation, feature engineering, and splitting into training, validation, and test sets.

Why it matters

Training data quality is the single biggest factor in model performance. Garbage in, garbage out applies to AI more than anywhere else — biased or noisy data produces biased or unreliable models.

Training (AI model training)

Training is the process of teaching the model by exposing it to data and adjusting its parameters. The cycle repeats millions of times: input data flows through the network (forward pass), the model's predictions are compared against expected results (error calculation), errors propagate backward through the layers (backpropagation), and weights update via gradient descent.

This is compute-intensive work. Training Llama 3.1 (405 billion parameters) required roughly 38 yottaflops — that's 3.8 x 10^25 math operations. Training runs take days or weeks on clusters of specialized hardware.

The key distinction: training happens once (or rarely). Inference happens every time someone sends a prompt. The next two terms cover ways to adapt a trained model without starting from scratch.

Why it matters

Training is the most expensive and time-consuming phase of building an AI model. Understanding the forward pass, backpropagation cycle explains why retraining from scratch is avoided and why fine-tuning and RAG exist as alternatives.

Fine-tuning

Fine-tuning adapts a pre-trained model to a specific task or domain. Instead of training from zero, you take an existing model and refine it with specialized data.

Three main approaches:

  • Full fine-tuning updates all parameters. Effective but expensive.
  • Parameter-efficient fine-tuning (PEFT) updates only a small subset. LoRA (Low-Rank Adaptation) is the most popular technique — it injects small trainable matrices into transformer layers, reducing the number of parameters that need updating by orders of magnitude. LoRA can run on consumer GPUs with 24 GB of memory.
  • RLHF (Reinforcement Learning from Human Feedback) trains a reward model from human rankings, then optimizes the LLM to maximize those reward scores. This is how ChatGPT learned to be helpful rather than harmful.

The myth that fine-tuning requires massive compute is outdated. LoRA changed the game.

Why it matters

Fine-tuning lets you customize a general-purpose model for your specific domain without starting from scratch. With LoRA, this is now accessible on consumer hardware — the barrier to custom AI has collapsed.

RAG (retrieval-augmented generation)

RAG supplements a model's knowledge at inference time with external data. Instead of relying solely on what the model learned during training, RAG retrieves relevant information from an external knowledge base and injects it into the prompt.

The process follows four steps: embed the external data as vectors, retrieve the most relevant chunks based on the user's query, augment the prompt with retrieved context, and generate a response that incorporates that context.

RAG is cost-effective (no retraining needed), keeps information current, and enables source attribution. But it doesn't replace training — it supplements it. A poorly trained model won't produce good outputs even with perfect retrieval.

The progression: fine-tuning changes the model itself. RAG changes what the model sees at inference time.

Why it matters

RAG is the most practical way to keep AI responses current and grounded in your own data — without the cost and complexity of retraining. It's the foundation of most enterprise AI deployments.

Reliability and safety

Hallucination

A hallucination occurs when a model generates plausible-sounding but factually incorrect content. The model isn't lying — it's predicting the most statistically likely next token without any mechanism for fact-checking.

Causes include overfitting to training data, biased datasets, and the fundamental nature of statistical prediction. Leading models report hallucination rates as low as 0.7%–0.9%, while many widely used models fall between 2% and 5%.

Two factors from earlier sections connect directly: higher temperature increases randomness and can raise hallucination rates. Larger context windows help reduce hallucinations by giving the model more relevant information to work with. Understanding both concepts helps you manage hallucination risk in practice.

Why it matters

Hallucinations are the primary reliability risk in AI. Models confidently generate false information with no internal fact-checking. Knowing the causes — and how temperature and context window affect rates — is essential for responsible AI use.

AI bias

AI bias refers to systematic errors that produce unfair outcomes. Three primary sources feed it:

  • Biased training data — if the dataset underrepresents certain groups, the model inherits those gaps.
  • Algorithmic bias — design choices in the model architecture that reinforce existing patterns.
  • Human interpretation bias — people applying model outputs without questioning assumptions.

Mitigation requires diverse, representative training data, regular fairness audits, and human oversight at decision points. No model is bias-free, but awareness of these sources is the first step toward responsible use.

Why it matters

AI bias produces real-world harm when models make decisions affecting people. Understanding the three sources — data, algorithm, and human interpretation — is the first step toward building and using AI responsibly.

Guardrails

Guardrails are the technical and procedural controls that keep AI systems within safe boundaries. They work in three layers:

  • Input filtering screens what the model receives — blocking prompt injection attempts, PII exposure, and harmful requests.
  • Processing constraints limit model behavior during generation — enforcing topic boundaries and compliance rules.
  • Output enforcement validates responses before they reach the user — checking for harmful content, factual consistency, and policy alignment.

Guardrails don't make models infallible. They make the risks manageable. Every production AI deployment needs them.

Why it matters

Every production AI deployment needs guardrails. They're the difference between a useful tool and an unpredictable liability — managing input filtering, processing constraints, and output validation in three protective layers.

Natural language processing (NLP)

NLP

Natural language processing is the broader field that makes human-computer language interaction possible. It combines computational linguistics, statistical modeling, and deep learning into a pipeline: text preprocessing (cleaning and structuring raw text), feature extraction (identifying meaningful patterns), text analysis (applying models to understand meaning), and model training (improving accuracy with feedback). Google's machine learning glossary provides additional NLP definitions worth bookmarking.

Key tasks within NLP include named entity recognition (identifying people, places, and organizations in text), sentiment analysis (determining emotional tone), and part-of-speech tagging (classifying words by grammatical role). LLMs operate under the NLP umbrella — they're the most powerful NLP tools ever built, but NLP as a discipline predates them by decades.

Why it matters

NLP is the broader discipline that LLMs belong to. Understanding the pipeline — preprocessing, feature extraction, analysis, training — gives you context for why LLMs work the way they do and what came before them.

Multimodal AI

Multimodal AI processes and generates multiple types of data simultaneously — text, images, audio, video, and increasingly 3D spatial data. Unlike single-modality models that handle only text, multimodal systems interpret combinations of inputs.

Current examples: GPT-4o processes text, images, and audio. Claude handles text and images. Gemini works across text, images, audio, and video. The trajectory is clear — future AI systems will be natively multimodal, processing information the way humans do: through multiple senses at once.

Why it matters

The trajectory of AI is toward native multimodality. Understanding that models increasingly process text, images, audio, and video together prepares you for the next generation of AI tools and workflows.

How AI models are measured

MMLU (massive multitask language understanding)

MMLU tests general knowledge across 57 subjects — from STEM and law to nutrition and religion — using 15,908 multiple-choice questions. Released in September 2020 by Dan Hendrycks et al., it quickly became the standard measure of how well a model handles diverse factual knowledge. The original MMLU benchmark paper (Hendrycks et al., 2020) details the methodology.

By mid-2024, top models had nearly saturated the original benchmark, scoring so high that differences between them became statistically insignificant. This saturation sparked several spin-offs: MMLU-Pro (harder questions), MMMLU (multilingual version), and MMLU-Redux (corrected errors in the original).

A high MMLU score means a model performs well on factual recall across disciplines. It doesn't measure reasoning depth, creative ability, or real-world task completion. Treat it as one data point, not a verdict.

Why it matters

MMLU is the most widely cited benchmark for comparing AI models. Knowing its scope — factual recall across 57 subjects — and its limitations helps you interpret model comparison claims critically rather than at face value.

HumanEval

HumanEval measures coding ability through 164 hand-crafted Python programming challenges, each with a function signature, docstring, and unit tests (averaging 7.7 tests per problem). OpenAI developed and released it in 2021 alongside the Codex model. The test suite is available in OpenAI's HumanEval repository.

It uses the pass@k metric: the probability that at least one of k generated code samples passes all unit tests. This approach accounts for the variability in AI-generated code — the model might get it right on the third try even if the first two fail.

HumanEval remains the most-cited coding benchmark, though successor benchmarks like BigCodeBench are emerging to test more complex programming scenarios.

For a deeper look at how benchmark scores compare across models, see the AI tools comparison page.

Why it matters

HumanEval is the standard benchmark for AI coding ability. Understanding its pass@k metric and its 164-problem scope helps you evaluate AI coding tool claims and understand why code generation quality varies.

SWE-bench

SWE-bench measures AI coding ability using 2,294 real GitHub issues and pull requests sourced from popular open-source Python projects. Models must read the issue, understand the codebase, write a patch, and pass the project's existing test suite — without human assistance at any step.

SWE-bench Verified (500 human-validated problems) is the standard version cited in model releases. The metric is resolve rate: the percentage of issues the model correctly patches on the first attempt.

As of February 2026: Claude Opus 4.6 leads at 80.8%, GPT-5.2 at 80.0%, Gemini 3 Pro at 74.2%. Unlike HumanEval's 164 hand-crafted Python problems, SWE-bench tests production-grade software engineering — the gap between a 75% and 80% score translates to real differences in autonomous coding capability.

Why it matters

SWE-bench is the most credible measure of real-world coding ability. Its problems come from actual GitHub issues, making it the benchmark most directly predictive of whether an AI tool can handle production codebases.

ARC-AGI-2

ARC-AGI-2 (Abstraction and Reasoning Corpus) tests AI on novel visual grid puzzles that humans typically solve in minutes but that require genuine abstract reasoning — not pattern recall from training data. Each puzzle shows a series of colored grid transformations and asks the model to identify the underlying rule.

The benchmark is deliberately resistant to memorization: problems are generated from new rules each evaluation, so a model can't improve by studying past examples. Results are verified by the ARC Prize Foundation.

As of February 2026: Gemini's Deep Think mode leads at 84.6%, Claude Opus 4.6 scores 68.8%, GPT-5.2 Pro scores 54.2%. The gap between 54% and 84% represents fundamentally different abstract reasoning capability.

Why it matters

ARC-AGI-2 is the closest current benchmark to testing genuine problem-solving rather than sophisticated pattern recall. A high score suggests a model can handle truly novel situations, not just interpolate from training data.

Terminal-Bench

Terminal-Bench 2.0 measures AI ability on command-line and DevOps tasks: file manipulation, shell scripting, network diagnostics, process management, and system configuration. Problems run in live Linux environments and are evaluated on whether the AI's commands produce the correct system state — not just the correct output text.

This execution-based evaluation makes it harder than benchmarks that only check generated text. The model must run commands, observe results, and self-correct autonomously.

As of February 2026: Claude leads at 65.4%, GPT-5.2 at 56.2%. The 9-point gap reflects Claude's strength on multi-step agentic tasks in constrained environments.

Why it matters

Terminal-Bench scores are the most direct predictor of AI performance on infrastructure automation, DevOps, and systems administration. If your workflows involve the command line, this benchmark is more predictive than general-purpose scores.

GDPval-AA

GDPval-AA (General Document Processing and Valuation — Advanced Analysis) measures AI performance on business document tasks: financial statement analysis, contract review, earnings call interpretation, and strategic document synthesis. Scoring uses an Elo-based head-to-head comparison system — human evaluators compare two models' outputs on the same document and indicate which is stronger.

As of February 2026: Claude Opus 4.6 scores 1606 Elo versus Gemini 3 Pro's 1195 — a 411-point gap that represents a substantial practical difference in business analysis quality.

Why it matters

GDPval-AA is the benchmark most relevant to knowledge workers in business, finance, and legal contexts. Its use of real business documents and Elo scoring makes it more predictive of enterprise AI performance than generalist knowledge benchmarks.

Elo rating

Elo is a pairwise ranking system originally developed for chess. In AI evaluation, platforms like LMArena present outputs from two anonymous models on the same prompt to human raters, who choose which response they prefer. The Elo algorithm updates both models' scores based on the outcome — an upset (weaker model beats stronger) moves scores more than an expected result.

As of February 2026: Gemini 3 Pro leads LMArena at 1501, Grok 4.1 leads thinking mode at 1483, GPT-5.2 and Claude Opus 4.6 cluster near 1440–1460.

Elo ratings capture overall user preference on open-ended tasks. They correlate imperfectly with task-specific benchmarks — a model can lead on Elo but trail on SWE-bench or ARC-AGI-2.

Why it matters

Elo ratings are the most direct measure of which AI users prefer for open-ended tasks. They complement task-specific benchmarks — a model that tops Elo but underperforms on SWE-bench excels at conversation but less at autonomous coding.

AI generation architectures

Autoregressive generation

Autoregressive generation is how LLMs produce text: predict one token at a time, using all previously generated tokens as context for the next prediction. Each output token depends on all prior tokens — hence "autoregressive."

The decode loop: (input prompt + all tokens generated so far) → probability distribution over vocabulary → sample one token → append to context → repeat until done.

This constraint explains two core LLM behaviors. First, generation is sequential — you can't produce token 100 without first producing tokens 1–99, which is why longer responses take longer. Second, each token is statistically likely given prior context — the model has no end-to-end check whether the complete sentence will be factually correct, which is a root cause of hallucinations.

OpenAI applied autoregressive generation to image pixels rather than text tokens for GPT Image — contrasting with diffusion models that start from noise and iteratively refine it.

Why it matters

Autoregressive generation explains why LLMs produce text sequentially, why response time scales with length, and why mid-sentence hallucinations happen. It's the architectural fact behind behaviors that otherwise seem arbitrary.

Diffusion model

A diffusion model generates images (and increasingly audio and video) by reversing a noise-adding process. During training, the model learns to predict how to "de-noise" a partially noisy version of a real image, one step at a time. During generation, it starts from pure random noise and applies the learned de-noising process repeatedly until a coherent image emerges.

Key examples: DALL-E 3 (API deprecated November 2025), Midjourney V6 and V7, Stable Diffusion, Adobe Firefly.

The core contrast with autoregressive generation: diffusion models refine all parts of the image simultaneously across multiple steps, while autoregressive image models generate pixels sequentially. Diffusion models excel at coherent full-image compositions; they historically struggle with text rendering and precise instruction-following — two areas where OpenAI cited advantages when switching ChatGPT's image generation to an autoregressive architecture (GPT Image) in March 2025.

Why it matters

Diffusion models are the dominant architecture for artistic image generation. Knowing how they differ from autoregressive generation explains Midjourney's strengths and weaknesses versus ChatGPT Image — and why they produce such distinctly different aesthetic results.

AI capabilities and workflows

Agentic AI

Agentic AI refers to systems that autonomously execute multi-step tasks — planning, taking action, observing results, and correcting course — without human approval at each individual step. The contrast is with single-query chatbots that answer one question and wait for the next.

A coding agent, for example, receives a feature request, reads the codebase, writes code, runs tests, interprets failures, and revises until tests pass. Current examples: Claude Code, ChatGPT Codex, Gemini's Jules, Grok's DeepSearch.

Four properties distinguish agents from chatbots: persistence (maintaining state across multiple steps), tool use (reading files, executing code, browsing the web), self-correction (adjusting based on intermediate results), and goal-directedness (pursuing an outcome rather than answering a single prompt).

The glossary entry "AI agents are fully autonomous" captures the key limitation: current agents require well-defined scope, guardrails, and human oversight for production use. They're capable interns, not independent colleagues.

Why it matters

Agentic AI represents the shift from AI as a Q&A tool to AI as a task executor. Understanding the distinction between agent and chatbot is essential for evaluating which tools can actually automate workflows versus which ones only assist them.

Extended thinking

Extended thinking is a reasoning mode where the model works through intermediate steps internally before producing its final answer. Instead of predicting the most statistically likely next token immediately, the model generates a chain of reasoning — checking assumptions, exploring alternatives, catching errors — then produces an output informed by that reasoning trace.

Implementations differ across products: Claude Opus 4.6 uses adaptive thinking (four effort levels the model self-selects based on problem complexity). Gemini Ultra's Deep Think mode applies extended reasoning with a 192K-token internal budget. ChatGPT's Thinking mode offers Standard, Light, and Extended variants.

Extended thinking increases latency (seconds to tens of seconds) and token consumption. For simple lookups, the overhead isn't worth it. For complex reasoning, math, planning, and coding tasks, the accuracy improvement is significant.

Why it matters

Extended thinking is the practical mechanism behind AI models that feel more deliberate. Knowing when to activate it — complex problems only — balances speed and quality. It's also the internal automation of chain-of-thought prompting.

Deep Research

Deep Research is an agentic workflow where an AI model autonomously conducts multi-source research and synthesizes a structured report — without the user manually browsing and reading each source.

The workflow: receive a complex research question → decompose into sub-queries → search multiple sources → read and evaluate retrieved pages → synthesize findings → produce a cited report. Full execution takes 2–15 minutes depending on scope.

Current implementations: ChatGPT Deep Research (powered by o3 reasoning; Free 5/month, Plus 25/month, Pro 250/month), Gemini Deep Research (AI Pro and Ultra tiers), Grok's DeepSearch (distinctive in also searching X/Twitter posts alongside web sources).

The quality difference from simple web search grounding: Deep Research reads and reasons over dozens of sources in sequence, not just the top few results. Output is a synthesized report with structure and citations — not a list of retrieved snippets.

Why it matters

Deep Research compresses hours of research work into minutes. Understanding how it differs from basic web search — agentic multi-source synthesis vs. single-query retrieval — sets realistic expectations for what it produces and where it falls short.

Web search grounding

Web search grounding supplements AI responses with live search results, reducing reliance on potentially stale training data. When activated, the model searches the web, retrieves relevant pages, and incorporates their content into its response — typically with source citations.

Current implementations: ChatGPT Search (Bing-powered, available on all tiers including Free since February 2025), Gemini Search (Google-powered, default in Gemini app), Copilot (Bing-powered, always on), Perplexity (search-first AI).

The distinction from RAG: web search grounding retrieves from the live public internet. RAG retrieves from your own private documents. Both reduce hallucinations by giving the model specific text to work from rather than relying on training memory. Neither eliminates hallucination — the model still synthesizes retrieved content and can misinterpret sources.

Why it matters

Web search grounding is the practical solution to the knowledge cutoff problem. It's also a partial hallucination mitigation: responses grounded in retrieved text are more reliable than responses from training memory alone — but not perfectly reliable.

Source grounding

Source grounding restricts an AI model to responding only from specific documents you provide, rather than its training data or the web. Every response includes citations pointing back to exact passages in those documents.

NotebookLM is the purest example: upload PDFs, research papers, or meeting notes, and every answer the AI gives links to the source paragraph that supports it. If the answer isn't in your sources, the model says so rather than hallucinating from training data.

Source grounding makes hallucination structurally harder: the model can't fabricate information it wasn't given. Errors still occur — the model can misinterpret source text — but they're detectable because every claim has a traceable citation. This is the key advantage over web search grounding, which retrieves broadly rather than precisely.

Why it matters

Source grounding is the highest-reliability approach to document analysis. When answers need to be auditable and verifiable — legal review, academic research, compliance — source-grounded AI reduces hallucination risk more than any other technique available today.

Enterprise, integration, and compliance

MCP (Model Context Protocol)

MCP is an open standard that defines how AI models connect to external tools and data sources. Anthropic describes it as "USB-C for AI": a single connector specification that works across compatible systems rather than requiring a different integration for every model-tool combination.

Published by Anthropic in 2024, MCP lets developers build one integration that works with any MCP-compatible AI system. Current connectors include Slack, GitHub, Figma, Asana, Notion, databases, and file systems. Claude supports 50+ connectors; Copilot Studio has generally available MCP support.

Before MCP, connecting an AI to external tools required custom integration work per AI provider. MCP standardizes the interface so one integration serves all compatible models — reducing development overhead and preventing vendor lock-in.

Why it matters

MCP compatibility is increasingly a key criterion when choosing an AI platform. It's the open-standard alternative to proprietary plugin ecosystems — integrations built to MCP spec are portable across AI systems rather than locked to one provider.

Data residency

Data residency refers to requirements — legal, contractual, or organizational — specifying which country or region AI data (prompts, outputs, user data) must be stored and processed in.

Relevant regulations: GDPR requires data protection for EU residents but doesn't automatically mandate geographic storage. Some EU member states, plus sector-specific rules in healthcare and finance, impose stricter geographic constraints. Japan's APPI and similar national laws add regional layers.

Current AI provider options: ChatGPT Enterprise offers EU Data Boundary, US, and Japan. Microsoft Copilot supports data residency across EU, UK, US, Canada, Japan, South Korea, Singapore, India, Australia, and UAE. Google Workspace AI supports EU, US, and multi-region configurations.

For most individuals and SMBs, data residency is not a concern. For regulated industries (healthcare, finance, government) and EU-based enterprises, it's often a hard procurement requirement.

Why it matters

Data residency determines whether an AI tool can legally operate with your organization's data in regulated industries or jurisdictions. It's not just a technical spec — missing regional storage options can block adoption entirely.

Common AI buzzwords vs. reality

"More parameters always means a better model"

Parameters matter, but architecture, training data quality, and training technique matter more. DeepSeek R1 (671 billion parameters) outperforms some models with higher counts on specific benchmarks.

Why it matters

This myth leads people to assume the biggest model is the best model. In reality, a well-architected, well-trained smaller model can outperform a bloated one — parameter count alone is not a reliable quality signal.

"AI models understand what they generate"

LLMs predict statistically likely next tokens based on learned patterns. They don't possess semantic understanding. This is why hallucinations happen — the model produces plausible text without verifying facts.

Why it matters

Believing AI "understands" leads to over-trust. Models predict likely next tokens — they don't verify facts or grasp meaning. Recognizing this prevents costly mistakes when using AI for critical decisions.

"Temperature equals creativity"

Temperature controls randomness, not intelligence. High temperature doesn't make the model more creative — it makes it more random. The quality ceiling stays fixed; the floor drops.

Why it matters

This misconception causes people to crank up temperature expecting better creative output. In reality, higher temperature only increases randomness — the model's best possible output stays the same while its worst output gets worse.

"Fine-tuning requires massive compute"

LoRA and other parameter-efficient techniques can fine-tune models on consumer GPUs with 24 GB of memory. The barrier to entry has collapsed.

Why it matters

This outdated assumption stops people from customizing models for their use case. LoRA and similar techniques have made fine-tuning accessible on consumer hardware — the compute barrier is far lower than most assume.

"RAG replaces training"

RAG supplements a model's knowledge at inference time. It can't fix a fundamentally weak model — the base still needs proper training.

Why it matters

RAG is powerful but not a substitute for a well-trained base model. Treating RAG as a replacement for training leads to poor results when the underlying model can't reason well with the retrieved information.

"Context window equals memory"

The context window is temporary working memory for a single conversation. When the session ends or the window fills up, the model retains nothing. It's a whiteboard, not a hard drive.

Why it matters

Confusing the context window with persistent memory leads to frustration when the model "forgets" previous conversations. The context window is temporary and session-bound — nothing persists once it's gone.

"AI agents are fully autonomous"

Current AI agents operate with a degree of autonomy, but they require human oversight, guardrails, and well-defined boundaries. They're more like capable interns than independent colleagues.

Why it matters

The "autonomous agent" hype sets dangerous expectations. Current AI agents need human oversight, guardrails, and clear boundaries — deploying them without these safeguards creates real risk.

Build real skill with AI tools

AITutoro provides adaptive training for both ChatGPT and Claude. The platform adjusts to what you already know, so you skip the basics and focus on the techniques that move your work forward.

Frequently asked questions

Ready to master your AI workflow?

Whether you chose ChatGPT, Claude, or both, targeted skill-building turns a good tool into a competitive advantage.