Agent
Software that can plan, call tools, and act on a goal — not just respond. The keyword is "loop": observe → plan → act → observe.
Every buzzword in agentic AI translated into plain English. 26 terms and growing.
Software that can plan, call tools, and act on a goal — not just respond. The keyword is "loop": observe → plan → act → observe.
Software that does things on your behalf, often via tool calls and multi-step loops. Marketing uses it loosely; engineers mean: LLM + tools + loop.
OpenAI's framework for building apps inside ChatGPT. Uses MCP-style servers as the tool/app interface and adds ChatGPT-specific UI/widget rendering.
Agent-to-Agent protocol from Google. Lets agents discover and delegate to each other. Useful for multi-agent systems.
Asking the model to "show its working" before answering. Improves reasoning on complex tasks at the cost of tokens.
How much text the model can "see" at once. Bigger ≠ always better — there's a "lost in the middle" problem.
Microsoft's low-code platform for building agents inside M365 / Power Platform. Tightly integrated into corporate IT estate.
An agent defined by a JSON manifest (no code). Microsoft's pattern for M365 Copilot extensibility.
A vector representation of text. Used for semantic search, similarity, and RAG. "Closer in vector space" ≈ "more similar in meaning".
When an LLM produces structured JSON describing which function to invoke. The mechanism behind tool calling.
When a model confidently produces wrong information. Most-cited LLM failure mode. Reduced — not eliminated — by RAG, grounding, and tool calls.
Constrained-generation mode where the model only produces valid JSON. Useful when you need a parseable response.
100K+ tokens, sometimes millions. Doesn't mean the model uses it well — see "lost in the middle".
Phenomenon where models pay less attention to information in the middle of long contexts. Strong performance at start + end.
Model Context Protocol — Anthropic's open spec for connecting AI agents to tools, data, and apps. The most-likely-to-stick tool protocol of 2026.
Mixture of Experts — model architecture where only some "expert" sub-networks fire per token. Powers Mixtral, GPT-4, etc.
Two or more agents collaborating, often with role specialisation (planner / worker / reviewer). Useful when, but not always.
The coordination layer that decides which agent runs when, with what input. Examples: LangGraph, CrewAI, AutoGen.
When attacker text overrides the system prompt. Direct (in user input) or indirect (via fetched content). The #1 agent security concern.
Retrieval-Augmented Generation — fetch relevant docs before answering. Reduces hallucinations on factual questions.
Reasoning + Acting — agent loop pattern: thought → action → observation → thought... Most common modern agent loop.
Anthropic's name for a packaged set of agent capabilities (instructions + files + tools). Marketplace launched 2025.
The "operating instructions" given to the model, hidden from end users, sets behaviour and constraints.
When an LLM decides which tool to use, generates arguments, the runtime executes it, and the result is fed back. Foundation of agentic behaviour.
Unit of text the model sees. Roughly 0.75 of a word in English. Pricing and context limits are measured in tokens.
Database optimised for similarity search over embeddings. Examples: Pinecone, Qdrant, Weaviate, pgvector.