Plain-language AI glossary.
50 terms every European mid-market leader runs into. Plain definitions, why each matters, and how each relates to the arqmetrica AI Maturity Index.
53 / 53
AI Act
EU regulation (Regulation (EU) 2024/1689) that establishes risk-based rules for AI systems placed on the EU market. It is the first comprehensive horizontal AI law in any major jurisdiction.
Related dimension:Governance & ethics
AI Act conformity assessment
The pre-market evaluation that proves a high-risk AI system meets the AI Act requirements on data quality, documentation, transparency, human oversight, robustness and cybersecurity.
Related dimension:Governance & ethics
AI agent
An AI system that can plan and execute multi-step tasks autonomously, using tools (APIs, code, web browsing) to take actions in the world rather than just produce text.
Related dimension:Tooling & infrastructure
AI centre of excellence
A small, central team that sets AI standards, owns shared infrastructure, builds reusable patterns and supports business units delivering AI projects — without owning every project itself.
Related dimension:People & capability
AI change management
The structured set of activities — communications, training, role redesign, incentive change — that move an organisation from announcing AI to actually using it.
Related dimension:People & capability
AI coach
A specialist who works one-to-one or in small groups to upskill employees on practical AI use within their actual workflows — not generic training.
Related dimension:People & capability
AI incident reporting
The duty under AI Act Article 73 to notify market surveillance authorities of serious incidents and malfunctions involving high-risk AI systems.
Related dimension:Governance & ethics
AI investment prioritisation
A documented framework that ranks candidate AI initiatives against explicit ROI thresholds, risk gates and capability fit — not whoever shouts loudest.
Related dimension:Strategy & vision
AI literacy
The practical understanding employees need to use AI tools well and recognise their limits — covering how the tools work, what they get wrong, and when not to use them.
Related dimension:People & capability
AI literacy mandate
AI Act Article 4 obligation that providers and deployers ensure a sufficient level of AI literacy among their staff and any other person involved in the operation of AI systems on their behalf.
Related dimension:Governance & ethics
AI operating model
The structural choice of how AI capability is organised: a central team, embedded specialists in business units, a federated hub-and-spoke, or a centre of excellence supporting decentralised delivery.
Related dimension:Strategy & vision
AI pilot purgatory
The state in which AI pilots produce demos but never reach production. Symptoms: every quarter starts with new pilots and ends with no measurable business outcomes.
Related dimension:ROI & measurement
AI ROI scorecard
A structured tracker that records, for each AI initiative, the business outcome it targets, the baseline, the measured result, and the net financial impact.
Related dimension:ROI & measurement
AI strategy
A board-endorsed document that says where AI will create value for the business in the next 18 to 36 months, what will not be pursued, who owns delivery, and how much will be invested.
Related dimension:Strategy & vision
Attribution method
The technique used to isolate the share of a business outcome that an AI initiative actually caused — separating it from market effects, seasonality and other simultaneous changes.
Related dimension:ROI & measurement
Bias audit
A structured test of an AI system to detect whether outputs differ systematically across protected groups — for example, by gender, age, ethnicity or nationality.
Related dimension:Data foundations
Chief AI Officer (CAIO)
The named executive accountable for AI outcomes across an organisation — strategy, governance, capability building, vendor selection, and ROI reporting to the board.
Related dimension:Strategy & vision
Context window
The maximum amount of text — measured in tokens — that a language model can consider at once when generating a response. Frontier models in 2026 typically support 200,000 to 2,000,000 tokens.
Data governance
The organisational framework — roles, policies and controls — that defines who owns each data domain, how data quality is enforced, and how access is granted and revoked.
Related dimension:Data foundations
Data lineage
The end-to-end record of where a piece of data came from, how it was transformed, and where it is used. Sometimes called provenance.
Related dimension:Data foundations
Data residency
The geographic location where data is stored and processed. EU data residency means all storage and compute happens inside the European Union.
Related dimension:Data foundations
DPIA (Data Protection Impact Assessment)
A structured risk assessment required by GDPR Article 35 before processing personal data in ways likely to result in a high risk to individuals — including most AI use cases.
Related dimension:Data foundations
Embeddings
Numerical representations of text, images or other content as vectors of numbers, so that semantically similar items end up close together in the vector space.
EU AI Act risk classification
The four-tier system the AI Act uses to sort AI systems by societal risk: prohibited, high-risk, limited-risk and minimal-risk. Each tier triggers a different set of legal obligations.
Related dimension:Governance & ethics
FAIR data principles
A widely adopted set of guidelines saying data should be Findable, Accessible, Interoperable and Reusable. Originated in 2016 in the open-science community.
Related dimension:Data foundations
Fine-tuning
The process of taking an already-trained AI model and continuing training on a smaller, specific dataset so it performs better on a particular task or in a particular style.
Fractional CAIO
A senior AI executive engaged part-time — typically two to four days per month — instead of as a full-time hire. Common for mid-market companies that need executive-grade AI leadership without a 250k salary.
Related dimension:Strategy & vision
General-purpose AI (GPAI) model
An AI model trained on broad data that can perform a wide range of tasks and be integrated into many downstream systems — for example, the foundation models behind ChatGPT, Claude and Gemini.
Related dimension:Governance & ethics
Hallucination
When a language model produces output that is fluent, confident and plausible-sounding — but factually wrong. Sometimes called confabulation.
High-risk AI system
An AI system listed in AI Act Annex III or used as a safety component of a regulated product — for example, AI used in hiring, credit scoring, education, critical infrastructure or law enforcement.
Related dimension:Governance & ethics
Holdout test
An experimental design where part of the population is deliberately excluded from an AI intervention so the difference in outcomes between treatment and holdout can be measured.
Related dimension:ROI & measurement
Human oversight
The requirement that high-risk AI systems are designed and used so that natural persons can effectively monitor them, intervene, and override outputs — defined in AI Act Article 14.
Related dimension:Governance & ethics
Inference
The act of running a trained AI model to produce an output — as opposed to training, which is the process of building the model in the first place.
Kill criteria
Pre-agreed conditions under which an AI initiative will be stopped — defined before it starts, so the decision is automatic rather than political.
Related dimension:ROI & measurement
LLM (large language model)
An AI model trained on vast text corpora to predict the next token in a sequence — the technology underneath ChatGPT, Claude, Gemini and similar assistants.
MCP server (Model Context Protocol)
A standard introduced by Anthropic in late 2024 for connecting AI models to external tools and data sources. An MCP server exposes tools and data that any compatible AI client can use.
Related dimension:Tooling & infrastructure
MLOps
The engineering discipline that brings DevOps practices to machine learning — versioning models and data, automating deployment, monitoring drift, and managing retraining.
Related dimension:Tooling & infrastructure
Model card
A short, structured document that describes an AI model — its intended use, training data, performance metrics, known limitations, and ethical considerations.
Related dimension:Tooling & infrastructure
Model drift
The gradual decay in a model’s performance as the real-world data it sees diverges from the data it was trained on. Comes in two flavours: data drift (the inputs change) and concept drift (the underlying relationship changes).
Related dimension:Tooling & infrastructure
Multimodal AI
AI systems that can process and generate more than one type of content — for example, text plus images, or text plus audio plus video.
PII (personally identifiable information)
Any data that can identify a natural person — directly (name, email, ID number) or indirectly in combination with other data (postcode plus date of birth plus job title).
Related dimension:Data foundations
Prohibited AI practices
AI uses banned outright by AI Act Article 5, including social scoring, untargeted facial-image scraping, emotion recognition in workplaces and schools, and most real-time remote biometric identification in public spaces.
Related dimension:Governance & ethics
Prompt engineering
The discipline of designing, testing and refining the instructions given to a language model to produce reliable, accurate and useful outputs.
Related dimension:People & capability
RAG (retrieval-augmented generation)
A pattern where a language model answers a question by first retrieving relevant documents from a knowledge base and then generating an answer grounded in them.
Related dimension:Tooling & infrastructure
Responsible AI
The umbrella term for designing, building and operating AI systems that are fair, transparent, accountable, safe and aligned with human values — operationalised through frameworks such as NIST AI RMF and ISO/IEC 42001.
Synthetic data
Artificial data generated by a model to mimic the statistical structure of real data, without containing actual records of real people.
Related dimension:Data foundations
Temperature
A setting on a language model that controls how random its outputs are. Low temperature (near 0) produces deterministic, conservative outputs; high temperature produces more varied, creative outputs.
Tokenisation
The process of breaking text into the smaller units (tokens) that a language model actually processes. Roughly, one English word equals 1.3 tokens; one Portuguese word slightly more.
Transformer
The neural network architecture, introduced by Google in 2017, that powers nearly all modern large language models. Its key innovation is the attention mechanism that lets the model weigh the relevance of every word to every other word.
Transparency obligations
AI Act Article 50 duties to disclose AI involvement: users must be told when they are interacting with an AI system, and AI-generated or manipulated content (deepfakes, synthetic media) must be labelled as such.
Related dimension:Governance & ethics
Value realisation
The discipline of tracking an AI initiative beyond go-live, all the way to the financial outcome it was meant to deliver — typically over the following four to eight quarters.
Related dimension:ROI & measurement
Vector database
A database optimised for storing and searching high-dimensional embeddings — the numerical representations of text, images or other content that AI models work with.
Related dimension:Tooling & infrastructure
Vendor lock-in
The state of being tied to a single AI vendor because switching costs — re-training, re-integration, re-procurement — are prohibitive.
Related dimension:Tooling & infrastructure