AI

Terminology

AI Terminology Guide

Artificial intelligence is evolving rapidly, and with it comes a growing set of terms that can be confusing or misused. This guide is designed to provide clear, straightforward definitions of the most important AI concepts, from foundational ideas like machine learning and documentation to newer topics like generative AI and prompt engineering. Whether you are new to AI or looking to deepen your understanding, this page will help you navigate the language of modern AI with confidence.

Does your business need help safely implementing AI? Reach out to learn how STACK Cybersecurity can support your team's AI journey.

Adversarial AI

Techniques and attacks used to manipulate AI systems, causing them to make incorrect or unintended predictions or decisions. These techniques exploit vulnerabilities in AI models, often by subtly altering input data, training data, or model interactions to manipulate the AI system.

Agentic AI

A category of AI systems capable of independently making decisions, interacting with their environment, and optimizing processes without direct human intervention.

AI Agent

A system that autonomously perceives its environment, decides what to do, and takes actions to achieve its goals.

AI drift/decay

The tendency for an AI model's performance to degrade over time when deployed in a real-world setting with differing conditions from those present in training and testing.

AI model/system exploitation

Adversarial actions that exploit vulnerabilities an AI model or system to force misperformance against its intended objectives, disrupt access to its outputs or functionality, or enable unauthorized access to restricted or proprietary information.

AI governance

The set of organizational policies, rules, frameworks, roles, and oversight processes that direct how AI is adopted, developed, deployed, and monitored within the organization, with the objective of ensuring AI-related risks are identified, managed, and monitored across the AI lifecycle.

AI lifecycle

The set of phases an AI system goes through. These are plan and design, collect and process data, build and use model, verify and validate, deploy and use, and operate and monitor. These phases are often iterative, and not necessarily sequential.

AI model

A component of an information system that implements AI technology and uses computational, statistical, or machine-learning techniques to produce outputs from a given set of inputs.

AI risk assessment

A risk-management process for identifying, estimating, and prioritizing risks arising from the operation and use of an AI system, incorporating threat and vulnerability analyses and considering mitigations provided by controls planned or in place.

AI as a service (AIaaS)

Cloud-based systems providing on demand services to organizations and individuals to deploy, develop, train, and manage AI models.

AI system

The term 'artificial intelligence system'

(A) means any data system, software, application, tool, or utility that operates in whole or in part using dynamic or static machine learning algorithms or other forms of artificial intelligence, whether

(i) the data system, software, application, tool, or utility is established primarily for the purpose of researching, developing, or implementing artificial intelligence technology; or

(ii) artificial intelligence capability is integrated into another system or agency business process, operational activity, or technology system; and

(B) does not include any common commercial product within which artificial intelligence is embedded, such as a word processor or map navigation system.

AI use case

A specific scenario in which AI is designed, developed, procured, or used to achieve a particular objective, such as delivering a product or service, enhancing decision making, or providing a defined benefit.

AI use case inventory

A maintained repository or listing of an organization's AI use cases, intended to support governance, transparency, and risk management by documenting where and how AI is designed, developed, procured, or used, and the purpose and outputs associated with those uses.

Algorithm

A clearly specified mathematical process for computation; a set of rules that, if followed, will give a prescribed result.

Algorithmic trading system

A system that fundamentally depends upon computerized algorithms, and the data and technological infrastructure through which they operate, to address various decisions and tasks associated with trading financial instruments.

Anomaly detection system

A system for identifying the occurrence of a condition that deviates from expectations based on requirements specifications, design documents, user documents, or standards, or from someone's perceptions or experiences.

Artificial general intelligence (AGI)

The currently hypothetical level of AI capability that is able to understand or learn an intellectual task as human being can. It is an AI system that can perform across diverse cognitive domains with versatility and proficiency, rather than being limited to a narrow task or domain.

Artificial intelligence (AI)

The term 'artificial intelligence' means a machine-based system that can, for a given set of human-defined objectives, make predictions, recommendations or decisions influencing real or virtual environments. Artificial intelligence systems use machine and human-based inputs to

(A) perceive real and virtual environments;

(B) abstract such perceptions into models through analysis in an automated manner; and

(C) use model inference to formulate options for information or action.

Benchmarking

An alternative prediction or approach used to compare a model's inputs and outputs to estimates from alternative internal or external data or models.

Bias

A systematic distortion, as opposed to random error, that reduces the representativeness or accuracy of an AI system's outputs or performance for its intended purposes and operating conditions. Bias may be introduced inadvertently or purposely, and may also emerge as the AI system is used in an application; this could arise when the data used to develop or operate the system are not representative of the intended population or operating conditions. Common sources/subcategories or bias include statistical/computational, systemic, and human bias (not exhaustive).

Black box

The nature of some AI techniques whereby the inferential operations are complex, hidden, or otherwise opaque to their developers and end users in terms of providing an understanding of how classifications, recommendations, or actions are generated and what overall performance will be.

Capability evaluation

A comprehensive assessment of an AI model's or system's overall capabilities, including both planned capabilities and unplanned, emerging, or malicious capabilities. Unlike specific task-focused evaluations this evaluation seeks to understand the full range of an AI's capabilities. This includes evaluating how an AI might adapt or evolve beyond its initial training, identifying both beneficial emergent behaviors and potential risks that could arise from its autonomous operation or interaction with complex environments.

Computer vision

The digital process of perceiving and learning visual tasks in order to interpret and understand the world through cameras and sensors.

Data lineage

The history of processing of a data element, which may include point-to-point data flows and the data actions performed upon the data element.

Data poisoning

An attack that corrupts and contaminates training data to compromise an AL system's performance.

Data quality/validity

The usefulness, accuracy, and correctness of data for its application.

Deep learning

A machine learning implementation technique that uses large quantities of data, or feedback from interactions with a simulation or an environment, as training sets for a network with multiple hidden layers, called a deep neural network, often employing an iterative optimization technique called gradient descent, to tune large numbers of parameters that describe weights given to connections among units.

Deepfake

AI-generated or manipulated image, audio or video content that resembles existing persons, objects, places or other entities or events and would falsely appear to a person to be authentic or truthful.

Deterministic (algorithm / model)

An algorithm/model that, given the same inputs, always produces the same outputs.

Diffusion models

A type of generative AI model that produces output to match a prompt by iteratively refining noise. These types of models require substantial computational resources and processing time.

Documentation

The collection of records that describe an AI system's purpose and intended use, key design choices, training and operational data characteristics and provenance, testing and evaluation results, limitations, and version history, maintained to support transparency, oversight, and accountability across the AI lifecycle.

Explainability

Property of an AI system that enables a given human audience to comprehend the reasons for the system's behavior; the ability to understand an AI system's output and decision given certain inputs.

Federated learningy

A method of training AI models across multiple devices or organizations without sharing underlying data. This machine learning architecture helps preserve privacy while enabling collaborative machine learning.

Foundation models

Large machine learning models trained on vast amounts of raw and unlabeled data through unsupervised learning that can be adapted and applied to versatile downstream tasks. Large language models are common subsets of foundation models and underpin many generative AI applications in the financial sector.

General purpose AI

AI designed for use across a broad array of tasks across many different applications rather than for a specific domain.

Generative Adversarial Networks (GANs)

A machine learning framework in which two neural networks contest with each other in the form of a zero-sum game, where one agent's gain is another agent's loss. A GAN learns to generate new data with the same statistics as the training set.

Generative AI

The class of AI that emulate the structure and characteristics of input data in order to generate derived synthetic content. This can include images, videos, audio, text, and other digital content.

Guardrails

Layered safeguards to prevent access to bad information and behavior in an AI system. These may encompass policies, technical controls, and monitoring mechanisms, and may exist at the data, model, application, and infrastructure levels. These safeguards aim to ensure generative AI systems behave ethically, safely, and within organizational or regulatory boundaries by filtering training data, aligning model behavior, and enforcing post-deployment controls.

Hallucination

A phenomenon when AI produces output that is erroneous or flawed but is still in the form of a convincing narrative or presentation. Generative AI can still produce flawed information even if underlying data is free of defects.

Human biases

These biases reflect systematic errors in human thought based on a limited number of heuristic principles and predicting values to simpler judgmental operations…These biases are omnipresent in the institutional, group, and individual decision-making processes across the AI lifecycle, and in the use of AI applications once deployed.

Human-in-the-Loop (HITL)

A risk-control approach for AI where a human is integrated within the AI's decision-making process.

Interpretability

Transparency into the inner workings of AI output in the context of their designed functional purposes, which helps users gain deeper insights into the functionality and trustworthiness of the system and its outputs.

Large language model

A subset of machine learning that uses algorithms trained on large amounts of data through self-supervised machine learning to recognize patterns and respond to user requests in natural language.

Machine learning

An AI learning method that enables computational systems to learn patterns, make predictions, and optimize decisions from large amounts of data without being explicitly programmed for each task. Machine learning encompasses supervised, unsupervised, and reinforcement learning paradigms, serving as the technical foundation for data-driven intelligence and automation.

Model integrity

The process of protecting a model against improper information modification or destruction and ensuring information non-repudiation and authenticity.

Model risk

The potential for adverse consequences from decisions based on incorrect or misused model outputs and reports. Model risk can be from individual models and be in the aggregate. Aggregate model risk is affected by interaction and dependencies among models; reliance on common assumptions, data, or methodologies; and any other factors that could adversely affect several models and their outputs.

Multi-modal model

A model that processes and relates information from multiple data modalities, such as text, images, audio, and sensor data, among others.

Natural language processing

The ability of a machine to process, analyze, and mimic human language, either spoken or written.

Output validation

Systematic process of verifying and confirming that AI system outputs meet specified requirements, accuracy standards, and quality criteria before being used for downstream processes.

Override

Output or input that is ignored, altered, rejected, or reversed.

Performance monitoring

Ongoing activities that confirm an AI system is implemented appropriately, used as intended, and continues to perform as intended over time.

Performance threshold

A particular value or range of values of a performance measure or diagnostic that determines the acceptance or rejection of a model's performance.

Predictive analytics

A discipline within AI that leverages historical data, statistical algorithms, and machine learning techniques to identity patterns and forecast future outcomes, behaviors, or events. This discipline is distinguished by emphasis on forward-looking insights rather than descriptive analysis.

Prompt

Natural language text describing the task that an AI should perform.

Prompt injection

An attack on an AI system that exploits how an application combines untrusted input with a prompt written by a higher-trust party, such as the application designer, so the system follows the untrusted instructions.

Reinforcement learning

A type of machine learning in which a model learns to optimize its behavior according to a reward function by interacting with and receiving feedback from an environment.

Representation learning

Also known as feature learning, a set of techniques for automatically detecting feature patterns, replaces manual feature engineering.

Responsible AI

Conscientious design, deployment, and governance of AI systems aligned with ethical principles, societal values, and legal requirements.

Retrieval augmented generation (RAG)

A type of generative AI system in which a model is paired with a separate information retrieval system (or "knowledge base"). Based on a user query, the RAG system identifies relevant information within the knowledge base and provides it to the generative AI model in context for the model to use in formu­lating its response. RAG systems allow the internal knowledge of a generative AI model to be modified without the need for retraining.

Service level agreement (SLA)

Contractually binding terms, often incorporated into a broader services contract, between a service provider and a customer that specify the services to be delivered and the measurable performance and service-quality commitments, such as availability and response times. SLAs also typically define each party's responsibilities and provisions for monitoring/reporting, issue resolution, and remedies if service levels are not met.

Service Provider Concentration (Financial Institution)

The extent to which a financial institution relies on a service provider, directly or indirectly, to support the financial institution's activities, particularly critical activities.

Service Provider Concentration (Financial sector)

The extent to which financial institutions rely on a service provider, directly or indirectly, to support financial institutions' activities, particularly critical activities.

Service Provider Concentration Risk (Financial Institution)

The potential for disruption or degradation at a service provider(s) to threaten the ability of a financial institution to continue performing the financial institution's activities, particularly critical activities, or cause the financial institution to suffer significant adverse effects.

Service Provider Concentration Risk (Financial Sector)

The potential for disruption or degradation at a service provider(s) to threaten the ability of financial institutions to continue performing their activities, particularly critical activities, or cause the financial institutions to suffer significant adverse effects, with the potential for systemic impact to the financial sector.

Supervised learning

A process for training algorithms by example. The training data consists of inputs paired with the correct outputs. During training, the algorithm will search for patterns in the data that correlate with the desired outputs and learn to predict the correct output for newly presented input data over iterative training and model updates.

Structured data

Data that is divided into standardized pieces that are identifiable and accessible by both humans and computers.

Swarm

Swarm shows up in a few distinct AI contexts:

(A) Swarm intelligence is the oldest use -- it's a subfield of AI inspired by collective behavior in nature (ant colonies, bird flocking, bee swarms). Algorithms like Particle Swarm Optimization (PSO) and Ant Colony Optimization fall here. It's been around since the 1990s.

(B) Multi-agent swarms is the more current, buzzy usage. With the rise of agentic AI, "swarm" now commonly refers to networks of AI agents working in parallel or in coordination to complete complex tasks. OpenAI even released an experimental framework called Swarm in late 2024 specifically for orchestrating multi-agent workflows.

Synthetic data

Data that has been generated using a purpose built mathematical model or algorithm, that is statistically realistic but artificial, that can be used for activities like model development and training.

Synthetic identity

The use of a combination of real and fake personally identifiable information (PII) to fabricate a person or entity.

Text/word embedding

A numerical vector representation of text that machine learning and artificial intelligence systems use to work with meaning in text, such as comparing similarity between pieces of text.

Third-party AI risk

Risk that arises when an organization relies on another entity to develop, provide, host, operate, or support AI systems or key AI components such as models, data, and related infrastructure.

Traditional AI

Traditional AI, also referred to as symbolic or rule-based AI, is a subset of AI that focuses on performing discreet, preset tasks using predetermined algorithms and rules. These AI applications are designed to excel in a single activity or a restricted set of tasks, such as playing chess, diagnosing diseases, or translating languages.

Training data

A subset of input data samples used to train a machine learning model.

Unstructured data

Data that does not have a predefined data model or is not organized in a predefined way. This may also include data that is more free form, such as multimedia files, images, sound files, or unstructured text. Unstructured data does not necessarily follow any format or hierarchical sequence, nor does it follow any relational rules.

Unsupervised learning

A learning strategy that consists in observing and analyzing different entities and determining that some of their subsets can be grouped into certain classes, without any correctness test being performed on acquired knowledge through feedback from external knowledge sources.

Validation

Confirmation, through objective evidence, that an AI system or model meets requirements for a specific intended use or application and achieves its intended use in its intended operational environment.

Version control

Systematic practice of tracking, managing, and documenting changes to AI assets through their development and deployment lifecycle.

Cybersecurity Consultation

Do you know if your company is secure against cyber threats? Do you have the right security policies, tools, and practices in place to protect your data, reputation, and productivity? If you're not sure, it's time for a cybersecurity risk assessment (CSRA). STACK Cybersecurity's CSRA will meticulously identify and evaluate vulnerabilities and risks within your IT environment. We'll assess your network, systems, applications, and devices, and provide you a detailed report and action plan to improve your security posture. Don't wait until it's too late.

Schedule a Consultation Explore our Risk Assessment