AI Glossary - Yenra

Curated glossary of core artificial intelligence terms

Yenra AI Glossary
Yenra AI Glossary

This curated glossary focuses on the concepts most likely to help readers understand modern AI, machine learning, and generative systems without the long tail of highly specialized hardware, statistical, and research-only jargon.

A

Activation Function: A mathematical function that adds nonlinearity to a neural network so it can learn more than simple linear patterns.

Active Learning: A training strategy in which the model asks for labels on the most informative examples instead of labeling everything at once.

Account Takeover: When an attacker gains control of a legitimate user's account and begins acting as that user.

Adversarial Attack: A deliberate attempt to fool a model with input designed to make it fail.

Adversarial Example: A specially crafted input that looks normal to people but causes a model to make a mistake.

Adversarial Machine Learning: The field that studies how AI systems are attacked, manipulated, and defended.

AI (Artificial Intelligence): The broad field of building systems that can perform tasks associated with perception, language, reasoning, and decision-making.

AI Agent: A software system that can interpret goals, use tools, and take actions with some autonomy.

After-Call Work (ACW): The wrap-up work an agent completes after an interaction, such as notes, summaries, disposition codes, and CRM updates.

Agent Assist: Real-time AI support that helps a human agent with knowledge, prompts, summaries, and next-best actions during a live interaction.

Advanced Process Control (APC): A model-based control layer that adjusts process settings using measurements, predictions, and feedback.

Advanced Driver Assistance Systems (ADAS): Vehicle technologies that help a human driver with warnings, braking, steering, and workload reduction without making every car fully autonomous.

Advanced Metering Infrastructure (AMI): The metering, communications, and data systems that make interval electricity data and smarter grid interaction possible.

Affective Computing: AI systems that estimate, model, or respond to human affect and emotion from signals such as text, voice, facial expression, or behavior.

Augmentative and Alternative Communication (AAC): Communication methods and tools that support or replace speech when a person cannot rely on spoken language alone.

AI Content Moderation: Using AI to review, filter, rank, or escalate content that may violate rules or safety standards.

AI Data Labeling: The process of tagging data so supervised models can learn from it.

AI Fairness: The effort to make AI systems behave equitably across people, groups, and contexts.

AI Firewall: A security layer that inspects AI inputs, actions, and outputs for threats, misuse, or policy violations.

AI-Generated Content (AIGC): Text, images, audio, video, or code created by AI systems.

Algorithm: A set of instructions or rules used to solve a problem or perform a computation.

Algorithmic Trading: Using software rules and models to generate, route, or manage orders in financial markets.

Algorithmic Bias: Systematic skew in a system that leads to unfair or distorted outcomes.

Agent-Based Modeling: Simulating a system through many interacting agents whose behavior and feedback loops shape the larger outcome.

Alignment (AI Alignment): The effort to make AI systems follow human goals, instructions, and safety expectations.

Ambient Computing: Computing woven into devices and environments so assistance can appear in context instead of always requiring an explicit app session.

Anomaly Detection: Finding unusual data points or events that differ sharply from normal patterns.

Architecture (AI Model Architecture): The overall design and structure of a model, including its layers and connections.

Archives: Organized collections of records and materials preserved because they have long-term historical, legal, cultural, or operational value.

Artificial General Intelligence (AGI): A hypothetical AI with broad human-like ability across many tasks rather than strength in one narrow domain.

Attention Mechanism: A way for a model to focus on the most relevant parts of its input when producing an output.

Attribution: Assigning authorship, origin, source, or responsibility to a work, record, object, or other output.

Audience Segmentation: Grouping people into useful audience buckets or modeled cohorts for targeting, personalization, or measurement.

Authentication: Confirming that a person, document, object, or piece of content is genuine and really what it claims to be.

Autoencoder: A neural network trained to compress data into a compact representation and reconstruct it again.

Automatic Speech Recognition (ASR): Technology that converts spoken language into text.

Automated Machine Learning (AutoML): Using software to automate parts of model training, tuning, and evaluation.

B

Backpropagation: The training process that moves error information backward through a neural network so its weights can be updated.

Benchmarks: Standard tests used to compare models on common tasks or datasets.

BACnet: A widely used communications standard that helps building automation devices exchange data and commands across vendors.

Battery Management System (BMS): The control system that monitors an EV battery pack and manages charging, power limits, safety, and battery-health estimation.

Behavioral Biometrics: Authentication and fraud-detection methods that identify people by how they type, swipe, move, or otherwise behave.

BERT (Bidirectional Encoder Representations from Transformers): An influential transformer-based language model designed for understanding text with bidirectional context.

Bias: Systematic skew or error in data, modeling, or decisions that can distort results or create unfair outcomes.

Bias Mitigation: Methods for identifying, reducing, and monitoring unfair bias in AI systems.

Bias-Variance Tradeoff: The balance between a model that is too simple to capture patterns and one that is too sensitive to the training data.

Binary Classification: A task in which each example must be assigned to one of two classes.

Black Box Model: A model whose internal reasoning is hard for humans to inspect or explain directly.

Brand Lift: Measuring whether advertising changed awareness, recall, favorability, consideration, or related brand outcomes compared with a control group.

Brand Safety: Keeping ads and generated creative away from harmful, unsuitable, misleading, or policy-violating content and contexts.

C

Calibration: The degree to which a model's confidence matches what really happens.

Call Deflection: Resolving routine issues through self-service or alternate channels so avoidable live-agent calls never have to enter the voice queue.

Candidate Generation: The retrieval stage that narrows a huge pool of possible items into a smaller set worth ranking in detail.

Cataloging: The structured process of describing an item so it can be identified, found, and managed later.

Chain of Thought (CoT): A prompting style that encourages a model to work through intermediate reasoning steps.

Change Detection: Comparing observations across time to determine what changed, where it changed, and how much it changed.

Chatbot: A system that interacts with users in natural language through text or voice.

Classification: The task of assigning an input to one of several categories.

Clinical Decision Support: Software that uses patient data, rules, models, or retrieved evidence to help clinicians make safer and better-informed decisions.

Collaborative Robot (Cobot): A robot designed to work more safely near people and to make more mixed, flexible automation tasks practical.

Clustering: Grouping similar items together without using predefined labels.

Cold Start: The recommendation problem that appears when a new user, item, or context has too little history for confident prediction.

Collections Management: The ongoing work of organizing, tracking, preserving, and governing collections over time.

Combined Heat and Power (CHP): Generating electricity and useful heat from the same fuel source so less energy is wasted overall.

Conservation: The active care and protection of objects, records, and heritage materials to slow deterioration and maintain their integrity.

Computer Vision: The branch of AI that helps systems interpret images and video.

Confidence: The system's stated or implied degree of certainty about an output, prediction, or match.

Continuous Authentication: Reassessing identity during a session instead of trusting one successful login forever.

Confusion Matrix: A table that shows how often a classifier makes each kind of right and wrong prediction.

Convolutional Neural Network (CNN): A neural network architecture especially useful for images and other grid-like data.

Context Window: The amount of prior input a model can consider in a single interaction.

Contextual Targeting: Showing ads or recommendations based on the surrounding content or moment rather than mainly on long-term user identity.

Conversation Intelligence: Using AI to turn calls, meetings, and other conversations into searchable structure, topics, sentiment, and workflow signals.

Conversational Commerce: Shopping flows that use natural language, recommendations, and live product data to guide product discovery and buying.

Creative Fatigue: The decline in ad performance that happens when audiences see the same or too-similar creative too often.

Cross-Validation: A method of testing a model by training and validating it across multiple different splits of the data.

D

Data Augmentation: Expanding a dataset by creating modified versions of existing examples.

Data Clean Room: A controlled environment where multiple parties can analyze combined data with privacy rules and aggregation safeguards.

Data Drift: A change in the input data over time that can hurt model performance.

Data Governance: The policies and controls that determine how data is collected, managed, protected, and used.

Demand Response: Reducing or shifting electricity use in response to grid conditions, prices, or utility signals.

De-Identification: Removing or transforming identifying details so data can be analyzed or shared with lower privacy risk.

Device Fingerprinting: Using device, browser, network, and environment clues to estimate whether an access event fits a known pattern.

Data Labeling: Adding tags or annotations to data so a model can learn the target output.

Data Preprocessing: Cleaning and transforming raw data before training or inference.

Deep Learning: A branch of machine learning based on multi-layer neural networks.

Deepfake: AI-generated or AI-altered media designed to convincingly imitate a real person's appearance or voice.

Decoder: The part of a model that generates output from an internal representation or from encoded input.

Differential Privacy: A formal way to reduce how much any one person's data can be inferred from a result or model.

Diffusion Models: Generative models that start from noise and gradually turn it into structured output.

Direct Indexing: Owning the underlying securities of an index so a portfolio can be customized, rebalanced, and tax-managed more precisely.

Digital Thread: A connected flow of lifecycle data that links design, production, operation, and service information.

Digital Twin: A live digital representation of a physical asset or process that stays connected to operational data.

Digitization: Converting physical or analog material into digital form so it can be stored, searched, and reused more effectively.

Driver Monitoring System: A vehicle system that checks whether the human driver is attentive, alert, and ready to supervise or take over.

Document AI: AI systems that read, classify, extract, and route information from documents; often called intelligent document processing (IDP).

Dynamic Creative Optimization (DCO): Automatically combining, testing, and promoting ad assets so different viewers and placements get more effective creative variants.

E

Embedding: A numeric vector representation that captures semantic similarity between items such as words, images, or documents.

Edge Computing: Processing data and running software close to where it is generated instead of sending every decision to a faraway cloud.

Electronic Health Record (EHR): A digital longitudinal record of patient care data such as notes, medications, orders, labs, and encounter history.

Encoder: The part of a model that transforms input into a useful internal representation.

Entity Extraction and Linking: Identifying important entities in text and connecting each mention to the correct real-world or database entry.

Epoch: One full pass through the training dataset during model training.

Ethical AI: Building and using AI in ways that respect fairness, accountability, privacy, and human values.

Evidence: The records, measurements, retrieved sources, or other signals that support a claim, decision, or model output.

Explainability: The broader practice of making a system's outputs, reasoning, evidence, and limits understandable to people.

Explainable AI (XAI): The field of making model behavior and outputs easier for people to understand.

F

F1 Score: A metric that combines precision and recall into one balanced measure.

Factor Investing: Building or analyzing portfolios through systematic exposures such as value, momentum, quality, size, or low volatility.

Face Identification: Searching one face against many enrolled identities to find likely candidates or a match above threshold.

Face Verification: Checking whether one presented face matches one claimed or enrolled identity.

Facings: The number of visible units of a product shown to the shopper on the shelf.

Feed Ranking: The process of ordering eligible posts, videos, or updates so a personalized feed shows the most relevant items first.

Feature Engineering: Creating or refining input variables so a model can learn more effectively.

Federated Learning: Training a shared model across many devices or organizations without pooling all raw data in one place.

Fault Detection and Diagnostics (FDD): Using rules, models, and live operational data to find faults in equipment or controls and explain what is likely going wrong.

FHIR: Fast Healthcare Interoperability Resources, a standard for structuring and exchanging healthcare data through modern APIs.

Fine-Tuning: Adapting a pretrained model to a narrower task or domain with additional training.

Flue Gas Cleaning: The systems that remove particulates, acid gases, metals, and other pollutants from combustion exhaust before release.

Forgery: A deceptive imitation or alteration meant to pass as a genuine object, document, identity, or piece of media.

Fraud Detection: Using analytics and AI to identify suspicious behavior, transactions, or impersonation.

Function Calling: A way for a model to produce structured arguments for a tool or function instead of free text alone.

G

Generative Adversarial Network (GAN): A generative architecture in which one network creates outputs and another tries to detect whether they are fake.

Generative Artificial Intelligence (GenAI): AI systems that create new content such as text, images, audio, video, or code.

Geographic Information System (GIS): Software and data systems for storing, visualizing, analyzing, and managing spatial information.

Gesture Recognition: Using AI and sensors to interpret hand, body, or motion signals as interface commands.

GPT (Generative Pre-trained Transformer): A family of transformer-based language models trained to predict the next token in text.

Graph Neural Network (GNN): A neural network built to learn from graph-structured data such as molecules, transaction networks, and knowledge graphs.

Grounding: Connecting a model's output to trusted sources, retrieved evidence, tools, or real-world state.

Ground Truth: The verified real-world state or trusted label set used as the reference point for training, evaluation, or operational measurement.

Guardrails: Filters, rules, and runtime checks that keep an AI system within desired safety or workflow boundaries.

H

Hallucination: When an AI system produces plausible-sounding information that is false, unsupported, or ungrounded.

Handwriting Recognition: Using AI to read handwritten notes, forms, and manuscripts and convert them into machine-readable text.

Hyperspectral Imaging: Capturing many narrow wavelength bands so materials can be identified by their spectral signatures rather than by ordinary color alone.

Human in the Loop: A workflow in which people remain involved in review, correction, approval, or decision-making instead of leaving everything to the model.

Hyperparameter: A setting chosen by the developer, such as learning rate or batch size, rather than learned by the model itself.

Hyperparameter Tuning: The process of searching for better hyperparameter settings to improve model performance.

I

Identity Proofing: Establishing that a person is who they claim to be during enrollment before later authentication begins.

Image Classification: The task of assigning an image to one or more categories based on what it contains.

Image Generation: Creating new images with AI, often from text prompts or other visual input.

Inference: Using a trained model to produce output on new input data.

Inventory Visibility: Knowing what inventory exists, where it is, and how available it really is across stores, warehouses, and channels.

Incrementality: Measuring what effect a campaign or intervention caused beyond what would likely have happened anyway.

Intent Recognition: Identifying what a user is trying to accomplish so a conversational system can route the interaction toward the right next step.

Instruction Tuning: Post-training that teaches a model to follow user requests more reliably.

Interoperability: The ability of different systems to exchange data and use it consistently without losing meaning.

Interpretability: The degree to which humans can understand how and why a model behaves the way it does.

J

Jailbreaking: Manipulating a language model so it bypasses intended restrictions or safety rules.

K

Knowledge Distillation: Training a smaller model to imitate a larger one so it becomes cheaper to deploy.

Knowledge Graph: A structured network of entities and relationships that makes connected facts easier to query and reason over.

L

Large Language Model (LLM): A neural network trained on large amounts of text to generate and work with language.

Language Model (LM): A model that learns patterns in language and predicts likely token sequences.

Layout Analysis: Identifying the structural parts of a page so AI can preserve reading order, tables, fields, and other document context.

Latent Space: A compressed internal representation in which models capture the underlying structure of data.

Learning Rate: A training setting that controls how large each update step is during optimization.

Link Prediction: Estimating which relationships are likely missing from a graph based on the structure already present.

Liveness Detection: Anti-spoofing checks that try to confirm a real person is present instead of a photo, replay, mask, or deepfake.

LoRA (Low-Rank Adaptation): A parameter-efficient way to adapt a large model without retraining all of its weights.

Loss Function: A formula that measures how wrong a model's predictions are during training.

Live-Agent Handoff: Transferring a conversation from a bot to a human agent with enough context that the customer does not have to start over.

M

ML (Machine Learning): The branch of AI in which systems learn patterns from data instead of relying only on fixed hand-written rules.

Machine Translation: Using AI to translate meaning from one language into another.

Market Microstructure: How order books, spreads, queues, and venue rules shape the way real trades are executed.

Marketing Mix Modeling (MMM): Using aggregated data and statistical models to estimate how channels and external factors affect marketing outcomes over time.

Matter: A smart-home interoperability standard that helps accessories from different brands work together more cleanly inside the same home.

Material Recovery Facility (MRF): A recycling facility that receives mixed material, separates it into usable commodities, and prepares those streams for sale or further processing.

Metadata Enrichment: Adding useful tags, descriptions, relationships, and context to content so it is easier to search, organize, connect, and reuse.

Microgrid: A local energy system that can coordinate generation, storage, and loads together and sometimes operate independently from the wider grid.

Model Card: A document that explains a model's purpose, evaluation, limitations, and intended use.

Model Compression: Techniques that make models smaller, faster, or cheaper to run.

Model Predictive Control (MPC): A control approach that uses a model of the system to predict future behavior and choose actions that work best under constraints.

Model Drift: A decline or change in model behavior over time as conditions shift.

Model Evaluation: Testing a model to understand how well it performs and where it fails.

Model Explainability: The ability to communicate why a model produced a particular result.

Model Fairness: Whether a specific model behaves equitably across different groups or contexts.

Model Monitoring: Tracking a deployed model so drift, degradation, and unusual behavior are caught after launch.

Model Parameters: The values a model learns during training, such as neural network weights.

Multimodal Large Language Models (MLLMs): Models that combine language with images, audio, video, or other input types.

Multimodal Learning: Learning from or generating across more than one kind of data, such as text plus images.

N

Natural Language Processing (NLP): The branch of AI focused on understanding and generating human language.

Named Entity Recognition (NER): Identifying references to people, organizations, places, dates, and other key entities in text.

Neural Networks: Computing systems inspired by layered neuron-like structures that learn patterns from data.

Nowcasting: Estimating the current state of the economy before the official statistics are fully available.

O

Object Detection: Identifying both what objects appear in an image and where they are located.

On-Device AI: Running AI features locally on a phone, laptop, vehicle, or other device instead of sending every request to a remote server.

Onboard Autonomy: Letting a spacecraft, rover, or other remote system make limited decisions locally when waiting for human commands is too slow.

Open Banking: User-permissioned access to bank and financial account data so apps and services can help with budgeting, payments, and account aggregation.

Optical Character Recognition (OCR): Converting text in scans, images, or PDFs into machine-readable text.

Optical Sorting: Using cameras and other sensors to identify materials on a moving line and separate them automatically.

Operational Design Domain (ODD): The specific conditions under which an autonomous or assisted system is designed and validated to operate.

Ontology: A formal description of the important concepts, categories, and relationships in a domain.

Overfitting: When a model learns the training data too specifically and performs poorly on new examples.

P

Personally Identifiable Information (PII): Information that can identify a specific person and therefore requires careful handling and protection.

Phenotyping: Identifying meaningful patient traits, disease patterns, or clinical subgroups from health data.

Physical AI: AI systems that act in the physical world through robots, vehicles, sensors, and control systems rather than only through software interfaces.

Path Planning: Choosing a route or movement sequence that helps a robot reach its goal safely and efficiently.

Passkey: A phishing-resistant sign-in credential that often uses device biometrics or a PIN for local user verification.

Planogram: A structured retail shelf layout that specifies where products should go, how much space they get, and how the shelf should be merchandised.

Planogram Compliance: Checking whether products, facings, prices, and placement match the intended retail merchandising plan.

Precision: The share of positive predictions that are actually correct.

Predictive Analytics: Using data and models to estimate likely future outcomes, risks, or trends.

Predictive Maintenance: Using data and models to estimate when equipment is likely to degrade or fail so maintenance can happen before an outage.

Post-Quantum Cryptography: Cryptographic methods designed to remain secure even if large-scale quantum computers become practical.

Preservation: The long-term work of keeping information, media, artifacts, or records safe, usable, and accessible over time.

Pre-trained Model: A model that has already been trained on a large dataset before being adapted to a narrower task.

Pretraining: The large-scale initial training stage that teaches a model broad patterns before task-specific adaptation.

Presence-Based Automation: Automations that respond to whether people are home, away, arriving, or moving through a space instead of relying only on fixed schedules.

Pronunciation Assessment: Using structured scoring to judge how closely a spoken sound, word, or phrase matches a target production.

Prosody: The rhythm, pitch, stress, pacing, and intonation patterns in speech that carry meaning beyond the transcript alone.

Prompt: The text, instructions, or input given to a model to guide its output.

Prompt Engineering: Designing prompts so a model produces more useful, accurate, or structured responses.

Prompt Injection: A security attack in which malicious instructions hidden in content try to override a model's intended behavior.

Provenance: The documented origin and ownership or custody history of an object, record, or other item of value.

Q

Quantization: Reducing numeric precision in a model to save memory and speed up inference.

R

Random Forest: A machine learning method that combines many decision trees to improve prediction stability.

Recall: The share of real positive cases that the model successfully catches.

Recommender System: An AI system that ranks and suggests items a particular user is likely to value.

Red Teaming: Structured adversarial testing used to uncover safety, security, bias, and reliability failures.

Regression: A task in which a model predicts a numeric value rather than a category.

Retail Media: Advertising inside retailer-owned shopping environments and data ecosystems using commerce signals close to purchase.

RFID (Radio-Frequency Identification): Using radio tags and readers to identify physical items automatically as they move through stores, stockrooms, and fitting rooms.

Retro-Commissioning: A systematic process of testing and tuning an existing building so its systems perform closer to intended design and operational goals.

Responsive Search Ads (RSA): Search ads that mix and match multiple headlines and descriptions to find stronger combinations for different queries and contexts.

Robo-Adviser: An automated digital investment service that uses questionnaires, algorithms, and portfolio rules to guide or manage investing.

Reinforcement Learning (RL): A learning approach in which an agent improves through rewards and penalties from its environment.

Reinforcement Learning from Human Feedback (RLHF): A method that uses human preferences to shape model behavior after pretraining.

Regularization: Training techniques that reduce overfitting and help a model generalize better.

Replenishment: Restocking inventory so products are available where and when they are needed without carrying too much excess stock.

Remote Sensing: Collecting imagery or other measurements from a distance so AI systems can analyze the Earth, oceans, atmosphere, or planetary surfaces.

Restoration: Repairing or reconstructing damaged, degraded, faded, or incomplete material so it becomes more legible, usable, or understandable.

Risk-Based Authentication: Adjusting authentication requirements based on how risky the current sign-in or action appears to be.

Responsible AI: Building AI systems that are useful, safe, fair, accountable, and governable.

Retrieval Augmented Generation (RAG): A pattern that combines a model with retrieved external information so answers stay fresher and better grounded.

Robustness: The ability of a system to keep working under noise, shift, unusual input, or active attack.

S

Self-Supervised Learning: Learning from data that provides its own training signal instead of relying only on manual labels.

Semantic Search: Search that finds results by meaning and intent, not only exact keyword matches.

Sensor Fusion: Combining signals from multiple sensors into one more reliable estimate of what is happening.

Send-Time Optimization: Using data and models to decide when a message is most likely to be opened, read, or acted on.

Sentiment Analysis: Using AI to identify whether language expresses positive, negative, neutral, or mixed attitude.

SLAM (Simultaneous Localization and Mapping): A robotics method for figuring out where the robot is while building a map of the space around it.

Smart Charging: Using software to decide when, where, and how fast an electric vehicle should charge.

Smart Grid: An electricity system that uses sensing, communication, software, and automation to manage supply, demand, and reliability more intelligently.

Slippage: The gap between the price a trader expected and the price the market actually delivered.

SOAR: Security orchestration, automation, and response systems that connect tools and run playbooks faster.

Social Listening: Using AI to monitor, organize, and interpret social posts, comments, reviews, and other public audience signals at scale.

Speaker Diarization: Figuring out who spoke when in a recording so transcripts preserve conversational structure instead of collapsing everyone together.

Speech Biofeedback: Using visual, acoustic, or sensor-based feedback to help a speaker see or hear aspects of speech production that are otherwise hard to monitor directly.

Shelf Intelligence: Using AI and computer vision to understand stock levels, placement, pricing, and other real shelf conditions in stores.

Stable Diffusion: A widely known family of diffusion-based image generation models.

Stress Testing: Testing how a system, market, or portfolio behaves under difficult but plausible conditions.

Structural Break: A change in the underlying relationships that a model relied on before.

Supervised Learning: Learning from labeled examples where the correct target is already known.

Support Vector Machine (SVM): A classical machine learning method often used for classification and sometimes regression.

Surrogate Model: A simplified fast-running model that approximates a more complex simulation or physical process.

Synthetic Data: Artificially generated data used to train, test, or evaluate models.

System Prompt: A high-priority instruction that sets a model's role, behavior, or constraints for an interaction.

T

Tax-Loss Harvesting: Selling positions at a loss to offset taxable gains while keeping the portfolio aligned to its long-term strategy.

Test Set: A held-out dataset used to measure how well a model generalizes after training.

TEFCA: The Trusted Exchange Framework and Common Agreement, a U.S. framework for nationwide health information exchange.

Telemetry: Operational signals such as events, metrics, logs, traces, and state changes that show what a device or system is doing over time.

Text Summarization: Condensing a longer document or conversation into a shorter version that preserves the main ideas.

Teleoperation: Remote control or supervision of a robot by a human operator, often blended with autonomy.

Time Series Forecasting: Predicting future values from time-ordered data such as demand, occupancy, emissions, or sensor readings.

Token: A unit of text a language model processes, such as a word piece, symbol, or short sequence of characters.

Tokenization: Breaking text into the tokens a model can process.

Tool Use: Letting a model call external tools, APIs, or services as part of solving a task.

Toxicity: Harmful, abusive, or hostile content that AI systems may generate, amplify, or help detect.

Training Set: The portion of data used to fit a model's parameters during learning.

Transfer Learning: Reusing knowledge from one task or dataset to improve performance on another.

Transformer: A neural network architecture built around attention that powers many modern language and multimodal models.

U

Uncertainty: The degree to which a model, dataset, or decision remains ambiguous, incomplete, or not fully known.

Underfitting: When a model is too simple to capture meaningful structure in the data.

Unsupervised Learning: Learning patterns from unlabeled data without predefined target labels.

V

Validation Set: A dataset used during model development to compare choices and tune settings before final testing.

Vehicle-to-Grid (V2G): Using bidirectional EV charging so vehicles can supply power back to a building or the electric grid when it makes sense.

Vector Database: A database optimized for storing and searching embeddings.

Vector Search: Finding items by semantic similarity in embedding space rather than exact keyword matching.

Verification: Checking whether a claim, identity, document, output, or piece of media is correct, genuine, or supported by evidence.

Virtual Commissioning: Testing automation and control logic in simulation before real equipment goes live.

Virtual Power Plant (VPP): A coordinated network of distributed energy resources that can be managed as a flexible power system even though the assets are spread across many locations.

Virtual Metrology: Estimating measurement results from process and equipment data instead of physically measuring every unit.

Virtual Try-On: Using AI and visual overlays to preview how makeup, accessories, or apparel may look before buying.

Visual Search: Searching with an image or camera input so the system finds visually similar objects, scenes, or products.

Voice Biometrics: Using voice characteristics as an identity signal for personalization, verification, or low-friction access control.

W

Wake-Word Detection: The lightweight speech model that listens for an activation phrase so a voice device knows when to start paying closer attention.

Waste-to-Energy: Converting residual waste into usable energy through controlled thermal, biological, or other recovery processes.

Workflow Orchestration: Coordinating the sequence of models, rules, tools, approvals, and human review steps around an AI-driven process.

Z

Zero-Shot Learning: The ability to handle a task or class the model was not explicitly trained on.

Zero Trust: A security model that assumes no user, device, or network location should be trusted by default.