The Essential Glossary of AI: 50+ Key Terms to Know
September 13, 2024
AI Glossary: 50+ Essential Terms to Understand Artificial Intelligence
A
AI (Artificial Intelligence): The simulation of human intelligence processes by machines, including learning, reasoning, and self-correction.
AI Accelerator: Specialized hardware or systems designed to speed up AI tasks, with NVIDIA GPUs being a leader in this space.
Algorithm: A set of rules or instructions given to an AI system to help it learn, solve problems, and make decisions.
Ampere Architecture: NVIDIA's GPU architecture designed to accelerate AI, data analytics, and HPC workloads.
B
Bare Metal: A physical server dedicated to a single tenant, offering full control over its resources without virtualization.
Benchmarking: The process of measuring system or model performance against a standard to evaluate efficiency.
Big Data: Large, complex datasets requiring advanced tools for processing, often used to uncover patterns.
Black Box: An AI system where the decision-making process is opaque, even if inputs and outputs are visible.
C
Cluster: A group of interconnected computers (nodes) working as a unified system for complex computations.
Colocation: Housing privately-owned servers in a third-party data center.
Compute: The processing power available in cloud or data center environments for running applications.
Containerization: A lightweight virtualization method that packages applications and dependencies into a container.
Conversational AI: Technology that powers machines to engage in human-like conversations using natural language processing (NLP).
CUDA (Compute Unified Device Architecture): NVIDIA’s parallel computing platform enabling general-purpose processing on GPUs.
CUDA Core: The fundamental processing unit in an NVIDIA GPU, handling parallel computing operations.
D
Deep Learning: A subset of machine learning that uses neural networks with multiple layers to analyze complex data.
DGX System: NVIDIA’s AI supercomputers designed for deep learning and AI workloads.
F
Finetuning: The process of training a pre-trained AI model on a smaller, task-specific dataset to enhance performance.
FP16 (Half Precision Floating Point): A 16-bit floating point format that improves memory efficiency in AI training.
FP32 (Single Precision Floating Point): A 32-bit format used in AI training and gaming for representing real numbers.
G
GDDR Memory: A type of memory in GPUs that provides high bandwidth for data access, critical in gaming and AI.
Generative AI: AI that creates new content (text, images, video) based on patterns learned from existing data.
GPU (Graphics Processing Unit): A processor designed for parallel processing, widely used in AI and machine learning.
H
HPC (High-Performance Computing): The use of supercomputers for solving complex problems requiring large-scale processing.
Hypervisor: Software that runs virtual machines by separating the physical hardware from the operating system.
I
InfiniBand (IB): A high-speed networking standard used in HPC clusters for low-latency communication.
Internet of Things (IoT): A network of connected devices that exchange data, often enhanced by AI for predictive tasks.
IOPS (Input/Output Operations Per Second): A measurement of storage device performance.
K
Kubernetes: An open-source platform automating the deployment and management of containerized applications.
L
Latency: The time delay between data transmission and receipt, important for real-time applications.
LLM (Large Language Model): AI models trained on vast amounts of text data to generate human-like language.
Load Balancing: Distributing workloads across resources to optimize performance and avoid overloading any single system.
M
Machine Learning (ML): A subset of AI where machines use algorithms and statistical models to improve performance over time.
MIG (Multi-Instance GPU): A feature allowing an NVIDIA GPU to be partitioned into smaller, isolated instances.
N
Natural Language Processing (NLP): AI that allows computers to understand, interpret, and generate human language.
NIC (Network Interface Controller): A hardware component enabling a device to connect to a network.
Node: A single machine or device in a network or cluster that processes and exchanges data.
NVLink: NVIDIA’s high-speed interconnect for fast communication between GPUs and CPUs.
NVSwitch: NVIDIA’s high-bandwidth switch allowing multiple GPUs to work together within a server.
O
OOB (Out of Band): A management method allowing access to network devices via a separate channel, even if the main network is down.
Optimization: The process of improving system or model performance by adjusting parameters.
P
Pre-training: The initial training phase where AI models learn general patterns from large datasets.
R
RAID (Redundant Array of Independent Disks): A data storage technology combining multiple disks for redundancy and improved performance.
S
SLA (Service Level Agreement): A contract specifying the level of service expected from a provider.
SLM (Small Language Model): A smaller AI model used for tasks where large-scale models are unnecessary.
T
Tensor: A mathematical object used in machine learning and AI, particularly in tensor processing units (TPUs).
Tensor Core: Specialized cores in NVIDIA GPUs designed for matrix operations crucial to AI workloads.
TensorRT: NVIDIA’s deep learning inference library optimized for real-time AI applications.
Text-to-Image Generator: An AI tool that converts text descriptions into images.
Text-to-Speech: AI technology that converts written text into spoken words.
Throughput: The amount of data processed by a system in a given amount of time, often measured in bits per second (bps).
V
Virtualization: Creating virtual versions of resources like servers or storage to enable multiple OSs on one machine.
Volta Architecture: NVIDIA’s GPU architecture that introduced Tensor Cores, widely used in AI tasks.