You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 

366 lines
13 KiB

terms:
- term: "gb10"
full_name: "NVIDIA GB10 Superchip"
definition: |
System-on-chip combining an NVIDIA Grace CPU and Blackwell GPU
connected via NVLink-C2C. The core silicon in the Dell Pro Max GB10
and NVIDIA DGX Spark.
unit: null
typical_range: null
related_terms: ["grace-blackwell", "superchip", "nvlink-c2c"]
related_topics: ["gb10-superchip"]
- term: "grace-blackwell"
full_name: "Grace Blackwell Architecture"
definition: |
NVIDIA's combined CPU+GPU architecture pairing a Grace ARM CPU
with a Blackwell GPU via NVLink-C2C coherent interconnect.
unit: null
typical_range: null
related_terms: ["gb10", "blackwell-gpu", "grace-cpu"]
related_topics: ["gb10-superchip"]
- term: "superchip"
full_name: "Superchip"
definition: |
NVIDIA's term for a system-on-chip that integrates both CPU and GPU
dies on a single package with high-bandwidth interconnect.
unit: null
typical_range: null
related_terms: ["gb10", "soc"]
related_topics: ["gb10-superchip"]
- term: "soc"
full_name: "System-on-Chip"
definition: |
An integrated circuit that combines multiple components (CPU, GPU,
memory controller, I/O) on a single die or package.
unit: null
typical_range: null
related_terms: ["gb10", "superchip"]
related_topics: ["gb10-superchip"]
- term: "cortex-x925"
full_name: "ARM Cortex-X925"
definition: |
ARM's high-performance CPU core design (ARMv9.2 architecture).
The GB10 contains 10 of these as its "big" cores.
unit: null
typical_range: null
related_terms: ["cortex-a725", "gb10"]
related_topics: ["gb10-superchip"]
- term: "cortex-a725"
full_name: "ARM Cortex-A725"
definition: |
ARM's efficiency-focused CPU core design (ARMv9.2 architecture).
The GB10 contains 10 of these as its "LITTLE" cores.
unit: null
typical_range: null
related_terms: ["cortex-x925", "gb10"]
related_topics: ["gb10-superchip"]
- term: "blackwell-gpu"
full_name: "NVIDIA Blackwell GPU"
definition: |
NVIDIA's GPU architecture generation. In the GB10, it provides
6,144 CUDA cores and 5th-gen Tensor Cores.
unit: null
typical_range: null
related_terms: ["cuda-core", "tensor-core", "gb10"]
related_topics: ["gb10-superchip"]
- term: "cuda-core"
full_name: "CUDA Core"
definition: |
NVIDIA's basic parallel processing unit for general-purpose GPU
computing. The GB10 has 6,144 CUDA cores.
unit: "cores"
typical_range: "6,144 in GB10"
related_terms: ["blackwell-gpu", "tensor-core"]
related_topics: ["gb10-superchip"]
- term: "tensor-core"
full_name: "Tensor Core (5th Generation)"
definition: |
Specialized GPU cores for matrix multiply-accumulate operations,
critical for deep learning inference and training. 5th-gen Tensor
Cores in Blackwell support FP4, FP8, FP16, and other precisions.
unit: "cores"
typical_range: null
related_terms: ["blackwell-gpu", "fp4", "fp8"]
related_topics: ["gb10-superchip", "ai-workloads"]
- term: "nvlink-c2c"
full_name: "NVLink Chip-to-Chip"
definition: |
NVIDIA's proprietary die-to-die interconnect connecting the Grace CPU
and Blackwell GPU within the GB10 superchip. Provides 600 GB/s
bidirectional bandwidth and enables unified coherent memory.
unit: "GB/s"
typical_range: "600 GB/s bidirectional"
related_terms: ["gb10", "unified-memory"]
related_topics: ["gb10-superchip", "memory-and-storage"]
- term: "unified-memory"
full_name: "Unified Coherent Memory"
definition: |
Memory architecture where CPU and GPU share the same physical memory
pool with hardware cache coherence. Eliminates explicit host-device
memory copies. In the GB10, both processors see the full 128 GB.
unit: "GB"
typical_range: "128 GB in GB10"
related_terms: ["lpddr5x", "nvlink-c2c"]
related_topics: ["memory-and-storage", "gb10-superchip"]
- term: "lpddr5x"
full_name: "Low-Power DDR5X"
definition: |
Latest generation of low-power DRAM. In the GB10, runs at up to
9,400 MT/s providing 273 GB/s of memory bandwidth.
unit: "MT/s"
typical_range: "9,400 MT/s in GB10"
related_terms: ["unified-memory"]
related_topics: ["memory-and-storage"]
- term: "tflops"
full_name: "Tera Floating-Point Operations Per Second"
definition: |
Unit of compute performance. 1 TFLOPS = 10^12 floating-point
operations per second. ALWAYS specify the precision (FP4, FP8,
FP16, FP32) when quoting TFLOPS figures.
unit: "TFLOPS"
typical_range: "1,000 TFLOPS FP4 for GB10"
related_terms: ["pflop", "fp4"]
related_topics: ["gb10-superchip", "equations-and-bounds"]
- term: "pflop"
full_name: "Peta Floating-Point Operations Per Second"
definition: |
1 PFLOP = 1,000 TFLOPS = 10^15 floating-point operations per second.
The GB10's headline figure is 1 PFLOP at FP4 precision.
unit: "PFLOP"
typical_range: "1 PFLOP FP4 for GB10"
related_terms: ["tflops", "fp4"]
related_topics: ["gb10-superchip", "equations-and-bounds"]
- term: "fp4"
full_name: "4-bit Floating Point"
definition: |
Ultra-low precision numerical format using 4 bits per value.
Used for quantized inference. The GB10's 1 PFLOP headline
is measured at FP4 precision.
unit: "bits"
typical_range: null
related_terms: ["fp8", "fp16", "quantization", "tflops"]
related_topics: ["ai-workloads", "equations-and-bounds"]
- term: "fp8"
full_name: "8-bit Floating Point"
definition: |
Low-precision numerical format using 8 bits per value. Common
for quantized LLM inference with good accuracy/performance tradeoff.
unit: "bits"
typical_range: null
related_terms: ["fp4", "fp16", "quantization"]
related_topics: ["ai-workloads", "equations-and-bounds"]
- term: "fp16"
full_name: "16-bit Floating Point (Half Precision)"
definition: |
Standard training precision for many deep learning models.
Good balance of range, precision, and memory efficiency.
unit: "bits"
typical_range: null
related_terms: ["fp4", "fp8", "fp32"]
related_topics: ["ai-workloads", "equations-and-bounds"]
- term: "quantization"
full_name: "Model Quantization"
definition: |
Technique for reducing model memory footprint by using lower-precision
number formats (FP4, FP8, INT4, INT8) for model weights. Enables
running larger models in limited memory at some accuracy cost.
unit: null
typical_range: null
related_terms: ["fp4", "fp8", "parameter-count"]
related_topics: ["ai-workloads"]
- term: "parameter-count"
full_name: "Model Parameter Count"
definition: |
The number of trainable weights in a neural network, typically
expressed in billions (B). Determines memory requirements and
roughly correlates with model capability.
unit: "billions (B)"
typical_range: "7B-200B on single GB10, up to 400B stacked"
related_terms: ["quantization", "unified-memory"]
related_topics: ["ai-workloads", "memory-and-storage"]
- term: "dgx-os"
full_name: "NVIDIA DGX OS 7"
definition: |
NVIDIA's customized Linux distribution based on Ubuntu 24.04 LTS.
Includes pre-configured GPU drivers, CUDA toolkit, and platform
optimizations for DGX/DGX Spark hardware.
unit: null
typical_range: null
related_terms: ["ubuntu", "cuda"]
related_topics: ["dgx-os-software"]
- term: "dgx-spark"
full_name: "NVIDIA DGX Spark"
definition: |
NVIDIA's own-branded desktop AI computer using the GB10 superchip.
Same hardware as the Dell Pro Max GB10, different branding and
support channel. Priced at $2,999.
unit: null
typical_range: null
related_terms: ["gb10"]
related_topics: ["skus-and-pricing"]
- term: "connectx-7"
full_name: "NVIDIA ConnectX-7 SmartNIC"
definition: |
High-performance network interface card integrated into the
Dell Pro Max GB10. Provides 2x QSFP 200 Gbps ports, primarily
used for multi-unit stacking.
unit: "Gbps"
typical_range: "200 Gbps per port"
related_terms: ["qsfp", "smartnic"]
related_topics: ["connectivity", "multi-unit-stacking"]
- term: "qsfp"
full_name: "Quad Small Form-factor Pluggable"
definition: |
High-speed networking connector standard. The Dell Pro Max GB10
has 2x QSFP ports supporting 200 Gbps each via ConnectX-7.
unit: "Gbps"
typical_range: "200 Gbps per port in GB10"
related_terms: ["connectx-7"]
related_topics: ["connectivity", "multi-unit-stacking"]
- term: "smartnic"
full_name: "Smart Network Interface Card"
definition: |
Network adapter with onboard processing capability for offloading
network tasks from the main CPU. The ConnectX-7 in the GB10 is
a SmartNIC.
unit: null
typical_range: null
related_terms: ["connectx-7", "qsfp"]
related_topics: ["connectivity"]
- term: "10gbe"
full_name: "10 Gigabit Ethernet"
definition: |
Standard Ethernet networking at 10 Gbps. The Dell Pro Max GB10
includes one 10GbE RJ45 port for general network connectivity.
unit: "Gbps"
typical_range: "10 Gbps"
related_terms: []
related_topics: ["connectivity"]
- term: "pytorch"
full_name: "PyTorch"
definition: |
Open-source deep learning framework. Primary ML framework
supported on the GB10 with ARM64-native builds and full
CUDA acceleration.
unit: null
typical_range: null
related_terms: ["cuda", "nemo"]
related_topics: ["ai-frameworks"]
- term: "nemo"
full_name: "NVIDIA NeMo"
definition: |
NVIDIA's framework for building, customizing, and deploying
generative AI models. Supports fine-tuning (SFT, RLHF) and
is optimized for NVIDIA hardware.
unit: null
typical_range: null
related_terms: ["pytorch", "cuda"]
related_topics: ["ai-frameworks"]
- term: "rapids"
full_name: "NVIDIA RAPIDS"
definition: |
Suite of GPU-accelerated data science libraries including cuDF
(DataFrames), cuML (ML), and cuGraph (graph analytics). Drop-in
replacements for pandas, scikit-learn, and NetworkX.
unit: null
typical_range: null
related_terms: ["cuda"]
related_topics: ["ai-frameworks"]
- term: "cuda"
full_name: "Compute Unified Device Architecture"
definition: |
NVIDIA's parallel computing platform and API for GPU-accelerated
computing. Pre-installed on the GB10 via DGX OS.
unit: null
typical_range: null
related_terms: ["cuda-core", "pytorch", "nemo"]
related_topics: ["ai-frameworks", "dgx-os-software"]
- term: "ngc"
full_name: "NVIDIA NGC Catalog"
definition: |
NVIDIA's hub for GPU-optimized AI software including pre-trained
models, containers, SDKs, and Helm charts.
unit: null
typical_range: null
related_terms: ["cuda", "nemo"]
related_topics: ["ai-frameworks"]
- term: "llama-cpp"
full_name: "llama.cpp"
definition: |
Open-source C/C++ inference engine for running quantized LLMs.
Supports ARM-optimized builds for GB10 and GGUF model format.
unit: null
typical_range: null
related_terms: ["quantization"]
related_topics: ["ai-frameworks", "ai-workloads"]
- term: "fcm1253"
full_name: "Dell Pro Max FCM1253"
definition: |
Dell's model number for the Pro Max with GB10 desktop system.
Available in 2TB and 4TB storage configurations.
unit: null
typical_range: null
related_terms: ["gb10"]
related_topics: ["skus-and-pricing"]
- term: "sed"
full_name: "Self-Encrypting Drive"
definition: |
Storage drive with built-in hardware encryption. Available
on the 4TB configuration of the Dell Pro Max GB10.
unit: null
typical_range: null
related_terms: []
related_topics: ["memory-and-storage", "skus-and-pricing"]
- term: "tdp"
full_name: "Thermal Design Power"
definition: |
Maximum amount of heat a cooling system must dissipate.
The GB10 system TDP is approximately 140W.
unit: "watts"
typical_range: "~140W for GB10 system"
related_terms: []
related_topics: ["physical-specs", "gb10-superchip"]
- term: "displayport-alt-mode"
full_name: "DisplayPort Alternate Mode"
definition: |
Protocol allowing DisplayPort video signals to be carried
over a USB Type-C connector. Used for display output on
the GB10's USB-C ports.
unit: null
typical_range: null
related_terms: ["usb-c", "hdmi"]
related_topics: ["connectivity"]