You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
366 lines
13 KiB
366 lines
13 KiB
terms:
|
|
- term: "gb10"
|
|
full_name: "NVIDIA GB10 Superchip"
|
|
definition: |
|
|
System-on-chip combining an NVIDIA Grace CPU and Blackwell GPU
|
|
connected via NVLink-C2C. The core silicon in the Dell Pro Max GB10
|
|
and NVIDIA DGX Spark.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["grace-blackwell", "superchip", "nvlink-c2c"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "grace-blackwell"
|
|
full_name: "Grace Blackwell Architecture"
|
|
definition: |
|
|
NVIDIA's combined CPU+GPU architecture pairing a Grace ARM CPU
|
|
with a Blackwell GPU via NVLink-C2C coherent interconnect.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["gb10", "blackwell-gpu", "grace-cpu"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "superchip"
|
|
full_name: "Superchip"
|
|
definition: |
|
|
NVIDIA's term for a system-on-chip that integrates both CPU and GPU
|
|
dies on a single package with high-bandwidth interconnect.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["gb10", "soc"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "soc"
|
|
full_name: "System-on-Chip"
|
|
definition: |
|
|
An integrated circuit that combines multiple components (CPU, GPU,
|
|
memory controller, I/O) on a single die or package.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["gb10", "superchip"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "cortex-x925"
|
|
full_name: "ARM Cortex-X925"
|
|
definition: |
|
|
ARM's high-performance CPU core design (ARMv9.2 architecture).
|
|
The GB10 contains 10 of these as its "big" cores.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cortex-a725", "gb10"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "cortex-a725"
|
|
full_name: "ARM Cortex-A725"
|
|
definition: |
|
|
ARM's efficiency-focused CPU core design (ARMv9.2 architecture).
|
|
The GB10 contains 10 of these as its "LITTLE" cores.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cortex-x925", "gb10"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "blackwell-gpu"
|
|
full_name: "NVIDIA Blackwell GPU"
|
|
definition: |
|
|
NVIDIA's GPU architecture generation. In the GB10, it provides
|
|
6,144 CUDA cores and 5th-gen Tensor Cores.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cuda-core", "tensor-core", "gb10"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "cuda-core"
|
|
full_name: "CUDA Core"
|
|
definition: |
|
|
NVIDIA's basic parallel processing unit for general-purpose GPU
|
|
computing. The GB10 has 6,144 CUDA cores.
|
|
unit: "cores"
|
|
typical_range: "6,144 in GB10"
|
|
related_terms: ["blackwell-gpu", "tensor-core"]
|
|
related_topics: ["gb10-superchip"]
|
|
|
|
- term: "tensor-core"
|
|
full_name: "Tensor Core (5th Generation)"
|
|
definition: |
|
|
Specialized GPU cores for matrix multiply-accumulate operations,
|
|
critical for deep learning inference and training. 5th-gen Tensor
|
|
Cores in Blackwell support FP4, FP8, FP16, and other precisions.
|
|
unit: "cores"
|
|
typical_range: null
|
|
related_terms: ["blackwell-gpu", "fp4", "fp8"]
|
|
related_topics: ["gb10-superchip", "ai-workloads"]
|
|
|
|
- term: "nvlink-c2c"
|
|
full_name: "NVLink Chip-to-Chip"
|
|
definition: |
|
|
NVIDIA's proprietary die-to-die interconnect connecting the Grace CPU
|
|
and Blackwell GPU within the GB10 superchip. Provides 600 GB/s
|
|
bidirectional bandwidth and enables unified coherent memory.
|
|
unit: "GB/s"
|
|
typical_range: "600 GB/s bidirectional"
|
|
related_terms: ["gb10", "unified-memory"]
|
|
related_topics: ["gb10-superchip", "memory-and-storage"]
|
|
|
|
- term: "unified-memory"
|
|
full_name: "Unified Coherent Memory"
|
|
definition: |
|
|
Memory architecture where CPU and GPU share the same physical memory
|
|
pool with hardware cache coherence. Eliminates explicit host-device
|
|
memory copies. In the GB10, both processors see the full 128 GB.
|
|
unit: "GB"
|
|
typical_range: "128 GB in GB10"
|
|
related_terms: ["lpddr5x", "nvlink-c2c"]
|
|
related_topics: ["memory-and-storage", "gb10-superchip"]
|
|
|
|
- term: "lpddr5x"
|
|
full_name: "Low-Power DDR5X"
|
|
definition: |
|
|
Latest generation of low-power DRAM. In the GB10, runs at up to
|
|
9,400 MT/s providing 273 GB/s of memory bandwidth.
|
|
unit: "MT/s"
|
|
typical_range: "9,400 MT/s in GB10"
|
|
related_terms: ["unified-memory"]
|
|
related_topics: ["memory-and-storage"]
|
|
|
|
- term: "tflops"
|
|
full_name: "Tera Floating-Point Operations Per Second"
|
|
definition: |
|
|
Unit of compute performance. 1 TFLOPS = 10^12 floating-point
|
|
operations per second. ALWAYS specify the precision (FP4, FP8,
|
|
FP16, FP32) when quoting TFLOPS figures.
|
|
unit: "TFLOPS"
|
|
typical_range: "1,000 TFLOPS FP4 for GB10"
|
|
related_terms: ["pflop", "fp4"]
|
|
related_topics: ["gb10-superchip", "equations-and-bounds"]
|
|
|
|
- term: "pflop"
|
|
full_name: "Peta Floating-Point Operations Per Second"
|
|
definition: |
|
|
1 PFLOP = 1,000 TFLOPS = 10^15 floating-point operations per second.
|
|
The GB10's headline figure is 1 PFLOP at FP4 precision.
|
|
unit: "PFLOP"
|
|
typical_range: "1 PFLOP FP4 for GB10"
|
|
related_terms: ["tflops", "fp4"]
|
|
related_topics: ["gb10-superchip", "equations-and-bounds"]
|
|
|
|
- term: "fp4"
|
|
full_name: "4-bit Floating Point"
|
|
definition: |
|
|
Ultra-low precision numerical format using 4 bits per value.
|
|
Used for quantized inference. The GB10's 1 PFLOP headline
|
|
is measured at FP4 precision.
|
|
unit: "bits"
|
|
typical_range: null
|
|
related_terms: ["fp8", "fp16", "quantization", "tflops"]
|
|
related_topics: ["ai-workloads", "equations-and-bounds"]
|
|
|
|
- term: "fp8"
|
|
full_name: "8-bit Floating Point"
|
|
definition: |
|
|
Low-precision numerical format using 8 bits per value. Common
|
|
for quantized LLM inference with good accuracy/performance tradeoff.
|
|
unit: "bits"
|
|
typical_range: null
|
|
related_terms: ["fp4", "fp16", "quantization"]
|
|
related_topics: ["ai-workloads", "equations-and-bounds"]
|
|
|
|
- term: "fp16"
|
|
full_name: "16-bit Floating Point (Half Precision)"
|
|
definition: |
|
|
Standard training precision for many deep learning models.
|
|
Good balance of range, precision, and memory efficiency.
|
|
unit: "bits"
|
|
typical_range: null
|
|
related_terms: ["fp4", "fp8", "fp32"]
|
|
related_topics: ["ai-workloads", "equations-and-bounds"]
|
|
|
|
- term: "quantization"
|
|
full_name: "Model Quantization"
|
|
definition: |
|
|
Technique for reducing model memory footprint by using lower-precision
|
|
number formats (FP4, FP8, INT4, INT8) for model weights. Enables
|
|
running larger models in limited memory at some accuracy cost.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["fp4", "fp8", "parameter-count"]
|
|
related_topics: ["ai-workloads"]
|
|
|
|
- term: "parameter-count"
|
|
full_name: "Model Parameter Count"
|
|
definition: |
|
|
The number of trainable weights in a neural network, typically
|
|
expressed in billions (B). Determines memory requirements and
|
|
roughly correlates with model capability.
|
|
unit: "billions (B)"
|
|
typical_range: "7B-200B on single GB10, up to 400B stacked"
|
|
related_terms: ["quantization", "unified-memory"]
|
|
related_topics: ["ai-workloads", "memory-and-storage"]
|
|
|
|
- term: "dgx-os"
|
|
full_name: "NVIDIA DGX OS 7"
|
|
definition: |
|
|
NVIDIA's customized Linux distribution based on Ubuntu 24.04 LTS.
|
|
Includes pre-configured GPU drivers, CUDA toolkit, and platform
|
|
optimizations for DGX/DGX Spark hardware.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["ubuntu", "cuda"]
|
|
related_topics: ["dgx-os-software"]
|
|
|
|
- term: "dgx-spark"
|
|
full_name: "NVIDIA DGX Spark"
|
|
definition: |
|
|
NVIDIA's own-branded desktop AI computer using the GB10 superchip.
|
|
Same hardware as the Dell Pro Max GB10, different branding and
|
|
support channel. Priced at $2,999.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["gb10"]
|
|
related_topics: ["skus-and-pricing"]
|
|
|
|
- term: "connectx-7"
|
|
full_name: "NVIDIA ConnectX-7 SmartNIC"
|
|
definition: |
|
|
High-performance network interface card integrated into the
|
|
Dell Pro Max GB10. Provides 2x QSFP 200 Gbps ports, primarily
|
|
used for multi-unit stacking.
|
|
unit: "Gbps"
|
|
typical_range: "200 Gbps per port"
|
|
related_terms: ["qsfp", "smartnic"]
|
|
related_topics: ["connectivity", "multi-unit-stacking"]
|
|
|
|
- term: "qsfp"
|
|
full_name: "Quad Small Form-factor Pluggable"
|
|
definition: |
|
|
High-speed networking connector standard. The Dell Pro Max GB10
|
|
has 2x QSFP ports supporting 200 Gbps each via ConnectX-7.
|
|
unit: "Gbps"
|
|
typical_range: "200 Gbps per port in GB10"
|
|
related_terms: ["connectx-7"]
|
|
related_topics: ["connectivity", "multi-unit-stacking"]
|
|
|
|
- term: "smartnic"
|
|
full_name: "Smart Network Interface Card"
|
|
definition: |
|
|
Network adapter with onboard processing capability for offloading
|
|
network tasks from the main CPU. The ConnectX-7 in the GB10 is
|
|
a SmartNIC.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["connectx-7", "qsfp"]
|
|
related_topics: ["connectivity"]
|
|
|
|
- term: "10gbe"
|
|
full_name: "10 Gigabit Ethernet"
|
|
definition: |
|
|
Standard Ethernet networking at 10 Gbps. The Dell Pro Max GB10
|
|
includes one 10GbE RJ45 port for general network connectivity.
|
|
unit: "Gbps"
|
|
typical_range: "10 Gbps"
|
|
related_terms: []
|
|
related_topics: ["connectivity"]
|
|
|
|
- term: "pytorch"
|
|
full_name: "PyTorch"
|
|
definition: |
|
|
Open-source deep learning framework. Primary ML framework
|
|
supported on the GB10 with ARM64-native builds and full
|
|
CUDA acceleration.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cuda", "nemo"]
|
|
related_topics: ["ai-frameworks"]
|
|
|
|
- term: "nemo"
|
|
full_name: "NVIDIA NeMo"
|
|
definition: |
|
|
NVIDIA's framework for building, customizing, and deploying
|
|
generative AI models. Supports fine-tuning (SFT, RLHF) and
|
|
is optimized for NVIDIA hardware.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["pytorch", "cuda"]
|
|
related_topics: ["ai-frameworks"]
|
|
|
|
- term: "rapids"
|
|
full_name: "NVIDIA RAPIDS"
|
|
definition: |
|
|
Suite of GPU-accelerated data science libraries including cuDF
|
|
(DataFrames), cuML (ML), and cuGraph (graph analytics). Drop-in
|
|
replacements for pandas, scikit-learn, and NetworkX.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cuda"]
|
|
related_topics: ["ai-frameworks"]
|
|
|
|
- term: "cuda"
|
|
full_name: "Compute Unified Device Architecture"
|
|
definition: |
|
|
NVIDIA's parallel computing platform and API for GPU-accelerated
|
|
computing. Pre-installed on the GB10 via DGX OS.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cuda-core", "pytorch", "nemo"]
|
|
related_topics: ["ai-frameworks", "dgx-os-software"]
|
|
|
|
- term: "ngc"
|
|
full_name: "NVIDIA NGC Catalog"
|
|
definition: |
|
|
NVIDIA's hub for GPU-optimized AI software including pre-trained
|
|
models, containers, SDKs, and Helm charts.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["cuda", "nemo"]
|
|
related_topics: ["ai-frameworks"]
|
|
|
|
- term: "llama-cpp"
|
|
full_name: "llama.cpp"
|
|
definition: |
|
|
Open-source C/C++ inference engine for running quantized LLMs.
|
|
Supports ARM-optimized builds for GB10 and GGUF model format.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["quantization"]
|
|
related_topics: ["ai-frameworks", "ai-workloads"]
|
|
|
|
- term: "fcm1253"
|
|
full_name: "Dell Pro Max FCM1253"
|
|
definition: |
|
|
Dell's model number for the Pro Max with GB10 desktop system.
|
|
Available in 2TB and 4TB storage configurations.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["gb10"]
|
|
related_topics: ["skus-and-pricing"]
|
|
|
|
- term: "sed"
|
|
full_name: "Self-Encrypting Drive"
|
|
definition: |
|
|
Storage drive with built-in hardware encryption. Available
|
|
on the 4TB configuration of the Dell Pro Max GB10.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: []
|
|
related_topics: ["memory-and-storage", "skus-and-pricing"]
|
|
|
|
- term: "tdp"
|
|
full_name: "Thermal Design Power"
|
|
definition: |
|
|
Maximum amount of heat a cooling system must dissipate.
|
|
The GB10 system TDP is approximately 140W.
|
|
unit: "watts"
|
|
typical_range: "~140W for GB10 system"
|
|
related_terms: []
|
|
related_topics: ["physical-specs", "gb10-superchip"]
|
|
|
|
- term: "displayport-alt-mode"
|
|
full_name: "DisplayPort Alternate Mode"
|
|
definition: |
|
|
Protocol allowing DisplayPort video signals to be carried
|
|
over a USB Type-C connector. Used for display output on
|
|
the GB10's USB-C ports.
|
|
unit: null
|
|
typical_range: null
|
|
related_terms: ["usb-c", "hdmi"]
|
|
related_topics: ["connectivity"]
|