Snowpark Container Services: Azure Instance Families

Instance families are grouped as current generation or previous generation. Azure instance families are available in three types:

  • General Compute (GEN): Best price-performance for general-purpose containerized workloads.
  • High Memory (MEM): High memory-to-vCPU ratio for applications that require large amounts of RAM, such as CPU-based model serving, large-scale in-memory data processing, and vector index serving.
  • GPU Accelerated (GPU): For machine learning training, inference, and AI workloads requiring GPU acceleration.

For pricing information, see the Snowflake Service Consumption Table.

Note

Region availability is subject to change. To retrieve current availability and instance family specifications programmatically, use SHOW COMPUTE POOL INSTANCE FAMILIES.

Current Generation Instance Families

General Compute Instance Families (Current Generation)

Current generation x86 instances offering the best price-performance for general-purpose workloads.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
GEN_X64_G2_21610012.5150Available everywhere
GEN_X64_G2_431310012.5150Available everywhere
GEN_X64_G2_862810012.5150Available everywhere
GEN_X64_G2_16145810012.5150Available everywhere
GEN_X64_G2_322811610016.0150Available everywhere

High Memory Instance Families (Current Generation)

Current generation x86 instances optimized for memory-intensive workloads.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
MEM_X64_G2_865810012.5150Available everywhere
MEM_X64_G2_322824010012.5150Available everywhere
MEM_X64_G2_646049210016.0150Available everywhere
MEM_X64_G2_969265210016.0150Available everywhere

GPU Accelerated Instance Families (Current Generation)

Azure GPU instance families feature three NVIDIA GPU architectures, each suited to different AI and ML workloads.

NVIDIA T4

Turing GPU for cost-effective inference and light ML workloads.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_NV_XS3261008.01 NVIDIA T41610Not available in Switzerland North, UAE North, Central US, and UK South regions

NVIDIA A10

Mid-range Ampere GPU for ML model development and inference on small to medium models.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_NV_SM3242410040.01 NVIDIA A102410Not available in Central US
GPU_NV_2M6885810080.02 NVIDIA A10485Not available in Central US

NVIDIA A100

High-throughput Ampere GPU for large-scale model training and large dataset processing.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_NV_3M4442410040.02 NVIDIA A100160On RequestNot available in Central US, North Europe, and UAE North
GPU_NV_SL9285810080.04 NVIDIA A100320On RequestNot available in Central US, North Europe, and UAE North

Previous Generation Instance Families

General Compute Instance Families (Previous Generation)

Previous generation x86 instances. For new workloads, use GEN_X64_G2 instances instead.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
CPU_X64_XS1610012.5500Available everywhere
CPU_X64_S31310012.5500Available everywhere
CPU_X64_M62810012.5500Available everywhere
CPU_X64_SL145810012.5500Available everywhere
CPU_X64_L2811610016.0500Available everywhere

High Memory Instance Families (Previous Generation)

Previous generation x86 instances optimized for memory. For new workloads, use MEM_X64_G2 instances instead.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
HIGHMEM_X64_S6581008.0150Available everywhere
HIGHMEM_X64_M2824010016.0150Available everywhere
HIGHMEM_X64_SL6049210032.0150Available everywhere