Snowpark Container Services: AWS Instance Families

Instance families are grouped as current generation or previous generation. AWS instance families are available in three types:

  • General Compute (GEN): Best price-performance for general-purpose containerized workloads.
  • High Memory (MEM): High memory-to-vCPU ratio for applications that require large amounts of RAM, such as CPU-based model serving, large-scale in-memory data processing, and vector index serving.
  • GPU Accelerated (GPU): For machine learning training, inference, and AI workloads requiring GPU acceleration.

For pricing information, see the Snowflake Service Consumption Table.

Note

Region availability is subject to change. To retrieve current availability and instance family specifications programmatically, use SHOW COMPUTE POOL INSTANCE FAMILIES.

Current Generation Instance Families

General Compute Instance Families (Current Generation)

Current generation x86 instances offering the best price-performance for general-purpose workloads.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
GEN_X64_G2_21610012.5150Not available in af-south-1, cn-northwest-1
GEN_X64_G2_431310012.5150Not available in af-south-1, cn-northwest-1
GEN_X64_G2_862810012.5150Not available in af-south-1, cn-northwest-1
GEN_X64_G2_322811610012.5150Not available in af-south-1, cn-northwest-1

High Memory Instance Families (Current Generation)

Current generation x86 instances optimized for memory-intensive workloads.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
MEM_X64_G2_865810012.5150Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1
MEM_X64_G2_322824010012.5150Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1
MEM_X64_G2_646049210025.0150Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1
MEM_X64_G2_192188143610050.0150Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1

GPU Accelerated Instance Families (Current Generation)

AWS GPU instance families feature four NVIDIA GPU architectures, each suited to different AI and ML workloads.

NVIDIA A10G

Mid-range Ampere GPU for ML model development and inference on small to medium models.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_NV_S62845010.01 NVIDIA A10G24150Not available in ap-southeast-1, eu-central-2, eu-west-3, ap-northeast-3
GPU_NV_M44178380040.04 NVIDIA A10G9610Not available in Gov Regions, ap-southeast-1, eu-central-2, eu-west-3, ap-northeast-3

NVIDIA A100

High-throughput Ampere GPU for large-scale model training and large dataset processing.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_NV_L921112100400.08 NVIDIA A100320On RequestAvailable only in AWS US West and US East non-gov regions; limited availability in other regions upon request

NVIDIA L40S

Ada Lovelace GPU optimized for GenAI inference and fine-tuning.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_L40S_G1_865845025.01 NVIDIA L40S485Only available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2
GPU_L40S_G1_161411660025.01 NVIDIA L40S485Only available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2
GPU_L40S_G1_48443683800100.04 NVIDIA L40S192On RequestOnly available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2
GPU_L40S_G1_19218814363800400.08 NVIDIA L40S384On RequestOnly available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2

NVIDIA RTX PRO 6000

Blackwell GPU with 96GB VRAM per GPU for large-scale inference and data-intensive workloads.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)GPUGPU Memory per GPU (GB)Node limitRegion Availability
GPU_R6K_G1_8658190050.01 NVIDIA RTX PRO 6000965Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1
GPU_R6K_G1_1614116190050.01 NVIDIA RTX PRO 6000965Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1
GPU_R6K_G1_32282401900100.01 NVIDIA RTX PRO 6000965Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1
GPU_R6K_G1_48444903800400.02 NVIDIA RTX PRO 6000192On RequestOnly available in us-west-2, us-east-1, us-east-2, ap-northeast-1
GPU_R6K_G1_96929847600800.04 NVIDIA RTX PRO 6000384On RequestOnly available in us-west-2, us-east-1, us-east-2, ap-northeast-1
GPU_R6K_G1_1921881843114001600.08 NVIDIA RTX PRO 6000768On RequestOnly available in us-west-2, us-east-1, us-east-2, ap-northeast-1

Previous Generation Instance Families

General Compute Instance Families (Previous Generation)

Previous generation x86 instances. For new workloads, use GEN_X64_G2 instances instead.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
CPU_X64_XS1610012.5500Available everywhere
CPU_X64_S31310012.5500Available everywhere
CPU_X64_M62810012.5500Available everywhere
CPU_X64_SL145810012.5500Not available in China
CPU_X64_L2811610012.5500Available everywhere

High Memory Instance Families (Previous Generation)

Previous generation x86 instances optimized for memory. For new workloads, use MEM_X64_G2 instances instead.

Instance FamilyvCPUMemory (GiB)Storage (GB)Bandwidth limit (Gbps)Node limitRegion Availability
HIGHMEM_X64_S65810012.5500Available everywhere
HIGHMEM_X64_M2824010012.5500Available everywhere
HIGHMEM_X64_L12498410050.0500Available everywhere