Snowpark Container Services: AWS Instance Families¶
Instance families are grouped as current generation or previous generation. AWS instance families are available in three types:
General Compute (GEN): Best price-performance for general-purpose containerized workloads.
High Memory (MEM): High memory-to-vCPU ratio for applications that require large amounts of RAM, such as CPU-based model serving, large-scale in-memory data processing, and vector index serving.
GPU Accelerated (GPU): For machine learning training, inference, and AI workloads requiring GPU acceleration.
For pricing information, see the Snowflake Service Consumption Table.
Note
Region availability is subject to change. To retrieve current availability and instance family specifications programmatically, use SHOW COMPUTE POOL INSTANCE FAMILIES.
Current Generation Instance Families¶
General Compute Instance Families (Current Generation)¶
Current generation x86 instances offering the best price-performance for general-purpose workloads.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|
GEN_X64_G2_2 |
1 |
6 |
100 |
12.5 |
150 |
Not available in af-south-1, cn-northwest-1 |
GEN_X64_G2_4 |
3 |
13 |
100 |
12.5 |
150 |
Not available in af-south-1, cn-northwest-1 |
GEN_X64_G2_8 |
6 |
28 |
100 |
12.5 |
150 |
Not available in af-south-1, cn-northwest-1 |
GEN_X64_G2_32 |
28 |
116 |
100 |
12.5 |
150 |
Not available in af-south-1, cn-northwest-1 |
High Memory Instance Families (Current Generation)¶
Current generation x86 instances optimized for memory-intensive workloads.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|
MEM_X64_G2_8 |
6 |
58 |
100 |
12.5 |
150 |
Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1 |
MEM_X64_G2_32 |
28 |
240 |
100 |
12.5 |
150 |
Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1 |
MEM_X64_G2_64 |
60 |
492 |
100 |
25.0 |
150 |
Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1 |
MEM_X64_G2_192 |
188 |
1436 |
100 |
50.0 |
150 |
Not available in af-south-1, cn-northwest-1, eu-central-2, me-central-1 |
GPU Accelerated Instance Families (Current Generation)¶
AWS GPU instance families feature four NVIDIA GPU architectures, each suited to different AI and ML workloads.
NVIDIA A10G¶
Mid-range Ampere GPU for ML model development and inference on small to medium models.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
GPU |
GPU Memory per GPU (GB) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|---|---|
GPU_NV_S |
6 |
28 |
450 |
10.0 |
1 NVIDIA A10G |
24 |
150 |
Not available in ap-southeast-1, eu-central-2, eu-west-3, ap-northeast-3 |
GPU_NV_M |
44 |
178 |
3800 |
40.0 |
4 NVIDIA A10G |
96 |
10 |
Not available in Gov Regions, ap-southeast-1, eu-central-2, eu-west-3, ap-northeast-3 |
NVIDIA A100¶
High-throughput Ampere GPU for large-scale model training and large dataset processing.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
GPU |
GPU Memory per GPU (GB) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|---|---|
GPU_NV_L |
92 |
1112 |
100 |
400.0 |
8 NVIDIA A100 |
320 |
On Request |
Available only in AWS US West and US East non-gov regions; limited availability in other regions upon request |
NVIDIA L40S¶
Ada Lovelace GPU optimized for GenAI inference and fine-tuning.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
GPU |
GPU Memory per GPU (GB) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|---|---|
GPU_L40S_G1_8 |
6 |
58 |
450 |
25.0 |
1 NVIDIA L40S |
48 |
5 |
Only available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2 |
GPU_L40S_G1_16 |
14 |
116 |
600 |
25.0 |
1 NVIDIA L40S |
48 |
5 |
Only available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2 |
GPU_L40S_G1_48 |
44 |
368 |
3800 |
100.0 |
4 NVIDIA L40S |
192 |
On Request |
Only available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2 |
GPU_L40S_G1_192 |
188 |
1436 |
3800 |
400.0 |
8 NVIDIA L40S |
384 |
On Request |
Only available in us-east-1, us-east-2, us-west-2, eu-central-1, eu-north-1, ap-northeast-1, ap-northeast-2 |
NVIDIA RTX PRO 6000¶
Blackwell GPU with 96GB VRAM per GPU for large-scale inference and data-intensive workloads.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
GPU |
GPU Memory per GPU (GB) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|---|---|
GPU_R6K_G1_8 |
6 |
58 |
1900 |
50.0 |
1 NVIDIA RTX PRO 6000 |
96 |
5 |
Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1 |
GPU_R6K_G1_16 |
14 |
116 |
1900 |
50.0 |
1 NVIDIA RTX PRO 6000 |
96 |
5 |
Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1 |
GPU_R6K_G1_32 |
28 |
240 |
1900 |
100.0 |
1 NVIDIA RTX PRO 6000 |
96 |
5 |
Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1 |
GPU_R6K_G1_48 |
44 |
490 |
3800 |
400.0 |
2 NVIDIA RTX PRO 6000 |
192 |
On Request |
Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1 |
GPU_R6K_G1_96 |
92 |
984 |
7600 |
800.0 |
4 NVIDIA RTX PRO 6000 |
384 |
On Request |
Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1 |
GPU_R6K_G1_192 |
188 |
1843 |
11400 |
1600.0 |
8 NVIDIA RTX PRO 6000 |
768 |
On Request |
Only available in us-west-2, us-east-1, us-east-2, ap-northeast-1 |
Previous Generation Instance Families¶
General Compute Instance Families (Previous Generation)¶
Previous generation x86 instances. For new workloads, use GEN_X64_G2 instances instead.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|
CPU_X64_XS |
1 |
6 |
100 |
12.5 |
500 |
Available everywhere |
CPU_X64_S |
3 |
13 |
100 |
12.5 |
500 |
Available everywhere |
CPU_X64_M |
6 |
28 |
100 |
12.5 |
500 |
Available everywhere |
CPU_X64_SL |
14 |
58 |
100 |
12.5 |
500 |
Not available in China |
CPU_X64_L |
28 |
116 |
100 |
12.5 |
500 |
Available everywhere |
High Memory Instance Families (Previous Generation)¶
Previous generation x86 instances optimized for memory. For new workloads, use MEM_X64_G2 instances instead.
Instance Family |
vCPU |
Memory (GiB) |
Storage (GB) |
Bandwidth limit (Gbps) |
Node limit |
Region Availability |
|---|---|---|---|---|---|---|
HIGHMEM_X64_S |
6 |
58 |
100 |
12.5 |
500 |
Available everywhere |
HIGHMEM_X64_M |
28 |
240 |
100 |
12.5 |
500 |
Available everywhere |
HIGHMEM_X64_L |
124 |
984 |
100 |
50.0 |
500 |
Available everywhere |