inf2 Instance Family

The Amazon EC2 Inf2 instance family, generally available since April 13, 2023, is designed for deep learning inference and powered by AWS Inferentia2 chips. These instances feature dual AMD EPYC 7R13 processors and offer up to 100 Gbps of networking bandwidth. They provide up to 384 GB of shared accelerator memory and EBS-only storage. Inf2 instances are optimized for generative AI models, delivering high throughput and low latency.

Source

vCPUs
Up to 192
Memory
Up to 768 GiB
Architecture
x86_64
Processor
AMD
Ranked by CPU performance (Passmark)
#1
m8a, c8a, r8a
3846 Up to 192 vCPU
#3
m7a, c7a, r7a
2887 Up to 192 vCPU
#4
2595 Up to 192 vCPU
#5
2303 Up to 128 vCPU
#6
1853 Up to 192 vCPU
#7
m5dn, c4, r5dn, r5n, t2
1764 Up to 96 vCPU
#8
m4, m5, m5n, m7g, m7gd, c3, c7g, c7gd, c7gn, i3, i3en, r3, r4, r5, r5b, r5d, r7g, r7gd, t3
1648 Up to 96 vCPU
#9
1419 Up to 96 vCPU
#10
1098 Up to 64 vCPU
🇯🇵 ap-northeast-1
🇰🇷 ap-northeast-2
🇮🇳 ap-south-1
🇸🇬 ap-southeast-1
🇦🇺 ap-southeast-2
🇩🇪 eu-central-1
🇸🇪 eu-north-1
🇮🇪 eu-west-1
🇬🇧 eu-west-2
🇫🇷 eu-west-3
🇧🇷 sa-east-1
🇺🇸 us-east-1
🇺🇸 us-east-2
🇺🇸 us-west-2
inf2.xlarge 4 vCPU, 16 GiB

On-Demand Price

Spot Price

inf2.8xlarge 32 vCPU, 128 GiB

On-Demand Price

Spot Price

inf2.24xlarge 96 vCPU, 384 GiB

On-Demand Price

Spot Price

inf2.48xlarge 192 vCPU, 768 GiB

On-Demand Price

Spot Price