inf2 Instance Family

The Amazon EC2 Inf2 instance family, generally available since April 13, 2023, is designed for deep learning inference and powered by AWS Inferentia2 chips. These instances feature dual AMD EPYC 7R13 processors and offer up to 100 Gbps of networking bandwidth. They provide up to 384 GB of shared accelerator memory and EBS-only storage. Inf2 instances are optimized for generative AI models, delivering high throughput and low latency.

Source

vCPUs
Up to 192
Memory
Up to 768 GiB
Architecture
x86_64
Processor
AMD
Ranked by CPU performance (Passmark)
🇯🇵 ap-northeast-1
🇰🇷 ap-northeast-2
🇮🇳 ap-south-1
🇸🇬 ap-southeast-1
🇦🇺 ap-southeast-2
🇩🇪 eu-central-1
🇸🇪 eu-north-1
🇮🇪 eu-west-1
🇬🇧 eu-west-2
🇫🇷 eu-west-3
🇧🇷 sa-east-1
🇺🇸 us-east-1
🇺🇸 us-east-2
🇺🇸 us-west-2
inf2.xlarge 4 vCPU, 16 GiB

On-Demand Price

Spot Price

inf2.8xlarge 32 vCPU, 128 GiB

On-Demand Price

Spot Price

inf2.24xlarge 96 vCPU, 384 GiB

On-Demand Price

Spot Price

inf2.48xlarge 192 vCPU, 768 GiB

On-Demand Price

Spot Price