AWS GPU Instances

No 1 P3

P3's

P3 instances, the next-generation of EC2 compute-optimized GPU instances, are powered by up to 8 of the latest-generation NVIDIA Tesla V100 GPUs and are ideal for computationally advanced workloads such as machine learning, high performance computing, scientific computing and simulations, financial analytics, image and video processing, and data compression.

  • 1 PetaFLOP of computational performance
    • 14x better than P2
  • 300 GB/s GPU-to-GPU communication (NVLink)
    • 9X better than P2
    • 16GB GPU memory with 900 GB/sec peak GPU memory bandwidth

    Instance Size
    GPUs - Tesla V100
    GPU Peer to Peer
    GPU Memory (GB)
    vCPUs
    Memory (GB)
    Network Bandwidth
    EBS Bandwidth
    p3.2xlarge 1 N/A 16 8 61 Up to 10 Gbps 1.5 Gbps
    p3.8xlarge 4 NVLink 64 32 244 10 Gbps 7 Gbps
    p3.16xlarge 8 NVLink 128 64 488 25 Gbps 14 Gbps

No 2 G3

G3's

Amazon EC2 G3 instances are the latest generation of Amazon EC2 GPU graphics instances that deliver a powerful combination of CPU, host memory, and GPU capacity. G3 instances are ideal for graphics-intensive applications such as 3D visualizations, mid to high-end virtual workstations, virtual application software, 3D rendering, application streaming, video encoding, gaming, and other server-side graphics workloads.

  • Powered by NVIDIA Tesla M60 GPUs
  • Each GPU has up to 2,048 parallel processing cores
  • Contains hardware encoder supporting up to 10 H.265 (HEVC) 1080p30 streams and up to 18 H.264 1080p30 streams
  • Supports OpenGL, DirectX, CUDA, OpenCL, and Capture SDK (formerly known as GRID SDK)

Name GPUs vCPU Memory (GiB) GPU Memory (GiB)
g3.4xlarge 1 16 122 8
g3.8xlarge 2 32 244 16
g3.16xlarge 4 64 488 32