To address this, AWS released the new G4 instances that leverage the latest generation NVIDIA T4 GPUs, custom 2nd Generation Intel Xeon Scalable (Cascade Lake) processors, up to 100 Gbps of networking throughput, and up to 1.8TB of local NVMe storage, to deliver the most cost-effective GPU instances for machine learning inference. The new G4 instances can deliver up to 65 TFLOPs of mixed-precision performance, great for inferencing. G4’s can also be used on cost-effectively for small-scale and entry-level machine learning training jobs that are less sensitive to time-to-train. For graphics-intensive workloads, the new instances offer up to 1.8x increase in graphics performance and up to 2x video transcoding capability over the previous generation G3 instances.
AWS Releases EC2 G4 Instances With NVIDIA T4 Tensor Core GPUs