Skip to main content
  1. Home
  2. >
  3. AWS
  4. >
  5. SAP-C02
  6. >
  7. AWS SAP-C02 Exam Scenarios
  8. >
  9. EC2 Placement Groups—Speed vs Resilience | SAP-C02

EC2 Placement Groups—Speed vs Resilience | SAP-C02

Jeff Taakey
Author
Jeff Taakey
21+ Year Enterprise Architect | Multi-Cloud Architect & Strategist.

While preparing for the AWS SAP-C02, many candidates confuse the three placement group types or assume Enhanced Networking alone solves latency issues. In the real world, this is fundamentally a decision about network topology optimization vs. infrastructure resilience. Let’s drill into a simulated scenario.

The Scenario
#

A financial analytics firm, QuantEdge, is deploying a real-time trading signal processor that runs distributed calculations across exactly five Amazon EC2 instances in the us-east-1 region. The application uses a custom UDP-based protocol for inter-node communication and requires:

  • Sub-millisecond latency between instances
  • 25 Gbps+ aggregate throughput for data exchange
  • No built-in fault tolerance (the application terminates and restarts if any node fails)

The architecture team must select an EC2 deployment strategy that maximizes network performance without over-engineering resilience features the application doesn’t use.

Key Requirements
#

Deploy five EC2 instances with the highest possible network performance while adhering to the application’s single-region, no-fault-tolerance design.

The Options
#

  • A) Launch five new EC2 instances in a Cluster Placement Group, ensuring the instance type supports Enhanced Networking.
  • B) Launch five new EC2 instances in an Auto Scaling Group within the same Availability Zone, attaching an additional Elastic Network Interface (ENI) to each instance.
  • C) Launch five new EC2 instances in a Partition Placement Group, ensuring the instance type supports Enhanced Networking.
  • D) Launch five new EC2 instances in a Spread Placement Group, attaching an additional Elastic Network Interface (ENI) to each instance.

Correct Answer
#

Option A — Cluster Placement Group with Enhanced Networking.

Step-by-Step Winning Logic
#

  1. Cluster Placement Groups place instances in a single rack within a single AZ, enabling:

    • 10 Gbps+ single-flow bandwidth (up to 100 Gbps with supported instance types like c5n.18xlarge)
    • Sub-100 microsecond latency via same-spine network topology
    • Full bisection bandwidth for tightly coupled workloads
  2. Enhanced Networking (ENA or Intel 82599 VF) provides:

    • Up to 100 Gbps network throughput
    • Lower jitter and CPU utilization for packet processing
    • Required for cluster placement group benefits to materialize
  3. No Fault Tolerance Requirement = No need for multi-AZ distribution:

    • The application restarts on failure (no HA requirement)
    • Cluster groups’ single-point-of-failure risk is acceptable here

💎 Professional-Level Analysis
#

This section breaks down the scenario from a professional exam perspective, focusing on constraints, trade-offs, and the decision signals used to eliminate incorrect options.

🔐 Expert Deep Dive: Why Options Fail
#

This walkthrough explains how the exam expects you to reason through the scenario step by step, highlighting the constraints and trade-offs that invalidate each incorrect option.

Prefer a quick walkthrough before diving deep?
[Video coming soon] This short walkthrough video explains the core scenario, the key trade-off being tested, and why the correct option stands out, so you can follow the deeper analysis with clarity.

🔐 The Traps (Distractor Analysis)
#

This section explains why each incorrect option looks reasonable at first glance, and the specific assumptions or constraints that ultimately make it fail.

The difference between the correct answer and the distractors comes down to one decision assumption most candidates overlook.

  • Why not B (Auto Scaling Group + Extra ENI)?

    • ASGs don’t guarantee physical proximity—instances may land on different racks/switches
    • Additional ENIs do not increase total bandwidth (they share the instance’s network cap)
    • No network topology optimization = higher latency (typically 100-300μs)
  • Why not C (Partition Placement Group)?

    • Partition groups spread instances across isolated hardware racks to reduce correlated failures
    • This increases latency (cross-rack communication) compared to Cluster groups
    • Designed for distributed systems like Hadoop/Cassandra, not HPC/low-latency apps
  • Why not D (Spread Placement Group + ENI)?

    • Spread groups enforce 7 instances max per AZ, each on distinct hardware
    • Worst latency of all placement options (cross-rack, cross-switch paths)
    • Extra ENIs are irrelevant—network performance is constrained by physical distance

💎 Professional Decision Matrix

This SAP-C02 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access

🔐 The Solution Blueprint
#

This blueprint visualizes the expected solution, showing how services interact and which architectural pattern the exam is testing.

Seeing the full solution end to end often makes the trade-offs—and the failure points of simpler options—immediately clear.

graph TB
    subgraph "Single Availability Zone - us-east-1a"
        subgraph "Cluster Placement Group
(Same Rack - Low Latency Network)" EC2_1[EC2 Instance 1
c5n.9xlarge
Enhanced Networking] EC2_2[EC2 Instance 2
c5n.9xlarge
Enhanced Networking] EC2_3[EC2 Instance 3
c5n.9xlarge
Enhanced Networking] EC2_4[EC2 Instance 4
c5n.9xlarge
Enhanced Networking] EC2_5[EC2 Instance 5
c5n.9xlarge
Enhanced Networking] end end EC2_1 <--> EC2_2 EC2_2 <--> EC2_3 EC2_3 <--> EC2_4 EC2_4 <--> EC2_5 EC2_5 <--> EC2_1 Note[UDP High-Frequency Trading Data
Sub-100μs Latency
50 Gbps Aggregate Throughput] style EC2_1 fill:#FF9900,stroke:#232F3E,stroke-width:2px,color:#fff style EC2_2 fill:#FF9900,stroke:#232F3E,stroke-width:2px,color:#fff style EC2_3 fill:#FF9900,stroke:#232F3E,stroke-width:2px,color:#fff style EC2_4 fill:#FF9900,stroke:#232F3E,stroke-width:2px,color:#fff style EC2_5 fill:#FF9900,stroke:#232F3E,stroke-width:2px,color:#fff style Note fill:#232F3E,stroke:#FF9900,stroke-width:2px,color:#fff

Diagram Note: All instances reside on the same physical rack, connected via a non-blocking network fabric, enabling full-mesh 50 Gbps communication with sub-100μs latency.

💎 Professional Decision Matrix

This SAP-C02 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access

🔐 The Decision Matrix
#

This matrix compares all options across cost, complexity, and operational impact, making the trade-offs explicit and the correct choice logically defensible.

At the professional level, the exam expects you to justify your choice by explicitly comparing cost, complexity, and operational impact.

Option Est. Complexity Est. Monthly Cost Network Latency Max Throughput Pros Cons
A - Cluster + ENA Low Baseline ($1,500/mo for 5× c5n.9xlarge) 50-100 μs Up to 50 Gbps ✅ Lowest latency
✅ Highest throughput
✅ No extra cost
⚠️ Single AZ (no HA)
⚠️ Limited to 1 AZ capacity
B - ASG + Extra ENI Medium Baseline + $25/mo (5 ENIs × $5/mo) 100-300 μs 10-25 Gbps ✅ ASG automation ❌ No placement guarantee
❌ ENI doesn’t boost bandwidth
C - Partition + ENA Medium Baseline 200-500 μs 10-25 Gbps ✅ Hardware isolation ❌ Higher latency (cross-rack)
❌ Designed for fault-tolerant systems
D - Spread + ENI Medium Baseline + $25/mo 300-800 μs 10 Gbps ✅ Max isolation (7 instances) ❌ Worst latency
❌ Limits scale to 7/AZ
❌ Unnecessary isolation

FinOps Insight
#

  • Option A delivers 5-8x better latency than Spread groups at zero incremental cost
  • Enhanced Networking is free (included in modern instance types like C5n, C6gn, P4d)
  • The only risk is AZ-level failure—acceptable for apps with no HA requirement

💎 Professional Decision Matrix

This SAP-C02 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access

🔐 Real-World Practitioner Insight
#

This section connects the exam scenario to real production environments, highlighting how similar decisions are made—and often misjudged—in practice.

This is the kind of decision that frequently looks correct on paper, but creates long-term friction once deployed in production.

Exam Rule
#

For the SAP-C02 exam:

  • Cluster Placement Group = “highest performance,” “low latency,” “HPC,” “tightly coupled”
  • Spread Placement Group = “max availability,” “7 instances/AZ limit”
  • Partition Placement Group = “large distributed systems,” “Hadoop/Cassandra”

Key Signal: The phrase “no fault tolerance requirement” = you can sacrifice multi-AZ for performance.

Real World
#

In production at QuantEdge, we would likely:

  1. Add monitoring: CloudWatch + custom UDP packet loss metrics to detect rack-level issues
  2. Implement graceful degradation: If cluster capacity is exhausted, fail over to Partition group in a secondary region
  3. Use Elastic Fabric Adapter (EFA) for MPI-based HPC workloads requiring RDMA (not mentioned in the question, but critical for ML training)
  4. Test instance launch failure rates: Cluster groups can fail to launch during capacity constraints—have a fallback script to retry in a different AZ

The hidden trade-off: AWS doesn’t guarantee cluster placement group capacity. For SLA-critical apps, you’d negotiate Reserved Capacity or use Capacity Reservations (adds ~5% cost premium).

💎 Professional Decision Matrix

This SAP-C02 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access