Skip to main content
  1. Home
  2. >
  3. AWS
  4. >
  5. SAA-C03
  6. >
  7. AWS SAA-C03 Exam Scenarios
  8. >
  9. Spot vs On-Demand for Stateless Containers | SAA-C03

Spot vs On-Demand for Stateless Containers | SAA-C03

Jeff Taakey
Author
Jeff Taakey
21+ Year Enterprise Architect | Multi-Cloud Architect & Strategist.

While preparing for the AWS SAA-C03, many candidates get confused by container hosting options and pricing models. In the real world, this is fundamentally a decision about operational overhead vs. compute cost optimization. Let’s drill into a simulated scenario.

The Scenario
#

StreamPulse Analytics, a growing media monitoring startup, is migrating its log-processing pipeline to AWS. The pipeline consists of 12 microservices packaged as Docker containers that parse social media streams, perform sentiment analysis, and aggregate results into S3-based data lakes.

Key Technical Context:

  • All services are stateless (no persistent storage on compute layer)
  • Workloads can gracefully handle infrastructure interruptions (built-in retry logic and idempotency)
  • Traffic patterns are unpredictable, requiring elastic scaling
  • The CFO mandates minimum operational overhead — the team has no dedicated Kubernetes experts

Key Requirements
#

Design a container hosting solution that minimizes both compute costs and operational management burden while accommodating interruption-tolerant workloads.

The Options
#

  • A) Deploy containers on EC2 Spot Instances managed by an Auto Scaling group
  • B) Deploy containers on Amazon EKS using Spot Instances in managed node groups
  • C) Deploy containers on EC2 On-Demand Instances managed by an Auto Scaling group
  • D) Deploy containers on Amazon EKS using On-Demand Instances in managed node groups

Correct Answer
#

Option A — EC2 Auto Scaling with Spot Instances.

Step-by-Step Winning Logic
#

This solution represents the optimal cost-complexity intersection for the stated requirements:

  1. Cost Optimization: Spot Instances provide 70-90% savings vs. On-Demand pricing
  2. Interruption Tolerance: The workload is explicitly designed to handle instance terminations
  3. Minimal Operational Overhead:
    • No need to manage Kubernetes control plane complexity
    • Auto Scaling groups handle basic lifecycle management
    • Native integration with Application Load Balancers for service discovery
  4. Stateless Architecture: No persistent data means instance replacements have no data loss risk

Key Exam Principle: When a scenario explicitly states “stateless” + “interruption-tolerant” + “minimal operational overhead,” always prefer the simplest compute orchestration model (EC2 ASG) over heavyweight solutions (EKS).


💎 The Architect’s Deep Dive: Why Options Fail
#

The Traps (Distractor Analysis)
#

Why not Option B? (EKS + Spot)
#

  • Unnecessary Complexity: While EKS supports Spot Instances, you’re paying for:
    • EKS control plane costs ($0.10/hour = ~$73/month)
    • Additional operational burden (managing node groups, RBAC, CNI plugins)
    • Kubernetes expertise requirements
  • FinOps Impact: The control plane cost alone eliminates 5-10% of Spot savings
  • Scenario Mismatch: No mention of advanced orchestration needs (service mesh, complex scheduling, multi-tenant isolation)

Why not Option C? (EC2 ASG + On-Demand)
#

  • Cost Inefficiency: Violates the “minimize cost” requirement
  • Wasted Reliability: Paying On-Demand premiums for workloads that don’t need 99.99% availability
  • FinOps Failure: Could be spending 10x more than Option A for identical functionality

Why not Option D? (EKS + On-Demand)
#

  • Worst of Both Worlds: Highest cost (On-Demand + EKS control plane) + highest complexity
  • Anti-Pattern: Using enterprise Kubernetes for simple stateless microservices without justification

💎 Professional Decision Matrix

This SAA-C03 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access

The Architect Blueprint
#

graph TD
    A[Application Load Balancer] --> B[Target Group]
    B --> C[EC2 Auto Scaling Group
Spot Instances] C --> D1[Container: Parser Service] C --> D2[Container: Sentiment Service] C --> D3[Container: Aggregator Service] D1 --> E[(S3 Data Lake)] D2 --> E D3 --> E F[CloudWatch Metrics] -->|Scaling Policies| C G[Spot Instance Pool
Multiple AZs] -.->|Launch Templates| C style C fill:#ff9900,stroke:#232f3e,stroke-width:3px,color:#fff style E fill:#569a31,stroke:#232f3e,stroke-width:2px style A fill:#8c4fff,stroke:#232f3e,stroke-width:2px

💎 Professional Decision Matrix

This SAA-C03 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access

Diagram Note: Spot Instances in an Auto Scaling group automatically replace interrupted instances while the ALB routes traffic only to healthy targets, ensuring zero application-level downtime despite infrastructure interruptions.

Real-World Practitioner Insight
#

Exam Rule
#

For the SAA-C03 exam, when you see “stateless” + “interruption-tolerant” + “minimize cost and operations”, immediately select EC2 Auto Scaling + Spot Instances over EKS-based solutions.

Real World
#

In production environments, we’d enhance this architecture with:

  1. Spot Fleet Diversification:

    • Mix multiple instance types (m5.large, m5a.large, c5.large) to reduce interruption probability
    • Use Capacity-Optimized allocation strategy (60% lower interruption rates vs. lowest-price)
  2. Hybrid Pricing Strategy:

    • Run 20% baseline capacity on On-Demand instances
    • Use Spot for burst capacity (80% of fleet)
    • Consider Savings Plans for predictable base load
  3. Graceful Shutdown Handlers:

    • Implement 2-minute Spot interruption notice handling
    • Use ALB connection draining (deregistration delay = 120s)
    • Add application-level checkpointing for long-running tasks
  4. When to Actually Use EKS:

    • Multi-tenant SaaS platforms requiring namespace isolation
    • Complex service mesh requirements (Istio, Linkerd)
    • Advanced scheduling needs (GPU affinity, topology constraints)
    • Teams with existing Kubernetes expertise and GitOps workflows

Cost Reality Check: For a 10-instance cluster running 24/7:

  • Option A: ~$150/month (m5.large Spot average)
  • Option D: ~$1,825/month (m5.large On-Demand + EKS control plane)

The 12x cost difference makes Option A a clear winner for the stated requirements.

💎 Professional Decision Matrix

This SAA-C03 professional section is locked.
Free beta access reveals the exam logic.

100% Free Beta Access