v6 uses ONE Lambda type: v6-worker at 10GB.
v6-worker (10GB, ~6 vCPUs): invoke: $0.0000002 per request compute: 10GB × duration × $0.0000166667/GB-s Leaf node (~200ms): $0.000033 Internal node (~100ms): $0.000017 Total per worker: ~$0.000025 avg
Binary tree: 2×leaves − 1 nodes per layer. All layers run in parallel.
| Scale | Layers | Workers/layer | Total workers | Wall clock | Cost |
|---|---|---|---|---|---|
| 250 | 5L | ~7 | ~35 | 2.0s | $0.001 |
| 1K | 5L | ~25 | ~125 | 2.4s | $0.003 |
| 5K | 15L | ~130 | ~1,950 | 5.3s | $0.05 |
| 15K | 5L | ~400 | ~2,000 | 7.3s | $0.05 |
| 150K | 5L | ~4,000 | ~20,000 | 19.8s | $0.50 |
| 1M | 5L | ~27,000 | ~135,000 | ~30s (est) | $3.40 |
| Scale | v5 cost | v5 wall | v6 cost | v6 wall | Speedup |
|---|---|---|---|---|---|
| 1K (5L) | $0.002 | ~5s | $0.003 | 2.4s | 2.1× |
| 5K (15L) | $0.02 | 46s | $0.05 | 5.3s | 8.7× |
| 15K (5L) | $0.02 | 20s | $0.05 | 7.3s | 2.7× |
| 150K (5L) | $0.14 | 270s | $0.50 | 19.8s | 13.6× |
v6 costs 2–3x more per training (10GB workers vs 2GB) but runs 3–14x faster. The cost-per-second is dramatically better. For time-sensitive workloads, v6 dominates.
v6 workers need 10GB not for memory but for CPU. At 10GB, Lambda provides ~6 vCPUs. The worker does:
Downgrading to 2GB would halve cost but triple wall clock. Not worth it for the binary recursion pattern where each hop adds ~100ms.
| Scale | Layers | Peak concurrent | Within 10K? |
|---|---|---|---|
| 5K | 15L | ~200 | ✓ |
| 15K | 15L | ~1,500 | ✓ |
| 30K | 15L | ~3,000 | ✓ |
| 150K | 5L | ~4,000 | ✓ |
| 150K | 15L | ~12,000 | ✗ (queue) |
| Scale | Layers | Peak concurrent | Within 100K? |
|---|---|---|---|
| 150K | 15L | ~12,000 | ✓ |
| 150K | 50L | ~40,000 | ✓ |
| 1M | 5L | ~27,000 | ✓ |
| 1M | 15L | ~80,000 | ✓ |
| 1M | 25L | ~135,000 | ✗ |
| Component | Monthly | Type |
|---|---|---|
| EC2 t4g.micro (proxy) | $3 | Fixed |
| S3 storage | ~$0.50 | Fixed |
| Route53 + CloudWatch | ~$1.50 | Fixed |
| Total fixed | $5/mo | |
| Lambda compute | $0.003–$3.40/train | Variable |
v6’s 3–10x speed improvement enables new use cases:
| Scale | v6 cost | Per row | Comparable to |
|---|---|---|---|
| 1K | $0.003 | $0.000003 | Free tier |
| 15K | $0.05 | $0.0000033 | Cheaper than SageMaker inference |
| 150K | $0.50 | $0.0000033 | Same per-row at scale |
| 1M | $3.40 | $0.0000034 | Linear scaling |
Charles Dana · Monce SAS · snakebatch.aws.monce.ai · April 2026
Co-Authored-By: Claude (Anthropic)