Kubernetes Requests & Limits Calculator
Requests drive scheduling. This tool converts per-pod requests into cluster totals and estimates how many nodes you need given allocatable overhead and max pods per node. Use baseline vs peak to stress-test the sizing.
Maintained by CloudCostKit Editorial Team. Last updated: 2026-02-23. Editorial policy and methodology.
Best next steps
Use this calculator for the first estimate, then validate the answer with the closest guide or companion tool.
Inputs
Results
| Scenario | Pods | Nodes | CPU req (cores) | Mem req (GiB) |
|---|---|---|---|---|
| Baseline | 60 | 3 | 15 | 30 |
| Peak | 75 | 3 | 18.75 | 37.5 |
| Delta | 15 | 0 | 3.75 | 7.5 |
Limits (burst risk)
| Metric | Total |
|---|---|
| CPU limits | 30 cores |
| Memory limits | 60 GiB |
Use the peak multiplier to model traffic spikes, and set max pods per node to capture CNI pod caps.
Requests decide scheduling, limits decide risk
This page is fundamentally about scheduler truth. Requests decide where workloads can be placed. Limits define burst boundaries and failure behavior. If those two ideas are blurred together, the cluster can look efficient on paper and still be unstable or over-provisioned in production.
- Requests: the scheduler's reservation signal and the main sizing input.
- Limits: the ceiling that shapes burst behavior, throttling, and OOM risk.
- Other constraint: pod caps can beat CPU and memory as the real scaling boundary.
Where requests and limits models usually fail
- A calm average request is used without a separate burst or deployment scenario.
- Limits are set so high that they stop being useful operational guidance.
- CPU and memory are not tested independently, so the wrong bottleneck gets blamed.
- Pod-cap and allocatable constraints are ignored because the resource math looked fine.
What to validate before handing sizing to a cost page
- Use representative workload history, not a single quiet week or a single incident peak.
- Review CPU and memory as separate scheduling boundaries and keep the higher resulting node pressure visible.
- Leave headroom for rollouts, burst traffic, and noisy-neighbor protection instead of sizing to the edge.
- Only after requests and limits are believable should you move on to node or cluster pricing.
Baseline vs burst-sensitive scheduling scenarios
| Scenario | Pods | Requests | Pod cap |
|---|---|---|---|
| Baseline | Expected | Typical | 110 |
| Peak | High | Same | 110 |
How to review requests and limits in production
- Check whether CPU, memory, or pod count is the actual reason the cluster needed more nodes under load.
- If production scheduling differs from plan, identify whether the miss came from requests, limits, or hidden pod-cap constraints.
Next steps
Example scenario
- 60 pods with 250m CPU and 512Mi requests -> estimate total requests and node count for an 8 vCPU / 32 GiB node.
- 120 pods with a 110 pods/node cap -> see how pod limits increase node count.
Included
- Totals for CPU/memory requests and limits from per-pod values and pod count.
- Node count estimate based on allocatable percentage and max pods per node.
- Baseline vs peak comparison and bottleneck label.
Not included
- Bin packing constraints (affinities, taints, topology spread) and daemonset overhead.
- Network, storage, and control plane costs.
How we calculate
- Total requests = pods x per-pod request (CPU and memory).
- Allocatable per node = node capacity x allocatable percentage.
- Node estimate uses the largest of CPU, memory, and pod-limit counts.
- Compare baseline vs peak to understand scaling risk.
FAQ
Why not size based on limits?
What should I use for allocatable %?
What is max pods per node?
Does this include per-node overhead like daemonsets?
What about cluster autoscaler and bin packing?
How do I turn this into a cost estimate?
Related tools
Related guides
Disclaimer
Educational use only. Not legal, financial, or professional advice. Results are estimates based on the inputs and assumptions shown on this page. Verify pricing and limits with your providers and documentation.
Last updated: 2026-02-23. Reviewed against CloudCostKit methodology and current provider documentation. See the Editorial Policy .