Remember that piece I wrote a little while back, “Everything You Ever Wanted to Know About Network Bandwidth“? I promised then that we’d tackle the often-thorny issue of capacity planning. Well, today’s the day. So, what exactly is capacity planning?

At its heart, it’s about ensuring your network has the resources – the bandwidth, the port speed, the processing power – available when your applications and users need them, without drastically overspending or, worse, facing performance bottlenecks that cripple the business.

It sounds straightforward, but as you know, networks rarely are.

Do We Even Need Capacity Planning Anymore?

It’s a fair question, especially today. We live in an era of seemingly infinite cloud resources, elastic bandwidth from ISPs, and sophisticated SD-WAN solutions that dynamically route traffic. Doesn’t this flexibility negate the need for painstaking capacity planning?

My honest take? Absolutely not. In fact, it arguably makes smarter capacity planning more critical. Think about it:

“Elastic” isn’t free. Cloud providers and ISPs charge for bandwidth consumption. Unexpected bursts or sustained high usage driven by poorly understood demand can lead to shocking bills. Elasticity provides resilience, but cost control still demands foresight.

Performance isn’t guaranteed. SD-WAN can intelligently use multiple links, but if all those links are undersized for peak demand or if the path to a critical SaaS application is congested beyond your direct control, users still suffer.

The network edge has blurred. Your enterprise network now extends across multiple ISP backbones, cloud provider fabrics, and SaaS platforms. Understanding capacity requirements involves peering points, cloud interconnects, and gateways – areas where you might have less direct control but still bear the performance consequences.

So yes, the technologies have changed, the landscape is more complex, but the fundamental need to anticipate demand and provision accordingly remains. Ignoring it is simply kicking the can down a very expensive and potentially outage-prone road.

Building an Application-Aware Baseline

Before we even talk about prediction models, let’s talk about the foundation: understanding what is normal for your network, right now. And crucially, this baseline must be tied to application performance.

A 90% utilized WAN link – is that good or bad? Without context, it’s just a number. Is it supporting critical real-time voice traffic flawlessly, or is it choked with non-essential backups causing unacceptable latency for your CRM users?

Effective capacity planning starts with comprehensive network observability. You need granular data not just on link utilization, but on which applications are consuming bandwidth, what the user experience is for those applications (think latency, jitter, packet loss impacting specific services), and how these metrics trend over time (daily, weekly, monthly cycles). Without this application-centric view, you’re flying blind, potentially upgrading links that don’t need it while ignoring bottlenecks impacting key business processes.

The Rear-View Mirror Approach

For years, we’ve relied on traditional statistical or mathematical models. Think linear regression, moving averages, and maybe some seasonal trending (like predicting month-end batch processing spikes).

These methods are well-understood, relatively simple to implement (often built into monitoring tools), and can be effective for predictable, slow-changing workloads. They provide a quantifiable, data-driven starting point.

Their biggest weakness? They are fundamentally reactive, extrapolating the past into the future. They struggle massively with sudden, non-linear changes – think unexpected application rollouts, major cloud migrations, or the unpredictable bursts common in today’s environments. They often rely on lagging indicators and can be easily skewed by outliers or insufficient data granularity. It’s like driving by looking mostly in the rearview mirror.

The Promise of Proactive Insight?

This is where Artificial Intelligence (AI), particularly Machine Learning (ML), enters the conversation. AI-powered approaches promise to analyze vast datasets, identify complex patterns invisible to traditional methods, and potentially predict future needs with greater accuracy.

AI can model non-linear relationships and factor in many more variables simultaneously (time of day, application mix, geographic factors, even external events). It holds the potential for truly proactive capacity management, identifying subtle precursors to congestion or predicting the impact of planned changes with greater fidelity. Anomaly detection, a subset of AI, can also flag unusual traffic patterns that might signal future capacity issues or security concerns.

However, AI isn’t magic. Its predictive power is entirely dependent on the data it’s fed. It requires vast amounts of clean, high-quality, and granular data that accurately reflects the network’s behaviour and the applications running across it – think detailed flow records, path performance metrics, and application identifiers, not just basic counters. The old adage ‘garbage in, garbage out’ applies with even greater force here; incomplete, inaccurate, or biased data will inevitably lead to flawed predictions, no matter how sophisticated the algorithm.

Conclusion: It’s Not the Model, It’s the Data

So, should you ditch your spreadsheets for AI overnight? Not necessarily. Traditional methods still have a place in most scenarios. AI offers powerful potential for handling the complexity and dynamism of modern networks, enabling more proactive and accurate planning.

However, neither approach works effectively without the right foundation. The real differentiator isn’t choosing between statistical extrapolation and a sophisticated ML algorithm. It’s about having access to the right data: granular, real-time network performance metrics deeply correlated with application behavior and user experience.

Whether you use established formulas or cutting-edge AI, your capacity planning decisions will only be as good as the data feeding them. Invest in network observability that gives you that rich, application-aware context. Only then can you move beyond guesswork and make truly informed decisions to ensure your network is ready for whatever comes next, keeping your applications running smoothly and your users productive.