News & Events

Closer to the Action: Why Edge Computing is a Game-Changer for Real-Time Video

Unlock the full potential of Wavestore v6.44 -view our launch presentation today and explore the latest innovations in video management.

VIEW PRESENTATION

Modern high-definition IP cameras are remarkable pieces of technology, capable of capturing 4K resolution video at 30 frames per second with exceptional clarity. But this capability comes with a significant challenge: each camera can generate anywhere from 2 to 20 Mbps of continuous data, depending on resolution, compression, and scene complexity. In a deployment with 50, 100, or 500 cameras, the aggregate data flowing across your network becomes staggering—potentially hundreds of gigabits per second that must traverse network infrastructure, reach recording servers, and be stored for retention periods that might extend weeks or months.

This raises a fundamental architectural question that has profound implications for system performance, cost, and scalability: Where should video data be analysed? The traditional answer has been centralised processing—send all raw video streams to powerful servers or cloud infrastructure where analytics engines can examine the footage and generate alerts. But this approach creates bottlenecks, introduces latency, and requires expensive network infrastructure to handle the data volume.

Edge computing offers a fundamentally different solution. Rather than transmitting raw video for remote analysis, edge computing processes data "at the edge" of the network—directly on the camera itself or on nearby edge devices. By embedding intelligence at the point of capture, edge architectures can make instant decisions about what's important, transmit only actionable information, and dramatically reduce the burden on both network infrastructure and central systems. For organisations deploying large-scale surveillance or operating in bandwidth-constrained environments, edge computing isn't just an optimisation—it's often the only practical architecture that makes advanced video analytics feasible.

What is Edge Computing?

To understand edge computing, consider this analogy: Imagine a large corporation where every employee must consult headquarters before making any decision, no matter how minor. Even simple choices—whether to reorder office supplies, how to respond to a routine customer enquiry—require sending information to central management, waiting for analysis and a decision, then receiving instructions back. The headquarters becomes overwhelmed with trivial matters, communication channels are constantly congested, and decision-making is painfully slow.

Now imagine empowering employees with training and authority to make routine decisions independently. They handle standard situations on their own and only escalate exceptional cases to headquarters with a summary and recommendation. Headquarters can focus on strategic issues rather than operational minutiae, communication bandwidth is used only for what matters, and decisions happen in real time at the point of action.

Edge computing in video surveillance operates on exactly this principle. Each camera becomes an intelligent device with onboard processing capability—essentially a "mini-brain" that can analyse its own video stream. Rather than sending every frame to a central server for processing, the camera runs analytics locally: detecting objects, recognising behaviours, identifying events, and making initial assessments about what's significant. The camera then transmits only what matters: alerts when specific conditions are detected, metadata describing what was observed, and relevant video clips of actual events rather than continuous streams of empty car parks or unchanging corridors.

This distributed intelligence architecture means that a camera monitoring a secure perimeter can detect and classify a person approaching a fence, determine that this constitutes a breach, and send an alert with supporting video—all within a fraction of a second and using minimal network bandwidth. The video that doesn't show anything interesting never leaves the camera, conserving both network capacity and storage resources.

Modern edge-capable cameras incorporate dedicated processors (often called System on Chip or SoC designs) with specialised hardware for AI and video analytics. These chips can run sophisticated deep learning algorithms—the same object detection and behavioural analysis models that previously required powerful server GPUs—directly on the camera with minimal power consumption. As processor technology continues advancing, the gap between edge and centralised computing capability continues narrowing, making increasingly sophisticated analytics possible at the edge.

Solving the 3 Biggest Challenges of Centralised VMS

Edge computing directly addresses the fundamental architectural constraints that limit the scalability, performance, and resilience of traditional centralised video management systems:

The Bandwidth Bottleneck

Network bandwidth is often the primary limiting factor in video surveillance deployments. A single 4K camera streaming at full quality can consume 12-15 Mbps continuously. In a 100-camera deployment, that's 1.2-1.5 Gbps of sustained traffic—and that's before accounting for live viewing, playback requests, or any other network activity. For organisations with distributed locations connected via WAN links, or remote sites with mobile or satellite connectivity, streaming all video to a central location simply isn't feasible.

Edge computing transforms this equation by reducing transmitted data. When cameras perform local analytics and recording, they only need to send:

  • Alert notifications when events are detected (typically just a few kilobytes of metadata)
  • Relevant video clips surrounding events (perhaps 10-30 seconds of footage rather than continuous streams)
  • Periodic metadata describing scene activity (object counts, trajectory paths)
  • Lower-resolution proxy streams for remote live viewing when requested

The practical impact is profound. That same 100-camera deployment might now require only 50-150 Mbps of aggregate bandwidth—a reduction that makes advanced video analytics feasible on existing network infrastructure without costly upgrades. Organisations can deploy high-resolution cameras in bandwidth-constrained environments where centralised architectures would be impossible. Remote locations with limited connectivity options become viable for sophisticated video surveillance.

Edge computing also enables more efficient use of available bandwidth through dynamic prioritisation. When network congestion occurs, cameras can automatically adjust: reducing stream quality for routine footage whilst ensuring that event-triggered video transmits at full resolution. This intelligent bandwidth management maintains system functionality even under suboptimal network conditions.

The Latency Lag

In centralised architectures, there's an inherent delay between when something occurs in front of a camera and when the system can respond. The video must be encoded, transmitted across the network, decoded at the server, processed through analytics algorithms, and then—if an alert condition is met—a notification must be generated and transmitted back. Even with excellent network conditions, this round-trip process typically introduces 2-5 seconds of latency. In congested networks or systems with processing queues, delays can extend to 10-15 seconds or longer.

For many applications, this latency is unacceptable. A perimeter intrusion alert that arrives 10 seconds after a breach has occurred provides little opportunity for intervention. Automated response systems that trigger based on video analytics—locking doors, activating barriers, directing PTZ cameras—are ineffective if the threat has already moved past.

Edge computing eliminates this latency by performing analysis instantaneously at the point of capture. When a camera's onboard processor detects a person crossing a virtual tripwire, the alert is generated immediately—typically within 100-300 milliseconds, essentially real-time. This speed enables:

  • Immediate automated responses: Physical security systems can react to threats before they escalate
  • Real-time operational awareness: Staff receive notifications whilst they can still observe and respond to events in progress
  • Effective deterrence: Audio warnings or lighting changes can be triggered whilst intruders are still approaching, potentially preventing incidents entirely
  • Accurate tracking: Fast analytics enable continuous object tracking across multiple cameras with minimal gap

The latency advantage is particularly critical for safety applications. Fall detection in healthcare or elder care facilities, detection of workers entering hazardous zones in industrial settings, or identification of slip-and-fall risks in retail environments all require immediate alerting where seconds matter.

The Single Point of Failure

Centralised video management architectures create inherent vulnerability: if the network connection fails, cameras become disconnected from their intelligence. They may continue recording locally (if equipped with edge storage), but they lose the ability to generate alerts, trigger responses, or even be monitored remotely. Similarly, if the central VMS server experiences downtime—hardware failure, software crash, maintenance window—the entire surveillance system's analytical capability is compromised.

This single point of failure is particularly problematic for mission-critical applications and distributed deployments. A corporate headquarters losing connectivity to remote branch offices means those locations go "dark" from a management perspective. A server failure during an incident means losing real-time detection capability exactly when it's most needed.

Edge computing distributes intelligence across the system, creating resilience through redundancy. Because each camera has autonomous processing capability:

  • Cameras continue detecting and alerting even when isolated from the central system, using local rules and analytics
  • Edge storage can buffer footage locally during network outages, then synchronise to central storage when connectivity is restored
  • Distributed processing means a single server or network failure doesn't disable the entire system's analytical capability
  • Graceful degradation allows systems to maintain core functionality even under impaired conditions

This architecture is particularly valuable for:

  • Remote or distributed locations with less reliable network connectivity
  • Critical infrastructure where continuous operation is essential
  • Large-scale deployments where the likelihood of some degree of network or server issues is statistically inevitable
  • Hostile environments where network infrastructure might be deliberately targeted or compromised

Rather than creating a brittle system where everything depends on continuous connectivity to a central point, edge computing builds resilience through distributed architecture—no single component failure can take down the entire system.

Edge vs. Cloud: A Powerful Partnership

A common misconception is that edge computing and cloud-based video management are competing approaches—that choosing one means abandoning the other. In reality, the most powerful and flexible architectures leverage both, utilising each for what it does best.

Edge computing excels at real-time processing and decision-making. The camera's onboard processor can:

  • Instantly detect and classify objects entering the field of view
  • Recognise specific behaviours or events as they occur
  • Make immediate decisions about alert conditions
  • Trigger local automated responses without network round-trip delays
  • Filter and prioritise what data should be transmitted
  • Maintain functionality during network disruptions

Cloud computing excels at aggregation, long-term storage, and complex analysis. The cloud infrastructure provides:

  • Centralised management interface for configuring and monitoring all cameras across all locations
  • Scalable storage for long-term retention without local hardware investment
  • Processing power for computationally intensive analytics that would overwhelm edge devices
  • Correlation and analysis across multiple sites, cameras, and time periods
  • Machine learning model training and updates based on aggregated data
  • Accessible-anywhere interface for authorised users
  • Automatic backup and disaster recovery

The optimal architecture combines these strengths in a complementary workflow:

  1. Edge devices perform initial analysis in real-time, detecting objects, behaviours, and events as they occur
  2. Relevant events trigger transmission of associated video clips and metadata to the cloud
  3. Cloud storage retains footage for the required retention period, with automatic scaling
  4. Cloud analytics can perform secondary processing: searching across multiple cameras and sites, analysing patterns over time, or running computationally intensive algorithms
  5. Management functions occur through cloud interface: user administration, camera configuration, system health monitoring
  6. Video playback pulls footage from cloud storage or directly from cameras depending on what's being requested

This hybrid edge-cloud architecture delivers the best of both worlds: real-time responsiveness from edge processing, combined with the scalability, accessibility, and management simplicity of cloud infrastructure. Organisations aren't forced to choose between speed and sophistication—they get both.

Modern video management platforms are increasingly designed around this architecture, with seamless integration between edge analytics and cloud management. Cameras can be configured to adjust their behaviour based on network conditions: performing more processing locally when bandwidth is constrained, and offloading more to the cloud when connectivity is excellent. This adaptive approach ensures optimal performance across varying conditions and deployment scenarios.

Conclusion

Edge computing represents a fundamental architectural shift in how intelligent video surveillance systems are designed and deployed. By distributing processing capability to the point of capture rather than concentrating it in centralised infrastructure, edge architectures solve the three primary constraints that have traditionally limited surveillance system scale and effectiveness: bandwidth consumption, analytical latency, and single points of failure.

The implications for system design are substantial. Organisations can now:

  • Deploy sophisticated video analytics in bandwidth-constrained environments where it was previously impossible
  • Achieve real-time detection and response measured in milliseconds rather than seconds
  • Build resilient systems that maintain functionality even during network or server disruptions
  • Scale deployments cost-effectively by avoiding expensive network infrastructure upgrades
  • Support high-resolution cameras without proportionally increasing backend infrastructure

As camera processors continue becoming more powerful and AI algorithms become more efficient, the capabilities available at the edge will only expand. Functions that currently require server-class hardware will migrate to cameras, and the intelligence available for real-time decision-making will continue improving.

However, edge computing doesn't eliminate the need for centralised or cloud infrastructure—it changes the relationship. Modern video management systems should be architected to leverage both edge intelligence and centralised processing, utilising each for its strengths. Edge devices provide real-time awareness and immediate response; central systems provide management, storage, and complex cross-system analysis.

For system integrators and security directors designing surveillance deployments, particularly at scale or in challenging network environments, edge computing isn't just an option to consider—it's increasingly the foundation upon which effective systems must be built. The question is no longer whether to use edge processing, but how to optimally combine edge and centralised capabilities to create systems that are simultaneously intelligent, efficient, and resilient. That combination represents the current state of the art and the blueprint for future-ready video surveillance architecture.

A group of five diverse business professionals smiling and engaging in a lively meeting around a table with laptops.

View Wavestore v6.40 presentation

Solutions for a world we can't yet see. Discover v6.40 features helping people and businesses.

View Wavestore v6.44 presentation

Solutions for a world we can't yet see. Discover v6.44 features helping people and businesses.