What makes monitoring layer is where AI surveillance value is realised important? Installing AI-enabled cameras is necessary but not sufficient. The value of AI video analytics is not in the camera or the model β it is in what happens when an event is detected. The monitoring layer β who or what receives the alert, how it is evaluated, and what action follows β determines whether a surveillance system improves security outcomes or merely generates records after the fact. The central question for any CCTV monitoring deployment is: what combination of AI automation and human review provides the best coverage, response time, and cost efficiency for the specific environment? There is no universal answer. The right balance depends on the required response actions, the acceptable false positive rate in human review queues, and the consequence of missed events. For the technical foundation of observable CV pipelines that support this monitoring architecture, see observable CV pipelines for CCTV. Practical comparison Dimension AI Automated Monitoring Human Monitoring Coverage 100% of cameras, 24/7, simultaneous Limited by number of operators; attention degrades over time Consistency Consistent β same threshold applied to every frame Inconsistent β human attention varies by time, fatigue, workload Response to detected events Immediate (milliseconds) for configured event types Variable β seconds to minutes depending on alert queue and staffing Complex judgment Poor β AI classifies against trained categories Strong β humans contextualise, infer intent, assess ambiguity False positive filtering Limited β threshold tuning reduces but cannot eliminate FPs Effective β humans quickly discard obvious false positives Cost at scale Low marginal cost per camera Linear cost increase with camera count Auditability High β every inference logged with evidence Variable β human decisions not always documented Regulatory compliance evidence Strong β automated logs provide evidence chain Weaker β reliant on human documentation discipline The implication: AI automation is most valuable where consistent, rapid detection of specific, well-defined events is required across many cameras simultaneously. Human monitoring is most valuable where context, judgment, and response to ambiguous situations is required. What AI monitoring handles well After-hours perimeter monitoring: detecting any person entering a restricted zone outside business hours. The event definition is simple (person present in zone during hours when no one should be present), the environment is predictable, and false positives can be managed through zone configuration. In our experience, this is consistently the highest-reliability use case for AI monitoring. Access control verification: detecting that a person is present when an access credential is used, or detecting multiple people entering on a single credential (tailgating). The scenario is constrained, the camera placement is fixed, and the action is specific (log event, alert security desk). Parking and vehicle management: detecting unauthorised vehicles, detecting specific vehicle types, monitoring occupancy. Vehicles are large, visually distinct, and their presence is unambiguous. People counting and flow monitoring in defined zones. Alert routing and evidence assembly: AI can detect a potential event, clip the relevant footage, attach metadata (timestamp, camera, detection class, confidence), and route to the appropriate reviewer β reducing the cognitive load on human operators and ensuring all relevant footage is immediately accessible. What AI monitoring does not handle well Complex behavioural judgment: determining whether an interaction between two people is a dispute, a transaction, an assault, or a friendly argument requires human contextual understanding. AI can flag unusual proximity, movement patterns, or physical contact β but the classification of intent is beyond reliable automation. Novel event types: AI monitors detect what they were trained to detect. An event type not in the training distribution β a novel social engineering approach, an unusual method of entry, a new theft method β will not be detected reliably. Human monitors can notice βsomething looks wrongβ without an explicit category to match against. Cross-camera reasoning: tracking a subject across multiple cameras and reasoning about their route through a building, or correlating events on different cameras to reconstruct a sequence, requires either sophisticated multi-camera tracking systems or human synthesis. Current automated multi-camera tracking is reliable in controlled, low-occlusion environments; building-wide tracking with occlusion and camera handoffs remains difficult. Response actions beyond alerting: AI can detect and alert; it cannot physically respond. For events requiring a security response β dispatch to location, remote door lock, intercom contact β a human must make the decision and take the action. Cost comparison Human monitoring cost calculation for 24/7 operation: Minimum staffing: 1 operator per shift Γ 3 shifts Γ 365 days = 1,095 operator-shifts per year At a fully-loaded cost of Β£40,000/year per operator (UK benchmark including employer costs), 24/7 monitoring requires a minimum of 4β5 FTEs (to cover shifts, holidays, and illness): Β£160,000β200,000/year This assumes one operator monitors all cameras; effective monitoring typically limits one operator to 12β16 cameras with active scanning AI monitoring platform cost: Commercial AI VMS platforms: Β£50β150/camera/year for analytics licensing For a 50-camera system: Β£2,500β7,500/year Infrastructure (servers, network): Β£10,000β30,000 capital, Β£2,000β5,000/year maintenance Human review for alerts: 1β2 operators reviewing AI-generated alerts (lower cognitive load than continuous monitoring): Β£80,000β100,000/year Total cost comparison for 50-camera system: Model Annual Operating Cost Notes 24/7 human monitoring Β£160,000β200,000 Minimum coverage; attention limitations at night AI-only (alerts to on-call) Β£15,000β45,000 Response delay; unhandled event types AI + human review (hybrid) Β£95,000β130,000 Best balance; human review of AI-generated alerts The hybrid model β AI for detection and triage, human review for evaluation and response β delivers cost efficiency while retaining human judgment for complex decisions. Alert response workflow checklist Alert categories defined with explicit response procedures for each Response time SLA defined per alert category (intrusion: 30 seconds; loitering: 5 minutes) Alert routing configured β which alerts go to human review vs automated response Alert queue management in place β alerts must be acknowledged and resolved, not accumulate Escalation path defined for unacknowledged alerts Out-of-hours response procedure documented (on-call, remote access, third-party response) Alert review staffing calculated based on expected alert volume and response SLA Performance metrics tracked: mean time to acknowledge, false positive rate, miss rate Monitoring quality degradation over time Both human and AI monitoring degrade without active management. Human monitors experience vigilance decrement β attention drops after 20β30 minutes of continuous monitoring, which is why video wall monitoring is less effective than alert-driven review. AI models experience distribution shift β environmental changes cause false alarm rates to drift upward, and new event types enter the environment that the model was not trained to detect. Active monitoring quality management means: tracking false positive and false negative rates, recalibrating AI thresholds periodically, retraining models when environmental conditions change, and maintaining operator engagement through active tasking rather than passive observation. In our experience, systems deployed without a quality management process degrade within 6β12 months to a state where either operators ignore alerts or the alert volume is throttled to the point where real events are missed.