IQ and AI

Issue 6 2025 Editor's Choice, Surveillance, AI & Data Analytics

How do we balance the use of human CCTV control room surveillance operators and AI systems as societies and organisations rush to implement AI? At this stage, in a ‘shootout’ for effectiveness of crime detection on a single monitor showing one camera view, I would bet on the success of a skilled and trained human operator over an AI to detect theft activity almost every time over an hour.

The human IQ has several competitive advantages, including the essential capacity to view things in context. Aspects of the scene can be weighed up against each other to make sense of the scene and ascribe intention and purpose. Having said that, not everyone does this effectively, and it is a particular focus of my advanced surveillance training. Operators trained in crime behaviour detection are capable of spotting small or subtle crime behaviour indicators that are beyond the capabilities of working AI systems. The camouflaging of behaviour in places like crowded settings is also beyond AI’s ability to recognise and comprehend.

Yet it makes a lot of sense for AI to be used to offset human limitations and the cost of human resources to staff security systems. Always on, never tiring, and capable of monitoring hundreds of cameras simultaneously, this means organisations are considering balancing or even replacing natural or human surveillance with AI. Covering long perimeters of a major building, mine site, or factory for intrusion attempts is done more effectively with automated detection than with operators regularly sweeping the areas with cameras.

Automatically supplying information from facial recognition and ANPR, linked to networked personal profile information, is also expected to expand the information base for informed decision-making by operators. Finding the balance between this IQ versus AI approach will define the nature of security over the next 10 years.

Understanding AI beyond chat

When we discuss AI, our impressions are often shaped by interactions with systems like ChatGPT and Copilot. The creation of imaginary scenes and people, faking the appearance and voices of individuals, and the fabrication of what appears real, but has no basis in fact, is becoming increasingly prevalent in everyday media. These almost miraculous capabilities have been transferred to an expectation of AI in the security area, without much foundation.

Modern AI security perimeter systems are based on video analytics that date back several years, and movement detection that dates back even further. Although the AI capacity of algorithms providing analysis and interpretation of images has increased along with the processing power of technology, systems for areas such as perimeter protection are still largely based on pixel analysis of 2D video and images. Using pixels, we can detect changes in movement, shape, colour, and even temperature, but this is still far from the capacity of large-scale AI systems like ChatGPT and its competitors.

While mainstream AI systems may suffer from so-called ’hallucinations’, the false alarm rates involved in the detection of perimeter AI may amount to hundreds, thousands, or even tens of thousands in sites with long perimeters to protect. Currently, we accept that AI generates these false alarms. because these systems are being ‘nursemaided’ by human operators who are having to bear with these efforts, and often are supposed to verify and try and teach AIs.

As part of this care and development function, however, human operators are being distracted by excessive false alarms and pulled away from their key tasks, sometimes for several hours per shift. Despite the ‘it can learn’ marketing claims that AI product promoters make, I have found in my experiences that they do this very poorly. AI providers have shifted responsibility for learning to their clients, when such technology should largely have the capacity for accurate detection built in. I have seen multiple examples of cats triggering perimeter AI detection systems, which seems to be one of the more simple things that they should get right. You may need to consider investing in a specialised false alarm handling package.

Narrow range and restricted flexibility

Security AI systems require tightly defined parameters to function effectively. The greater the variability of what is being viewed and categorised, the higher the false alarm rate. In a sense, this kind of surveillance is akin to a railway – it works best within a narrow range and its flexibility is restricted. Machine learning needs extensive repetition to learn the characteristics or shape of even one type of object.

With perimeter protection, for example, the further the detection fields move away from the perimeter fence, the greater the variability of conditions and potential for false alarms. This is especially true when you have little control over conditions outside the perimeter, such as vegetation, weather conditions, insects, and animals.

Conversely, human surveillance is more akin to driving a car, allowing for easier and more flexible movement, and even stepping out to investigate specific issues with the use of PTZ cameras or digital zoom capabilities. The AI security systems are also not tailor-built for a client – they are typically off-the-shelf solutions that may have a range of capabilities tweaked for a client. Humans bring to the control room a knowledge of the site, the people and movements, expected behaviour and patterns of behaviour, and a knowledge of a range of policies and procedures that govern behaviour which make it easier to pick up unusual or incident circumstances. A human operator can generally adjust to even unique or unexpected conditions.

When reading about the introduction or marketing of AI systems, we see little reporting on the accuracy rates. Yet different products within the same categories have significantly different accuracy rates, for example, in facial recognition software. Different AI capabilities also vary in their accuracy rates. Part of this may be that what goes into the system determines what the AI can do with it. Even with generic large AI systems, it is often possible to obtain incorrect or inappropriate results.

Pixel-based analysis is more prone to this issue because it must deal with overlapping objects within the environment and distinguish between them. Firearm detection is also affected by this, and there have been some high-profile incidents recently where the lack of detection and activation has been blamed on the shooter’s position and location relative to cameras. More recently, a high-profile case of mistaken identification of a firearm in Baltimore resulted in multiple police vehicles responding to a teen’s empty bag of chips as a possible firearm.

A group of teens at a local high school waiting for a lift suddenly found themselves confronted by police, and the teen with the chips packet found himself cuffed and searched. The incident prompted a statement on social media from Baltimore County Councilman Izzy Patoka that “No child in our school system should be accosted by police for eating a bag of Doritos,” and calling on the school district “to review procedures around its AI-powered weapon detection system.”

Protocols and procedures, and context

The Baltimore incident reinforced the importance of having standardised protocols and procedures to deal with such events, covering the relevant parties and their involvement in any AI notification. There are questions when things like when firearm detection becomes effective relative to the incident – it is likely that the crisis has already begun by the time the firearm can be recognised easily by identifying someone pointing a gun. Ironically, in these situations, AI detection of behaviours like having hands up, lying on the floor, or groups of people running in the direction of an exit may be more effective in detecting firearm-related incidents.

Importantly, the Baltimore incident prompted a response from the company providing the service, which expressed regret over the incident and emphasised that its system was designed to identify possible threats and elevate them to human review. The need for handling consequences, from both accurate and false alarms, is something that needs to be built into the human side of the control room from an operations, public relations, credibility, and legal point of view.

The importance of human verification as part of the process is also highlighted, particularly in applications involving AI, such as perimeter protection, firearm detection, and facial recognition. Currently, AI providers probably pay less attention to this human function of using AI than they should.

Humans provide context in a way that many AI systems cannot. In areas such as perimeter protection, AI systems should be interacting with each other; however, most of these detection systems operate in parallel. Providing some integration context for the AI is likely to have a strong benefit in managing operational risk. For instance, if a system is triggered by motion detection and it is based on movement in the direction of the perimeter fence, one may give it a 30% risk rating.

Combined with object detection, which reports the shape of a person, may increase the risk rating to 50%, line crossing by the object over an electronic trip wire or boundary may boost risk to 85%, and the addition of a fence activation may heighten the risk of an incident happening to something like 95%. These are all working within certain tolerances, as movement may be triggered by a tree blowing in the wind, the object may be a cat rather than a human, line crossing may be triggered by something other than a person, as is the possibility of a fence alarm.

However, when considered in the context of all of them being activated, the context enhances the validity of the total risk notification. Unfortunately, integrating the alarms from all these AI inputs in real-world environments is sorely lacking, never mind the integrated linking of CCTV AI detection to other systems, such as voice monitoring, access control, ANPR, and facial recognition databases. At this stage, AI can provide detection using specific behaviours like crouching, crawling, lying on the ground, hands up, and wearing of PPE equipment, such as hard hats, but these remain basic and need to be easily identifiable relative to their surroundings. Human perspectives in reviewing the combination of events and CCTV footage remain critical to the effective operation of these systems.

The evolving balance between IQ and AI

So, what is the evolving balance of human operator insight and surveillance relative to system AI capabilities likely to look like?

• There are a range of CCTV surveillance areas which remain the preserve of human operators. Places in which there are people, activity, or behaviour are all done much better by people rather than AI. Human operators have a far greater ability to examine elements that can predict incidents, evaluate threat levels, identify groups of suspects, and anticipate how crisis situations will evolve. AI is likely to be used in these scenarios to augment surveillance and other information as it steadily improves and gets new capabilities.

• The nursemaid role of operators in resolving AI issues is likely to continue, although it is expected to decline over the next few years. AI, by itself, in areas such as perimeter security and object identification, is just too inefficient in providing accurate results, and without human intervention, it will compromise real threat detection and effective responses. Operators will still need to train the AI, but its success will be limited. Simply, AI suppliers will have to get their own capacity right, and specialised false alarm screener products have already shown the way. The danger in the interim is that human operators will be pulled away from core functions in order to fulfil this nursemaid role.

• Human verification will be more acknowledged across AI functions in the short term, as AI failures will highlight the response, reputational, and potential legal costs of using it without verification and structured reporting processes.

• Human operators will still largely be responsible for dispatching responses due to their capacity to verify footage and examine the situation and its components more closely. However, as the integration of multiple AI notifications provides better predictions, we will see more automated response preparation, even if this involves preparation to be picked up by an operator. AI will also provide operators with current information from available data sources, enabling them to make more informed decisions by updating live data inputs.

• AI will take over more extended functions where its strengths lie, like perimeter protection, movement in restricted areas, and simple behaviour identification. There is just too much appeal in being able to cover such broad areas without continual human coverage, but this is dependent on the development of greater accuracy and integration, as detailed below.

• AI integration of multiple detection products into a reconciled live risk analysis based on multiple criteria will increase confidence and acceptance. This, in turn, will make human verification easier, as part of the context evaluation is already being done.

We can assume that AI will become increasingly sophisticated and effective as algorithms evolve to better suit specific circumstances. Increased observation from multiple technologies is also likely to enhance the accuracy and predictability of AI findings, including more subtle signs of movement and behaviour, much like you are profiled on the internet and by instruments such as Alexa.

For the foreseeable future, people will still be more competent decision-makers. However, even now, we have situations where some operators working in control rooms do not have the insights we would want, and the gap and the need to fill it is only going to get worse. If people do not have the skills to detect crimes, evaluate situations and understand context, can they truly provide an effective role themselves, never mind confirmation of AI-sourced notifications?

Place the wrong kind of person in the human verification role, and you have a crisis waiting to happen. It places even more emphasis on what the key human contribution should be, where operators should be trained in recognising crime behaviour, situational awareness, and behavioural indicators such as types of body movements, positioning, posture, and behavioural and crime scene indicators. This is one of their key strengths, relative to AI, in delivering results from a CCTV system.



Dr Craig Donald.

About Craig Donald

Dr Craig Donald is a human factors specialist in security and CCTV. He is a director of Leaderware, which provides instruments for the selection of CCTV operators, X-ray screeners and other security personnel in major operations around the world. He also runs CCTV Surveillance Skills and Body Language, and Advanced Surveillance Body Language courses for CCTV operators, supervisors and managers internationally, and consults on CCTV management. He can be contacted on +27 11 787 7811 or craig.donald@leaderware.com


Credit(s)




Share this article:
Share via emailShare via LinkedInPrint this page



Further reading:

What is your ‘real’ security posture?
BlueVision Editor's Choice Information Security Infrastructure AI & Data Analytics
Many businesses operate under the illusion that their security controls, policies, and incident response plans will hold firm when tested by cybercriminals, but does this mean you are really safe?

Read more...
What is your ‘real’ security posture? (Part 2)
BlueVision Editor's Choice Information Security Infrastructure
In the second part of this series of articles from BlueVision, we explore the human element: social engineering and insider threats and how red teaming can expose and remedy them.

Read more...
ONVIF to end support for Profile S
News & Events Surveillance
ONVIF has announced that it will end support for ONVIF Profile S and recommends using its successor, Profile T. Profile S is the first-ever profile introduced by ONVIF in 2011.

Read more...
Onsite AI avoids cloud challenges
SMART Security Solutions Technews Publishing Editor's Choice Infrastructure AI & Data Analytics
Most AI programs today depend on constant cloud connections, which can be a liability for companies operating in secure or high-risk environments. That reliance exposes sensitive data to external networks, but also creates a single point of failure if connectivity drops.

Read more...
Toxic combinations
Editor's Choice
According to Panaseer’s latest research, 70% of major breaches are caused by toxic combinations: overlapping risks that compound and amplify each other, forming a critical vulnerability to be exploited.

Read more...
New Edge AI Plus PTZ cameras with analytics
Products & Solutions Surveillance
IDIS has unveiled two new PTZ cameras that are NDAA-compliant, delivering AI auto-tracking, rapid 40x zoom, EIS image stabilisation, and advanced automated AI functionality.

Read more...
Who has access to your face?
Access Control & Identity Management AI & Data Analytics
While you may be adjusting your privacy settings on social media or thinking twice about who is recording you at public events, the reality is that your facial features may be used in other contexts.

Read more...
The impact of AI on security
Technews Publishing Information Security AI & Data Analytics
Today’s threat actors have moved away from signature-based attacks that legacy antivirus software can detect, to ‘living-off-the-land’ using legitimate system tools to move laterally through networks. This is where AI has a critical role to play.

Read more...
Continuum launches centralised access and identity management
Editor's Choice Access Control & Identity Management Integrated Solutions Facilities & Building Management
Continuum Identity is a newly launched company in the identity management and access control sector, targeting the complexity of managing various Access and Identity Management (AIM) systems.

Read more...
Who has access to your face?
Access Control & Identity Management Residential Estate (Industry) AI & Data Analytics
While you may be adjusting your privacy settings on social media or thinking twice about who is recording you at public events, the reality is that your facial features may be used in other contexts,

Read more...










While every effort has been made to ensure the accuracy of the information contained herein, the publisher and its agents cannot be held responsible for any errors contained, or any loss incurred as a result. Articles published do not necessarily reflect the views of the publishers. The editor reserves the right to alter or cut copy. Articles submitted are deemed to have been cleared for publication. Advertisements and company contact details are published as provided by the advertiser. Technews Publishing (Pty) Ltd cannot be held responsible for the accuracy or veracity of supplied material.




© Technews Publishing (Pty) Ltd. | All Rights Reserved.