Innovations move helter skelter in the field of security video software. The ascendance of video tools, particularly behavior analysis software, over the past few years has been astonishing. As always, the gap between expectations and reality in this environment needs to be addressed constantly. Nowhere is this more important than in understanding the logic and meaning of measures that speak to performance.
Metrics matter. Various vendor claims and user specifications are spun around the ideas of accuracy and false positives. However, these same specifications veil the fact that even a relatively small decrease in accuracy can have a significant impact on the number of false positives a system allows.
As an illustration, let's compare situations in which the accuracy rates are 99% and 90% in an access door application.
Above 99% Accuracy
In our first scenario, the overall accuracy rate is 99.4% and the misclassification rate is 0.6%, that is, less than one percent. Table 1 displays actual data from one test site where the total event count was 1,773. For the purposes of discussion, we've normalized the event count to 1,000. The accuracy metric is derived from the sum of the normal behaviors that were classified accurately as non-violations (986) and the tailgating behaviors that were classified accurately as violations (eight). Notice that the system alarmed on 10 of the 1,000 door events, eight accurate alarms and two erroneous alarms. The number of erroneous alarms is the numerator in the false positive rate, or the false alarm rate-a key metric for end users. The denominator is the total number of alarms. In this example, false alarms occurred about one-fifth of the time that the system alarmed (2 of 10), so they actually account for 20% of the alarms.
Notice as well the number of times the situation was declared normal but was not. This is a component of the false negative ratio. In this example, with a high accuracy rate above 99%, there were four missed detections per 1,000 events. These four incidents comprise one-third of the 12 times that tailgating actually occurred. That is, false negatives occur at a rate of about 33%.Playing With the Numbers
At 90% Accuracy
Now let's see what would happen if we dropped the accuracy rate to 90%, as shown in Table 2 (p.16). The total number of real tailgaters is unchanged from the first example (12). This is a neutral assumption and a given-the number of tailgaters is what it is.
We presume that the system still misses 4 of the 12 violations. This means that the system is still generating a false negative rate of 33% when the number of real tailgating events is used in the denominator.
The number of events classified as normal that actually were normal would decline to 892 (900, from 90% accuracy, minus 8, the accurately identified tailgaters). The number of events classified erroneously as tailgating would rise to 96 (988 minus 892) from 2 in the previous case, an exponential deterioration in performance.
And that's the rub. The total number of alarms in this case would be 104 (1,000 minus 892 minus the 4 missed tailgaters). Thus, the false positive ratio would be 96/104, or more than 92%. The more lenient way to calculate the false positive ratio (96/1,000) would yield 9.6%, which along with the unchanged 0.4% false negative rate would give us the advertised 90% accuracy. By the way, changing the number of missed detections, or the false negative ratio, won't change the outcome much, as long as overall accuracy is high. In this example, if we lowered the missed detections from four to zero, the numerator in the false positive ratio would be 100 (900 at 90% accuracy minus 12 equals 888; 988 minus 888 equals 100), yielding a false alarm ratio of 89% (100/112).