One of the central elements in last week's Trapwire story involves the application of “behavioral recognition,” also known as “video analytics,” to camera feeds. What are we to make of this technology?
In essence, video analytics is a form of artificial intelligence that tries to automatically derive meaning from a video feed. Face recognition, license plate recognition, and red light cameras are each examples of the automated extraction of meaning from a video feed, but what I’m focused on here are technologies that aim to offer more general analysis of behaviors that are taking place in a camera’s field of view. Examples include the tracking of people throughout an area, zone or perimeter protection, determination of (and detection of deviations from) “normal” patterns of movement in an area, and the detection of abandoned objects. (This article at EE Times offers an extensive introduction to the technology.)
A video camera on its own is dumb, like your retina being hit by photons. Video analytics is an attempt to create a brain behind the eye to interpret those signals. Of course we know that this kind of thing is very, very difficult; when it comes to the visual realm a computer that can analyze a stack of statistics in a flash can’t tell a toaster from a toadstool. Although, we also know that computers’ abilities in this area are rapidly improving.
Security agencies are still in the relatively early stages of experimenting with this technology. To a great extent, the civil liberties issues raised by video analytics are only an intensification of the issues raised by video surveillance itself, which we at the ACLU have long sought to limit (see our discussion of The Four Problems With Video Surveillance). After all, once you come into view of a camera, you may or may not be under observation by a human being, a roomful of human beings, or perhaps even an entire television audience (should the video later be deemed significant for some reason). By itself that creates a significant potential for abuse, and for chilling effects.
What does automated video monitoring add to those effects? A few points:
Ultimately we have to ask, where is this all going to lead? We are seeing more and more surveillance cameras installed everywhere, and increasingly they are being networked together. As artificial intelligence improves, video analytics may become capable of tracking increasingly complicated behavior. Ultimately, we need to confront the central question facing us: how are we going to handle the increasing capability of machines to monitor us in ways large and small, wide and deep? (I discussed some aspects of that question in my post on privacy invasions by humans vs. those by computers).