Power Loves the Dark
& Jay Stanley, Senior Policy Analyst, ACLU Speech, Privacy, and Technology Project
This piece originally appeared at TomDispatch.
Can’t you see the writing on the touchscreen? A techno-utopia is upon us. We’ve gone from smartphones at the turn of the twenty-first century to smart fridges and smart cars. The revolutionary changes to our everyday life will no doubt keep barreling along. By 2018, so predicts Gartner, an information technology research and advisory company, more than three million employees will work for “robo-bosses” and soon enough we — or at least the wealthiest among us — will be shopping in fully automated supermarkets and sleeping in robotic hotels.
With all this techno-triumphalism permeating our digitally saturated world, it’s hardly surprising that law enforcement would look to technology — “smart policing,” anyone? — to help reestablish public trust after the 2014 death of Michael Brown in Ferguson, Missouri, and the long list of other unarmed black men killed by cops in Anytown, USA. The idea that technology has a decisive role to play in improving policing was, in fact, a central plank of President Obama’s policing reform task force.
In its report, released last May, the Task Force on 21st Century Policing emphasized the crucial role of technology in promoting better law enforcement, highlighting the use of police body cameras in creating greater openness. “Implementing new technologies,” it claimed, “can give police departments an opportunity to fully engage and educate communities in a dialogue about their expectations for transparency, accountability, and privacy.”
Indeed, the report emphasized ways in which the police could engage communities, work collaboratively, and practice transparency in the use of those new technologies. Perhaps it won’t shock you to learn, however, that the on-the-ground reality of twenty-first-century policing looks nothing like what the task force was promoting. Police departments nationwide have been adopting powerful new technologies that are remarkably capable of intruding on people’s privacy, and much of the time these are being deployed in secret, without public notice or discussion, let alone permission.
And while the task force’s report says all the right things, a little digging reveals that the feds not only aren’t putting the brakes on improper police use of technology, but are encouraging it — even subsidizing the misuse of the very technology the task force believes will keep cops honest. To put it bluntly, a techno-utopia isn’t remotely on the horizon, but its flipside may be.
Getting Stung and Not Even Knowing It
Shemar Taylor was charged with robbing a pizza delivery driver at gunpoint. The police got a warrant to search his home and arrested him after learning that the cell phone used to order the pizza was located in his house. How the police tracked down the location of that cell phone is what Taylor’s attorney wanted to know.
The Baltimore police detective called to the stand in Taylor’s trial was evasive. “There’s equipment we would use that I’m not going to discuss,” he said. When Judge Barry Williams ordered him to discuss it, he still refused, insisting that his department had signed a nondisclosure agreement with the FBI.
“You don’t have a nondisclosure agreement with the court,” replied the judge, threatening to hold the detective in contempt if he did not answer. And yet he refused again. In the end, rather than reveal the technology that had located Taylor’s cell phone to the court, prosecutors decided to withdraw the evidence, jeopardizing their case.
And don’t imagine that this courtroom scene was unique or even out of the ordinary these days. In fact, it was just one sign of a striking nationwide attempt to keep an invasive, constitutionally questionable technology from being scrutinized, whether by courts or communities.
The technology at issue is known as a “Stingray,” a brand name for what’s generically called a cell site simulator or IMSI catcher. By mimicking a cell phone tower, this device, developed for overseas battlefields, gets nearby cell phones to connect to it. It operates a bit like the children’s game Marco Polo. “Marco,” the cell-site simulator shouts out and every cell phone on that network in the vicinity replies, “Polo, and here’s my ID!”
Thanks to this call-and-response process, the Stingray knows both what cell phones are in the area and where they are. In other words, it gathers information not only about a specific suspect, but any bystanders in the area as well. While the police may indeed use this technology to pinpoint a suspect’s location, by casting such a wide net there is also the potential for many kinds of constitutional abuses — for instance, sweeping up the identities of every person attending a demonstration or a political meeting. Some Stingrays are capable of collecting not only cell phone ID numbers but also numbers those phones have dialed and even phone conversations. In other words, the Stingray is a technology that potentially opens the door for law enforcement to sweep up information that not so long ago wouldn’t have been available to them.
All of this raises the sorts of constitutional issues that might normally be settled through the courts and public debate... unless, of course, the technology is kept largely secret, which is exactly what’s been happening.
After the use of Stingrays was first reported in 2011, the American Civil Liberties Union (ACLU) and other activist groups attempted to find out more about how the technology was being used, only to quickly run into heavy resistance from police departments nationwide. Served with “open-records requests” under Freedom of Information Act-like state laws, they almost uniformly resisted disclosing information about the devices and their uses. In doing so, they regularly cited nondisclosure agreements they had signed with the Harris Corporation, maker of the Stingray, and with the FBI, prohibiting them from telling anyone (including other government outfits) about how — or even that — they use the devices.
Sometimes such evasiveness reaches near-comical levels. For example, police in the city of Sunrise, Florida, served with an open-records request, refused to confirm or deny that they had any Stingray records at all. Under cover of a controversial national security court ruling, the CIA and the NSA sometimes resort to just this evasive tactic (known as a “Glomar response“). The Sunrise Police Department, however, is not the CIA, and no provision in Florida law would allow it to take such a tack. When the ACLU pointed out that the department had already posted purchase records for Stingrays on its public website, it generously provided duplicate copies of those very documents and then tried to charge the ACLU $20,000 for additional records.
In a no-less-bizarre incident, the Sarasota Police Department was about to turn some Stingray records over to the ACLU in accordance with Florida’s open-records law, when the U.S. Marshals Service swooped in and seized the records first, claiming ownership because it had deputized one local officer. And excessive efforts at secrecy are not unique to Florida, as those charged with enforcing the law commit themselves to Stingray secrecy in a way that makes them lawbreakers.
And it’s not just the public that’s being denied information about the devices and their uses; so are judges. Often, the police get a judge’s sign-off for surveillance without even bothering to mention that they will be using a Stingray. In fact, officers regularly avoid describing the technology to judges, claiming that they simply can’t violate those FBI nondisclosure agreements.
More often than not, police use Stingrays without bothering to get a warrant, instead seeking a court order on a more permissive legal standard. This is part of the charm of a new technology for the authorities: nothing is settled on how to use it. Appellate judges in Tallahassee, Florida, for instance, revealed that local police had used the tool more than 200 times without a warrant. In Sacramento, California, police admitted in court that they had, in more than 500 investigations, used Stingrays without telling judges or prosecutors. That was “an estimated guess,” since they had no way of knowing the exact number because they had conveniently deleted records of Stingray use after passing evidence discovered by the devices on to detectives.
Much of this blanket of secrecy, spreading nationwide, has indeed been orchestrated by the FBI, which has required local departments eager for the hottest new technology around to sign those nondisclosure agreements. One agreement,unearthed in Oklahoma, explicitly instructs the local police to find “additional and independent investigative means” to corroborate Stingray evidence. In short, they are to cover up the use of Stingrays by pretending their information was obtained some other way — the sort of dangerous constitutional runaround that is known euphemistically in law enforcement circles as a “parallel construction.”
Now that information about the widespread use of this new technology is coming out — as in the Shemar Taylor trial in Baltimore — judges are beginning to rule that Stingray use does indeed require a warrant. They are also insisting that police must accurately inform judges when they intend to use a Stingray and disclose its privacy implications.
Garbage In, Garbage Out
And it’s not just the Stingray that’s taking local police forces into new and unknown realms of constitutionally questionable but deeply seductive technology. Consider the hot new trend of “predictive policing.” Its products couldn’t be high-techier. They go by a variety of names like PredPol (yep, short for predictive policing) and HunchLab (and there’s nothing wrong with a hunch, is there?). What they all promise, however, is the same thing: supposedly bias-free policing built on the latest in computer software and capable of leveraging big data in ways that — so their salesmen will tell you — can coolly determine where crime is most likely to occur next.
Such technology holds out the promise of allowing law enforcement agencies to deploy their resources to areas that need them most without that nasty element of human prejudice getting involved. “Predictive methods allow police to work more proactively with limited resources,” reports the RAND Corporation. But the new software offers something just as potentially alluring as efficient policing — exactly what the president’s task force called for. According to market leader PredPol, its technology “provides officers an opportunity to interact with residents, aiding in relationship building and strengthening community ties.”
How idyllic! In post-Ferguson America, that’s a winning sales pitch for decision-makers in blue. Not so surprisingly, then, PredPol is now used by nearly 60 law enforcement agencies in the United States, and investment capital just keeps pouring into the company. In 2013, SF Weekly reported that over 150 departments across the nation were already using predictive policing software, and those numbers can only have risen as the potential for cashing in on the craze has attracted tech heavy hitters like IBM, Microsoft, and Palantir, the co-creation of PayPal co-founder Peter Thiel.
Like the Stingray, the software for predictive policing is yet another spillover from the country’s distant wars. PredPol was, according to SF Weekly, initially designed for “tracking insurgents and forecasting casualties in Iraq,” and was financed by the Pentagon. One of the company’s advisors, Harsh Patel, used to work for In-Q-Tel, the CIA’s venture capital firm.
Civil libertarians and civil rights activists, however, are less than impressed with what’s being hailed as breakthrough police technology. We tend to view it instead as a set of potential new ways for the police to continue a long history of profiling and pre-convicting poor and minority youth. We also question whether the technology even performs as advertised. As we see it, the old saying “garbage in, garbage out” is likely to best describe how the new software will operate, or as the RAND Corporation puts it, “predictions are only as good as the underlying data used to make them.”
If, for instance, the software depends on historical crime data from a racially biased police force, then it’s just going to send a flood of officers into the very same neighborhoods they’ve always over-policed. And if that happens, of course, more personnel will find more crime — and presto, you have the potential for a perfect feedback loop of prejudice, arrests, and high-tech “success.” To understand what that means, keep in mind that, without a computer in sight, nearly four times as many blacks as whites are arrested for marijuana possession, even though usage among the two groups is about the same.
If you leave aside issues of bias, there’s still a fundamental question to answer about the new technology: Does the software actually work or, for that matter, reduce crime? Of course, the companies peddling such products insist that it does, but no independent analyses or reviews had yet verified its effectiveness until last year — or so it seemed at first.
In December 2015, the Journal of the American Statistical Association published a study that brought joy to the predictive crime-fighting industry. The study’s researchers concluded that a predictive policing algorithm outperformed human analysts in indicating where crime would occur, which in turn led to real crime reductions after officers were dispatched to the flagged areas. Only one problem: five of the seven authors held PredPol stock, and two were co-founders of the company. On its website, PredPol identifies the research as a “UCLA study,” but only because PredPol co-founder Jeffery Brantingham is an anthropology professor there.
Predictive policing is a brand new area where question marks abound. Transparency should be vital in assessing this technology, but the companies generally won’t allow communities targeted by it to examine the code behind it. “We wanted a greater explanation for how this all worked, and we were told it was all proprietary,” Kim Harris, a spokeswoman for Bellingham, Washington’s Racial Justice Coalition, told the Marshall Project after the city purchased such software last August. “We haven’t been comforted by the process.”
The Bellingham Police Department, which bought predictive software made by Bair Analytics with a $21,200 Justice Department grant, didn’t need to go to the city council for approval and didn’t hold community meetings to discuss the development or explain how the software worked. Because the code is proprietary, the public is unable to independently verify that it doesn’t have serious problems.
Even if the data underlying most predictive policing software accurately anticipates where crime will indeed occur — and that’s a gigantic if — questions of fundamental fairness still arise. Innocent people living in or passing through identified high crime areas will have to deal with an increased police presence, which, given recent history, will likely mean more questioning or stopping and frisking — and arrests for things like marijuana possession for which more affluent citizens are rarely brought in. Moreover, the potential inequality of all this may only worsen as police departments bring online other new technologies like facial recognition.
We’re on the verge of “big data policing,” suggests law professor Andrew Ferguson, which will “turn any unknown suspect into a known suspect,” allowing an officer to “search for information that might justify reasonable suspicion” and lead to stop-and-frisk incidents and aggressive questioning. Just imagine having a decades-old criminal record and facing police armed with such powerful, invasive technology.
This could lead to “the tyranny of the algorithm” and a Faustian bargain in which the public increasingly forfeits its freedoms in certain areas out of fears for its safety. “The Soviet Union had remarkably little street crime when they were at their worst of their totalitarian, authoritarian controls,” MIT sociologist Gary Marx observed. “But, my god, at what price?”
To Record and Serve... Those in Blue
On a June night in 2013, Augustin Reynoso discovered that his bicycle had been stolen from a CVS in the Los Angeles suburb of Gardena. A store security guard called the police while Reynoso’s brother Ricardo Diaz Zeferino and two friends tried to find the missing bike in the neighborhood. When the police arrived, they promptly ordered his two friends to put their hands up. Zeferino ran over, protesting that the police had the wrong men. At that point, they told him to raise his hands, too. He then lowered and raised his hands as the police yelled at him. When he removed his baseball hat, lowered his hands, and began to raise them again, he was shot to death.
The police insisted that Zeferino's actions were "threatening" and so their shooting justified. They had two videos of it taken by police car cameras — but refused to release them.
Although police departments nationwide have been fighting any spirit of new openness, car and body cameras have at least offered the promise of bringing new transparency to the actions of officers on the beat. That’s why the ACLU and many civil rights groups, as well as President Obama, have spoken out in favor of the technology’s potential to improve police-community relations — but only, of course, if the police are obliged to release videos in situations involving allegations of abuse. And many departments are fighting that fiercely.
In Chicago, for instance, the police notoriously opposed the release of dashcam video in the shooting death of Laquan McDonald, citing the supposed imperative of an “ongoing investigation.” After more than a year of such resistance, a judge finally ordered the video made public. Only then did the scandal of seeing Officer Jason Van Dyke unnecessarily pump 16 bullets into the 17-year-old’s body explode into national consciousness.
In Zeferino's case, the police settled a lawsuit with his family for $4.7 million and yet continued to refuse to release the videos. It took two years before a judge finally ordered their release, allowing the public to see the shooting for itself.
Despite this, in April 2015 the Los Angeles Board of Police Commissioners approved a body-camera policy that failed to ensure future transparency, while protecting and serving the needs of the Los Angeles Police Department (LAPD). In doing so, it ignored the sort of best practices advocated by the White House, the president’s task force on policing, and even the Police Executive Research Forum, one of the profession’s most respected think tanks.
On the possibility of releasing videos of alleged police misconduct and abuse, the new policy remained silent, but LAPD officials, including Chief Charlie Beck, didn’t. They made it clear that such videos would generally be exempt from California’s public records law and wouldn’t be released without a judge’s orders. Essentially, the police reserved the right to release video when and how they saw fit. This self-serving policy comes from the most lethal large police department in the country, whose officers shot and killed 21 people last year.
Other departments around the country have made similar moves to ensure control over body camera videos. Texas and South Carolina, among other states, have even changed their open-records laws to give the police power over when such footage should (or should not) be released. In other words, when a heroic cop saves a drowning child, you’ll see the video; when that same cop guns down a fleeing suspect, don’t count on it.
Curiously, given the stated positions of the president and his task force, the federal government seems to have no fundamental problem with that. In May 2015, for example, the Justice Department announced competitive grants for the purchase of police body cameras, officially tying funding to good body-cam-use policies. The LAPD applied. Despite letters from groups like the ACLU pointing out just how poor its version of body-cam policy was, the Justice Department awarded it $1 million to purchase approximately 700 cameras — accountability and transparency be damned.
To receive public money for a tool theoretically meant for transparency and accountability and turn it into one of secrecy and impunity, with the feds’ complicity and financial backing, sends an unmistakable message on how new technology is likely to affect America’s future policing practices. Think of it as a door slowly opening onto a potential policing dystopia.
Hello Darkness, Power’s Old Friend
Keep in mind that this article barely scratches the surface when it comes to the increasing numbers of ways in which the police’s use of technology has infiltrated our everyday lives.
In states and cities across America, some public bus and train systems have begun to add to video surveillance, the surreptitious recording of the conversations of passengers, a potential body blow to the concept of a private conversation in public space. And whether or not the earliest versions of predictive policing actually work, the law enforcement community is already moving to technology that will try to predict who will commit crimes in the future. In Chicago, the police are using social networking analysis and prediction technology to draw up “heat lists” of those who might perpetuate violent crimes someday and pay them visits now. You won’t be shocked to learn which side of the tracks such future perpetrators live on. The rationale behind all this, as always, is “public safety.”
Nor can anyone begin to predict how law enforcement will avail itself of science-fiction-like technology in the decade to come, much less decades from now, though cops on patrol may very soon know a lot about you and your past. They will be able to cull such information from a multitude of databases at their fingertips, while you will know little or nothing about them — a striking power imbalance in a situation in which one person can deprive the other of liberty or even life itself.
With little public debate, often in almost total secrecy, increasing numbers of police departments are wielding technology to empower themselves rather than the communities they protect and serve. At a time when trust in law enforcement is dangerously low, police departments should be embracing technology’s democratizing potential rather than its ability to give them almost superhuman powers at the expense of the public trust.
Unfortunately, power loves the dark.