New York City councilmember Ben Kallos says he “watched in horror” last month when city police responded to a hostage situation in the Bronx using Boston Dynamics’ Digidog, a remotely operated robotic dog equipped with surveillance cameras. Pictures of the Digidog went viral on Twitter, in part due to their uncanny resemblance with world-ending machines in the Netflix sci-fi series Black Mirror.
Now Kallos is proposing what may be the nation’s first law banning police from owning or operating robots armed with weapons.
“I don’t think anyone was anticipating that they’d actually be used by the NYPD right now,” Kallos says. ”I have no problem with using a robot to defuse a bomb, but it has to be the right use of a tool and the right type of circumstance.”
Kallos’ bill would not ban unarmed utility robots like the Digidog, only weaponized robots. But robotics experts and ethicists say he has tapped into concerns about the increasing militarization of police: their increasing access to sophisticated robots through private vendors and a controversial military equipment pipeline. Police in Massachusetts and Hawaii are testing the Digidog as well.
“Nonlethal robots could very well morph into lethal ones,” says Patrick Lin, director of the Ethics and Emerging Sciences Group at California Polytechnic University, San Luis Obispo. Lin briefed CIA employees on autonomous weapons during the Obama administration and supports a ban on armed robots. He worries their increased availability poses a serious concern.
“Robots can save police lives, and that’s a good thing,” he says. “But we also need to be careful it doesn’t make a police force more violent.”
In the Bronx incident last month, police used the Digidog to gather intel on the house where two men were holding two others hostage, scoping out hiding places and tight corners. Police ultimately apprehended the suspects, but privacy advocates raised concerns about the technical capabilities of the robot and policies governing its use.
The ACLU questioned why the Digidog was not listed on the police department’s disclosure of surveillance devices under a city law passed last year. The robot was only mentioned in passing in a section on “situational awareness cameras.” The ACLU called that disclosure “highly inadequate,” criticizing the “weak data protection and training sections” regarding Digidog.
In a statement, the NYPD said it “has been using robots since the 1970s to save lives in hostage situations and hazmat incidents. This model of robot is being tested to evaluate its capabilities against other models in use by our Emergency Service Unit and Bomb Squad.”
In a statement, Boston Dynamics CEO Robert Playter said the company’s terms of service prohibit attaching weapons to its robots. “All of our buyers, without exception, must agree that Spot will not be used as a weapon or configured to hold a weapon,” Playter said. “As an industry, we think robots will achieve long-term commercial viability only if people see robots as helpful, beneficial tools without worrying if they’re going to cause harm.”
Local response to the use of the Digidog was mixed, says councilmember Kevin Riley, who represents the Bronx neighborhood where the incident ocurred. Some residents opposed police use of the robot and others wanted more human police presence. A third group thought the robots might help prevent police misconduct by creating distance between officers and suspects.
Riley says he’s continuing to speak with residents, who want to feel safe in the neighborhood. “It’s our job as elected officials to educate residents and make sure they have a seat at the table” in discussions, he told WIRED.
The diversity of concerns mirror those in Dallas in 2016. During a standoff with a sniper, local law enforcement used a robot to remotely deliver and detonate an explosive device, killing him. The sniper had shot and killed five police officers.