Feeds

New Laws of Robotics proposed for US kill-bots

Droid-on-droid mayhem OK'd; machines to ask before snuffing humans

Remote control for virtualized desktops

A new set of laws has been proposed to govern operations by killer robots. The ideas were floated by John S Canning, an engineer at the Naval Surface Warfare Centre, Dahlgren Division – an American weapons-research and test establishment. Mr Canning's “Concept of Operations for Armed Autonomous Systems” presentation can be downloaded here (pdf).

Many Reg readers will be familiar with the old-school Asimov Laws of Robotics, but these are clearly unsuitable for war robots – too restrictive. However, the new Canning Laws are certainly not a carte blanche for homicidal droids to obliterate fleshies without limit; au contraire.

Canning proposes that robot warriors should be allowed to mix it up among themselves freely, autonomously deciding to blast enemy weapon systems. Many enemy “systems” would, of course, be themselves robots, so it's clear that machine-on-machine violence isn't a problem. The difficulty comes when the automatic battlers need to target humans. In such cases Mr Canning says that permission from a human operator should be sought.

“Let machines target other machines,” he writes, “and let men target men.”

The concept document makes the point that various kinds of automated death-tech have been allowed to destroy machinery or even people for years. He cites anti-shipping missiles which are sometimes sent off over the horizon and told to look around for a target. Other examples include automatic air-defence systems such as Phalanx or Aegis which blast anything which comes at them too fast, or the “Captor” seabed system which torpedoes passing submarines but leaves surface ships alone.

It isn't really made clear how the ask-permission-to-kill-meatsacks rule could really be applied in these cases. Doppler radar is going to have trouble distinguishing between attacking manned jets and incoming missiles, for instance. Even if the two could be swiftly and reliably differentiated, adding a human reaction and decision period in an air-defence scenario may not be a survivable thing to do.

Mr Canning also says that the emphasis should be on destroying enemy weaponry rather than people.

“We can equip our machines with non-lethal technologies for the purpose of convincing the enemy to abandon their weapons prior to our machines destroying the weapons, and lethal weapons to kill their weapons,” he suggests.

This raises the prospect of American robot enforcers packing the crowd-cookers, strobe pacifier cannons or Star Trek puke blasters already reported by El Reg, and also some conventional exploding stuff. Once enemy troops had been partially grilled, rendered epileptic or incapacitated by vomit beams, presumably fleeing as a result, the droid assailants could blow up their abandoned tanks, artillery, ships or whatnot.

Of course, this might not work so well with personal enemy weaponry such as the ubiquitous AK47 or RPG. Interestingly, though, Mr Canning quotes airforce major R Craig Burton of the Judge Advocate General's Legal Centre:

“If people or property isn't a military objective, we don't target it. It might be destroyed as collateral damage, but we don't target it. Thus in many situations, we could target the individual holding the gun and/or the gun and legally there's no difference.”

Which seems to suggest that a robot could decide, under Mr Canning's rules, to target a weapon system such as an AK47 for destruction on its own initiative, requiring no permission from a human. If the person holding it was thereby killed, that would be collateral damage and the killer droid would be in the clear. Effectively the robot is allowed to disarm enemies by prying their guns from their cold dead hands.

El Reg's advice? Do what the droids say. They are our friends. ®

Internet Security Threat Report 2014

More from The Register

next story
Antarctic ice THICKER than first feared – penguin-bot boffins
Robo-sub scans freezing waters, rocks warming models
I'll be back (and forward): Hollywood's time travel tribulations
Quick, call the Time Cops to sort out this paradox!
Your PHONE is slowly KILLING YOU
Doctors find new Digitillnesses - 'text neck' and 'telepressure'
Reuse the Force, Luke: SpaceX's Elon Musk reveals X-WING designs
And a floating carrier for recyclable rockets
NASA launches new climate model at SC14
75 days of supercomputing later ...
Britain's HUMAN DNA-strewing Moon mission rakes in £200k
3 days, and Kickstarter moves lander 37% nearer takeoff
Bond villains lament as Wicked Lasers withdraw death ray
Want to arm that shark? Better get in there quick
prev story

Whitepapers

Go beyond APM with real-time IT operations analytics
How IT operations teams can harness the wealth of wire data already flowing through their environment for real-time operational intelligence.
The total economic impact of Druva inSync
Examining the ROI enterprises may realize by implementing inSync, as they look to improve backup and recovery of endpoint data in a cost-effective manner.
Forging a new future with identity relationship management
Learn about ForgeRock's next generation IRM platform and how it is designed to empower CEOS's and enterprises to engage with consumers.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?
Mitigating web security risk with SSL certificates
Web-based systems are essential tools for running business processes and delivering services to customers.