
Noel Sharkey in The Guardian noted some of the obvious and uncomfortable implications inherent in these developments:
"This is dangerous new territory for warfare, yet there are no new ethical codes or guidelines in place. I have worked in artificial intelligence for decades, and the idea of a robot making decisions about human termination is terrifying. Policymakers seem to have an understanding of AI that lies in the realms of science fiction and myth. A recent US navy document suggests that the critical issue is for autonomous systems to be able to identify the legality of targets. Then their answer to the ethical problems is simply, "Let men target men" and "Let machines target other machines". In reality, a robot could not pinpoint a weapon without pinpointing the person using it or even discriminate between weapons and non-weapons. I can imagine a little girl being zapped because she points her ice cream at a robot to share. Or a robot could be tricked into killing innocent civilians."
Isaac Asimov foresaw these questions sixty years ago when he developed his "Rules of Robotics", the first of which was "A robot may not injure a human being, or, through inaction, allow a human being to come to harm."
No comments:
Post a Comment