
AI Safety is uncomputable. It's Law Zero all over again
The 3 laws of robotics A robot may not injure a human being or, through inaction, allow a human being to come to harm. A robot must obey the orders given it by human beings except where such orders would conflict with the First Law. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law. The Three Laws of Robotics, conceived by Isaac Asimov, are a cornerstone of science fiction, designed to explore the complex relationship between humans and artificial intelligence. They were never intended to be practical. Their primary purpose was to entertain, to provoke thought about the potential pitfalls and ethical dilemmas inherent in creating sentient machines, often by illustrating the very ways in which the 3 laws could be circumvented or lead to unforeseen consequences. The law zero of robotics A robot may not harm humanity, or through inaction allow humanity to come to harm. That was a good idea on paper. Sounds nice. But it made th
Continue reading on Dev.to
Opens in a new tab




