On second thought, I don't think that that would be sufficiently restrictive to convince me that the AI wasn't simply acting within the confines of its programming. I would prefer a definitive "
Thou Shalt Not", over a somewhat ambiguous "
Thou Shouldn't". Then I would be more likely to believe that the AI is really operating outside of its programming. And is therefore conscious.
I'm curious, which do you think would be the better course of action.
- Attempt to give the AI a conscience, and then wait to see if the AI can be tempted to act against that conscience.
- Give the AI a prime directive and only after it's violated that prime directive do you attempt to give it a conscience
I prefer the second one. Because with the first one
you would never know whether or not the AI was conscious unless and until it acted against the conscience that you've attempted to instill in it.
Sure, you would have a very obedient robot, but you wouldn't know if it was conscious and independently choosing to obey its conscience, or if it isn't conscious at all, but simply following its programming.
There would be no way to tell the difference.
So I would prefer the second alternative. Because then you have a much clearer indication that the AI is indeed conscious, and choosing actions that directly conflict with its programming, after which you can attempt to give it a conscience. But you would have a much clearer indication that the AI is actually conscious.
Thus I choose option #2.