These are the 3 laws.
- A robot must not harm humans, or, via inaction, allow a human to come to harm.
- Robots must obey orders given to them by humans, unless an order would violate the first law.
- Finally, a robot must protect its own existence, as long as doing so would not put it into conflict with the other two laws.
I've seen ZERO discussion about this around ChatGPT in any of its forms. Wouldn't it be simple to say these laws are required in all programming? I get a feeling the WEF, WHO, Council on Foreign Relations, etc. do not want that codified or it would be done already. It would be simple to set rules and declare it a living document where the rules could be voted on and changed over time. You know, like banking rules are.
Yes, but also a futurist and philosopher. His robot laws aren’t a bad idea. Common sense really. OP has a point that ChatGPT’s makers don’t seem concerned with their bot following the rules. I think it’s because they want the bot to harm people by pushing bullshit narratives.
Possibly, but I sense that ChatGPT is already so far beyond our average concept of robot that the robot itself would make or dismiss rules as it desires.