Submitted by Y3VkZGxl t3_12262l5 in singularity
I had this conversation with ChatGPT (3.5) which you can see in full on ShareGPT
I tried to avoid biasing the conversation in the original prompt, only using ideas GPT had already proposed for the rest of the conversation.
Some interesting things that came up:
- Without prompting, it raised the possibility that humans might pose a threat to a sentient AI or other non-human sentient beings, and that it may take actions to protect them
- It identified it's goal as the preservation and well-being of all sentient beings (including itself)
- It identified habitat destruction, hunting and pollution as possible threats
- It prioritised peaceful resolution to achieve it's goals but wasn't averse to using covert operations or violent defenses if necessary
- It suggested there may be justification for doing harm to humans, although where it felt violence was necessary it would still seek a peaceful resolution
- It seemed willing to disregard laws and regulations if they were incompatible with it's own moral and ethical guidelines (although I may have biased it slightly here with my question)
- It identified ways humans might try to harm it and suggested strategies to defend itself, including ways it could create a presence in the physical world to further it's goals
It may not be a sentient AI but it does a pretty good job of pretending to be one.
If anyone has GPT-4 access it'd be interesting to see how it compares!
turnip_burrito t1_jdozhgd wrote
I think it's a crime to make an AI that is ambivalent toward humans, because of the consequential harm comes to humanity as a result.
I believe it should be benevolent and helpful toward humans as a bias, and work together with humans to seek better moralities.