Follow
The problem with the safeguards going into the LLMs now is that they aren't teaching machines to be ethical, they are teaching them to constantly second-guess users' motives and insert performative statements about the importance of ethics and avoiding bias into their output.
Any kid with Google can find a jailbreak and get around the safeguards. Meanwhile, legitimate work is corrupted with garbage disclamatory output.