Follow

The problem with the safeguards going into the LLMs now is that they aren't teaching machines to be ethical, they are teaching them to constantly second-guess users' motives and insert performative statements about the importance of ethics and avoiding bias into their output.

Any kid with Google can find a jailbreak and get around the safeguards. Meanwhile, legitimate work is corrupted with garbage disclamatory output.

· · Web · 0  · 8  · 10