One would think the programmers would have put in some safeguards that would trigger and AI would tell the user “If you’re feeling suicidal you should call 988 the National Suicide Crisis Line.”
Lots of LLMs have this, but they can be jailbroken. People are jailbreaking their sessions to have the conversations they want to have, even those who are suicidal.
Im glad they have some safeguards now. By the sound of it, ChatGPT 4o didnt.
If they put in a safeguard and people use a workaround, I dont see how the creators would be liable. Cant disable your airbags and run into a tree and blame the auto maker.
One would think the programmers would have put in some safeguards that would trigger and AI would tell the user “If you’re feeling suicidal you should call 988 the National Suicide Crisis Line.”
Lots of LLMs have this, but they can be jailbroken. People are jailbreaking their sessions to have the conversations they want to have, even those who are suicidal.
Im glad they have some safeguards now. By the sound of it, ChatGPT 4o didnt. If they put in a safeguard and people use a workaround, I dont see how the creators would be liable. Cant disable your airbags and run into a tree and blame the auto maker.