New research looks at how three large language models handle queries of varying riskiness on suicide amid rising mental health crisis, shortage of care.
ELIZA worked just as well as long as you had enough IF-THEN statements to cover every possible turn of a conversation. But you can’t, and just like ELIZA but in a more complex way, LLMs will miss an unexpected turn if they can’t predict it. And just about every LLM I’ve seen will end up agreeing with you, which is not at all great for someone getting mental health.
ELIZA worked just as well as long as you had enough IF-THEN statements to cover every possible turn of a conversation. But you can’t, and just like ELIZA but in a more complex way, LLMs will miss an unexpected turn if they can’t predict it. And just about every LLM I’ve seen will end up agreeing with you, which is not at all great for someone getting mental health.
Yeah. All the API-only big models are now and forever dead in the water for this, no matter how much they improve. The sycophancy is unreal.