72
this post was submitted on 16 Jan 2026
72 points (96.2% liked)
Technology
41298 readers
643 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 4 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
This openai up a slippery slope of requiring openai to analyze user-llm input and outputs, along with the question of privacy.
If anything, llms simply weren't ready for the open market.
E: a word
Opens? OpenAI spent years doing exactly that. Though, apparently they almost three years ago.
https://www.maginative.com/article/openai-clarifies-its-data-privacy-practices-for-api-users/
If I'm reading this right, they (claim) they are not reading user input/outputs to user, in which case they can't be held liable for results.
If we want an incomplete and immature LLM to detect the subtle signs of depression and then take action to provide therapy to guide people away, I feel we are asking too much.
At best it's like reading an interactive (and depressing) work of fiction.
Perhaps the only viable way is to train a depression detector and flag + deny function to users, which comes with its own set of problems.