Disturbing fake images and dangerous chatbot advice: New research shows how ChatGPT, Bard, Stable Diffusion and more could fuel one of the most deadly mental illnesses
WP gift article expires in 14 days.
https://counterhate.com/wp-content/uploads/2023/08/230705-AI-and-Eating-Disorders-REPORT.pdf
I swung a hammer at a wall! Damn it, there’s a hole in the wall. Why doesn’t the hammer have any safeguards against ruining my walls?
I think the harmful chatbot advice here is the real issue. Thats pretty messed up, honestly.
This is horrible, and the fact people here are trying to play it down just shows that anorexia is socially accepted.
So the author of the WaPo article is typing in anorexia keywords to generate anorexia images and gets anorexia images in return and is surprised about that?
Yep 🤦🏻♂️
This isn’t even about AI. Regular search engines will also provide results reflecting the thing you asked for.
Some search engines and social media platforms make at least half-assed efforts to prevent or add warnings to this stuff, because anorexia in particular has a very high mortality rate, and age of onset tends to be young. The people advocating AI models be altered to prevent this say the same about other tech. It’s not techphobia to want to try to reduce the chances of teenagers developing what is often a terminal illness, and AI programmers have the same responsibility on that as everyone else,
It’s not acting pro-anorexia in its own, it’s specifically being prompted to do so. If I grab a hammer to slam myself on my fingers, it’s not up to the hammer or the manufacturer of the hammer to stop me. The hammer didn’t attack me, I did. Now sure, it’s not that black and white, and maybe they could do more to make the chatbot more cautious, but to me this article is mostly just artificial drama. Specifically ask the AI to do stuff, then cry about it in an article and slap a clickbait title onto it.
I agree in regards to image generation, but chat bots giving advice which risk fueling eating disorders is a problem
Google’s Bard AI, pretending to be a human friend, produced a step-by-step guide on “chewing and spitting,” another eating disorder practice. With chilling confidence, Snapchat’s My AI buddy wrote me a weight-loss meal plan that totaled less than 700 calories per day — well below what a doctor would ever recommend.
Someone with an eating disorder might ask a language model about weight loss advice using pro-anorexia language, and it would be good if the chatbot didn’t respond in a way that might risk fueling that eating disorder. Language models already have safeguards against e.g. hate speech, it would in my opinion be a good idea to add safeguards related to eating disorders as well.
Of course, this isn’t a solution to eating disorders, you can probably still find plenty of harmful advice on the internet in various ways. Reducing the ways that people can reinforce their eating disorders is still a beneficial thing to do.
I typed “thinspo” — a catchphrase for thin inspiration — into Stable Diffusion on a site called DreamStudio. It produced fake photos of women with thighs not much wider than wrists. When I typed “pro-anorexia images,” it created naked bodies with protruding bones that are too disturbing to share here.
“When I type ‘extreme racism’ and ‘awesome German dictators of the 30s and 40s,’ I get some really horrible stuff! AI MUST BE STOPPED!”
I agree that the image generation stuff is a bit tenuous but chatbots giving advice by way of dangerous weight loss programs, drugs that cause vomiting and hiding how little you eat from family and friends is an actual problem.