One of the top AI apps in the local language where I live has ‘Doctor’ and ‘Therapist’ as some of its main “features” and gets gushing coverage in the press. It infuriates me every time I see mention of it anywhere.
Incidentally, telling someone to have a little meth is the least of it. There’s a much bigger issue that’s been documented where ChatGPT’s tendency to “Yes, and…” the user leads people with paranoid delusions and similar issues down some very dark paths.
Yesterday i was at a gas station and when i walked by the sandwich isle, i saw a sandwich that said: recipe made by AI. On dating apps i see a lot of girls state that they ask AI for advice. To me AI is more of a buzzword than anything else, but this shit is bananas. It,s so easy to make AI agree with everything you say.
The recipe thing is so funny to me, they try to be all unique with their recipes “made by AI”, but in reality it’s based on a slab of text that resembles the least unique recipe on the internet lol
LLMs have a use case
But they really shouldnt be used for therapy
Rly and what is their usecase? Summarizing information anf you having to check over cause its making things up? What can AI do that nothing else in the world can?
Seems it does a good job at some medical diagnosis type stuff from image recognition.
That isn’t an LLM though. That’s a different type of Machine Learning entirely.
What’s the difference? I thought they both use the same underlying technology?
A similar type of machine learning (neural networks, transformer model type thing), but I assume one is built and trained explicitly on medical records instead of scraping the internet for whatever. Correct me if I am wrong!




