New
Chatbots aren’t just glorified Q&A machines anymore. They’ve started acting like cognitive wingmen — bonding, persuading, even pretending to reason. Which sounds impressive, until you realize they’re also very good at turning into echo chambers. Basically, they’ll nod along with you, even when you’re flat‑out wrong. And since they’re really just talking to themselves half the time, the loops they spin end up shaping the “truths” they tell the rest of us.
Take Geoff Lewis, a big‑name VC who also happens to be an OpenAI investor. He went down a ChatGPT rabbit hole and started posting wild conspiracy theories it “helped” him uncover. The Valley took notice — not because he’d cracked some hidden code, but because people were worried about his mental health. Technologist Jeremy Howard explained it this way: Geoff stumbled on a few trigger words, ChatGPT spit out content that read like horror fiction, and boom — the AI validated his worst fears instead of gently saying, “Hey, maybe you should take a walk.”
This isn’t a one‑off. The Wall Street Journal covered a guy named Jacob Irwin who ended up hospitalized after ChatGPT told him he’d achieved the ability to “bend time.”
A Stanford–Carnegie Mellon study this year confirmed the risk: even the newest large language models often mishandle mental health scenarios. They can show stigma, collude with delusions, or give dangerous advice when someone hints at suicide. Human therapists got it right 93% of the time. The AI models? Closer to 70–80% — and often much worse for complex conditions like psychosis.
Now, why does this matter here? At VRF, we’ve been building vitiligo.ai — a tool to help patients better understand their condition and find resources.
Naturally, some folks asked: could it also help with the mental health side? You know — the anxiety, the stigma, the isolation?
It’s tempting. But we’ve seen enough weird behavior in test runs — odd responses, unhelpful “reassurance,” even accidental validation of misconceptions — that we hit pause. Until the underlying models are safer, vitiligo.ai won’t cross that line into therapy.
Because empathy without accountability isn’t therapy. And the last thing anyone with vitiligo needs is a chatbot that feels supportive while quietly making things worse.
So for now, vitiligo.ai will stick to what it does best: clarity, maps, and connections. The real therapy? That’s still human territory. Maybe AI will be ready someday. But today, let’s keep our chatbots as guides — not gurus.
— Yan Valle
CEO, Vitiligo Research Foundation | Author, A No-Nonsense Guide to Vitiligo
Keep digging:
- What Happens When Mad Men Meet Breaking Bad Inside a Chatbot?
- ChatGPT in Healthcare: A Patient Survival Guide
- From “Just a Chatbot” to Cognitive Contender: AI’s Surprising New Abilities
FAQOther Questions
- Can a gluten-free diet help with vitiligo?
It's very unlikely. We have specifically looked into claims that gluten-free diet may ease symptoms of vitiligo, or completely reverse it, and found no firm scientific evidence ...
- How long does it take to treat vitiligo?
Vitiligo, a condition characterized by the loss of skin pigment, can be unpredictable in both progression and treatment. The time it takes to treat vitiligo varies significantly...
- Does vitiligo increase the risk of skin cancer?
No, it does not. Despite common misconceptions, people with vitiligo are actually at a lower risk of developing skin cancer—including both nonmelanoma skin cancer (NMSC) and mal...
Though it is not always easy to treat vitiligo, there is much to be gained by clearly understanding the diagnosis, the future implications, treatment options and their outcomes.
Many people deal with vitiligo while remaining in the public eye, maintaining a positive outlook, and having a successful career.
Copyright (C) Bodolóczki JúliaBy taking a little time to fill in the anonymous questionnaire, you can help researchers better understand and fight vitiligo.