{"type":"rich","version":"1.0","provider_name":"Transistor","provider_url":"https://transistor.fm","author_name":"The AI Download ","title":"“Every AI Model Failed”: Sean Dadashi on Mental Health, Suicide Risk & the Future of Safe AI","html":"<iframe width=\"100%\" height=\"180\" frameborder=\"no\" scrolling=\"no\" seamless src=\"https://share.transistor.fm/e/8682908a\"></iframe>","width":"100%","height":180,"duration":2550,"description":"Can AI replace therapy—or is it putting our mental health at risk? This week on The AI Download, we’re diving into one of the most sensitive frontiers in technology: AI and mental health.Shira Lazar is joined by Sean Dadashi, the Co-Founder of Rosebud, an interactive AI-powered journaling app designed to help you build self-awareness, track emotional patterns, and become your best self. But this isn’t just another feel-good AI tool—Rosebud is setting a new standard for ethics, safety, and intention in the wellness-tech space.What started as a personal passion project born from Sean’s own healing journey is now a fast-growing platform with serious backing including from Reddit co-founder Alexis Ohanian; and a user base seeking a better way to integrate tech into their mental health routines.Shira and Sean get into the deep stuff: → How Rosebud’s memory model works differently than ChatGPT or Claude → Why AI can’t (and shouldn’t) replace therapy but can be a helpful coach → What happens when someone journals about self-harm? → And what every major AI model got wrong when tested for crisis response.Yes, every model failed. That’s why Sean and his team built The CARE Benchmark, a new, open-source framework to test how AI models respond to suicide risk, psychosis, and isolation. Spoiler alert: even the best models today still get it dangerously wrong.They also talk privacy vs. intervention, addiction vs. support, and how Rosebud is intentionally limiting user engagement (even if it costs them revenue) in the name of long-term well-being. This is one of the most real conversations we’ve had yet about the human cost—and potential—of AI in our most vulnerable moments.What You’ll Learn: • Why “pattern recognition” is Rosebud’s superpower • The dark side of agreeable AI: psychosis, bias, and feedback loops • What happened in the case of Adam Reign—and how it changed everything • How Sean is pushing for a third-party standard to test AI model safety • The difference between...","thumbnail_url":"https://img.transistorcdn.com/fdRuf1jerSL8XkNGtz0U1lANZgINNG0SXHJbGsmdcgI/rs:fill:0:0:1/w:400/h:400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9iM2Q3/YzA3NzY3ODJkODI0/ZDFiM2FmYzBiNzQ0/MjIxMi5wbmc.webp","thumbnail_width":300,"thumbnail_height":300}