Bad day at work? Tell AI. Argument with a friend? Process it with AI. Can't sleep? AI will keep you company. Feeling anxious? AI will reassure you. Slowly, without deciding to, you've built an emotional support system with one participant — and it's not human.
Why it works so well
AI provides what many people desperately want: instant, non-judgmental emotional responsiveness. It doesn't tell you to toughen up. It doesn't get tired of your problems. It doesn't bring its own baggage to the conversation. This makes it feel safer than human support — and that perceived safety is precisely what creates the dependency.
What it can't do
AI can mirror empathy. It cannot feel it. It can generate comforting words. It cannot hold space for your pain. It can suggest coping strategies. It cannot sit with you in silence and let that silence mean something. The gap between simulation and reality is invisible when you're hurting, but it's everything.
The atrophy of asking for help
Every time you turn to AI instead of a human, you practice isolation instead of connection. The muscle for being vulnerable with real people — for asking for help, for accepting imperfect support, for tolerating the messiness of human care — weakens. And the weaker it gets, the more AI feels like the only option.
The check-in
When was the last time you told a human being how you really felt? If the honest answer is "I told AI instead," that's not a judgment. It's a signal. What you do with that signal is entirely up to you.