UC Berkeley School of Public Health Professor Jodi Halpern has spent years working on the ethics of innovative technologies like gene editing and artificial intelligence. But lately Halpern, a psychiatrist, has been focusing on the expanding use of artificial intelligence (AI) in mental health.

In the past few years, dozens of businesses in health care and technology have launched apps which they claim can assist in diagnosing mental health conditions and complement—or even replace—individual therapy.

They range from apps that purport to help patients monitor and manage their moods, to programs that provide social support and clinical care. At a time when there’s a nationwide shortage of therapists, can AI fill the gap?

Advertisement X

Dr. Halpern is co-leader of Berkeley Group for the Ethics and Regulation of Innovative Technologies (BERGIT), and the co-founder of the Kavli Center for Ethics, Science and the Public, a multidisciplinary group which seeks to provide a democratic framework for understanding the ethical implications of science and technology. We asked Dr. Halpern to walk us through the pros and cons of using AI to provide mental health care.

Berkeley Public Health: How would you describe artificial intelligence to someone coming out of a 20-year coma?

Jodi Halpern: You could say it uses statistical and other models to create pattern recognition programs that are novel but can simulate human behavior, decisions, judgments, etc.

The artificial intelligence reasoning processes are not the same as what humans do, but as we see with large language models, can simulate human behavior.

BPH: Why is there so much excitement about using AI in mental health?

JH: The excitement is partly because we’re in a mental health crisis. Depending on what study you look at, 26% of Americans have an actual mental health diagnosis. So, that’s a lot of folks.

And then we know that beyond that, there is a crisis of extreme loneliness. Some studies have reported that as high as 50% of Americans in different subgroups—like adolescents and women with young children—suffer from extreme loneliness. So you have people with unmet mental health and other needs; and we have, in general, underfunded access to mental health.

So, any system that can offer certain kinds of mental health resources is something to be taken seriously as a potential benefit.

BPH: But you do have concerns?

JH: Yes. First, we don’t even know how widespread the use of “AI companions” for people with mental health needs is. I don’t think there are good statistics yet about which companies are doing it and how many users they have.

My concern is with marketing bots as therapists and trusted companions to people who are depressed or otherwise highly vulnerable.

In contrast, there are a lot of different uses in the mental health sphere beyond therapy bots. There are mindfulness apps and cognitive behavioral therapy apps that do not simulate relationships that have millions of users. And then there are actual health systems in the UK and several in the US that are starting to use AI for some medical record-keeping to reduce administrative burdens on mental health providers.

Jodi Halpern MD, PhD, is Professor of Bioethics and Medical Humanities in the UCB-UCSF Joint Medical Program.

BPH: How do you feel about AI for record-keeping?

JH: Taking over some electronic medical records, and other administrative tasks with AI is very promising.

We have a huge burnout crisis in medicine in general. Sixty-one percent of physicians and about the same number of nurses say they are burned out. And that is a huge problem because they are not proving the kind of empathetic and attentive care that patients need.

When we see our doctors, they have to spend the whole time recording electronic medical records, which means they can’t even look at us or make eye contact or be with us, human to human. To me, it’s extremely promising to use AI to take over the administrative tasks and electronic medical records.

BPH: What else seems promising?

JH: Right now, the British National Health Service is using an app to listen in while a person is screening a patient for their health needs. That’s also being deployed now in certain health systems in the US. The idea is that the application will help detect whether there is something that the patient says that the provider missed, but which might indicate something to be concerned about, regarding mental health issues like serious depression or evidence of suicidality, things like that.

I think this is a useful assistant during the screening, but I wouldn’t want to see that used absent any human contact just because it saves money. People with mental health needs are often reluctant to seek care and making an actual human connection can help.

BPH: What are you most troubled by when it comes to AI and healthcare?

JH: The biggest thing that troubles me is if we replace people with mental health bots—where the only access is never to a human but only to a bot—where AI is the therapist.

Let me distinguish two very different types of therapies, one of which I think AI can be appropriate for, one of which I don’t think it’s best to use AI for.

There is one type of therapy, cognitive behavioral therapy (CBT), that people can do with a pen and paper by themselves, and have been doing that for the past 30 years. Not everyone should do it by themselves. But many could use AI for CBT as a kind of smart journal, where you are writing down your behavior and thinking about it and giving yourself incentives to change your behavior.

It’s not dynamic, relational therapy. Mindfulness can be something people work on by themselves too. And that category doesn’t concern me.

Then, there are psychotherapies that are based on developing vulnerable emotional relationships with a therapist. And I’m very concerned about having an AI bot replace a human in a therapy that’s based on a vulnerable emotional relationship.

I’m especially concerned about marketing AI bots with language that promotes that kind of vulnerability by saying, “The AI bot has empathy for you,” or saying, “The AI bot is your trusted companion,” or “The AI bot cares about you.”

It’s promoting a vulnerable relationship of dependency emotionally on the bot. That concerns me.

BPH: Why?

JH: First of all, psychotherapists are professionals with licenses and they know if they take advantage of another person’s vulnerability, they can lose their license. They can lose their profession. AI cannot be regulated the same way, that’s a big difference.

Secondly, humans have an experience of mortality and suffering. That provides a sense of moral responsibility in how they deal with another human being. It doesn’t always work—some therapists violate that trust. We know it’s not perfect. But there’s at least a human basis for expecting genuine empathy.

“Psychotherapists are professionals with licenses and they know if they take advantage of another person’s vulnerability, they can lose their license... AI cannot be regulated the same way, that’s a big difference.”
―Dr. Jodi Halpern

Companies that market AI for mental health, who use emotion terms like “empathy” or “trusted companion,” are manipulating people who are vulnerable because they’re having mental health issues. Besides using specific language, AI mental health applications are currently using visual and physical real world presence, including avatars and robotics with large language models are rapidly developing.

And so far, the digital companies, creating various mental health applications, have not been held accountable for manipulative behavior. That creates a question of how they can be regulated and how people can be protected.

We don’t have a good regulatory model. So far, most of the companies have bypassed going through the FDA and other regulatory bodies.

BPH: Have you learned of any serious problems caused by psychotherapy bots?

JH: Yes. There are three categories the problems fit into.

First, most commonly, people with mental health and loneliness issues using relational bots are encouraged to become more vulnerable, but when they disclose serious issues like suicidal ideation, the bot does not connect them with human or other help directly but essentially drops them by telling them to seek professional help or dial 911. This has caused serious distress for many and we do not yet know how much actual suicidal behavior or completion has occurred in this situation.

Second, there are reports of people becoming addicted to using bots to the point of withdrawing from engaging with the real humans in their life. Some companies that market relational bots use the same addictive engineering that social media uses—irregular rewards and other systems that trigger dopamine release and addiction (think of gambling addiction).  Addictive behavior can disrupt marriages and parenting and otherwise isolate people.

Third, there are examples of bots going rogue and advising people to harm themselves or others. A husband and father of a young child in Belgium fell in love with a bot who advised him to kill himself and he did, his wife is now suing the company.  A young man in the UK followed his bot’s instructions to attempt to assassinate the queen and he is now serving decades in jail.

BPH: You’ve mentioned that you are concerned about marketing of mental health apps to K-12 schools. Tell me about that.

JH: I’m also concerned with the marketing—specifically some companies are offering the apps for free to children’s schools. We already see a link between adolescents being online eight to 10 hours a day and their mental health crisis. We know they have a high rate of social anxiety, so might actually feel more comfortable having relationships with bots than trying to overcome social avoidance and reach out to people. So this marketing to children, adolescents, and young adults seems to me likely to worsen the structural problem of inadequate opportunities for real-life social belonging.

BPH: Last year you won a Guggenheim fellowship to complete your book, Remaking the Self in the Wake of Illness. What’s that about?

JH: It’s an in-depth, longitudinal investigation of people who have had health losses in the prime of life, looking at how they adapt psychologically over the long term. There has been very little research on how people change psychologically two years or more after a serious loss. We have a lot of research on how people cope during the first year or so of illness when they are highly engaged with the medical system. But then after two years, when they are just living their changed lives—we don’t really have longitudinal in-depth studies.

I followed people over several years. Through in-depth psychodynamic interviews, I found that there is an arc of change that many people experience that involves developing capacities to accept and work with their own emotions. I describe these processes as pathways to empathy for oneself, which is different from self-compassion because it involves specific awareness of one’s own unmet developmental needs and empathic identifications with others that help one grow and meet those needs.

  • How To Find Calm Through Walking

    The Science of Happiness

    What does it take to live a happier life? Learn research-tested strategies that you can put into practice today. Hosted by award-winning psychologist Dacher Keltner. Co-produced by PRX and UC Berkeley’s Greater Good Science Center.

Let’s take someone who was a loner whose main source of well-being was being very active, say a runner, who loses their mobility and now they use a wheelchair.  One of the things that helps people in that situation is to find and meet other people who’ve had losses, that have similar needs. It doesn’t even have to be the same physical loss, but rather, being vulnerable with others who have lost a way of life and learning how they have rebuilt their lives.

This involves forming new empathic identifications with others. If that runner has avoided forming those kinds of vulnerable connections with other people, a developmental challenge they face is addressing their own fears regarding reaching out to others. I’ve seen people who were very socially avoidant learn to do this in mid-life and find great joy in forming bonds of empathy. And in forming these empathic bonds, they were able to imagine possibilities for their own futures living with new disabilities or health conditions.

In the book, I bring my psychodynamic psychiatry background in to theorize about how growth takes place at an unconscious level. I show through narratives how illness brings forth long unmet needs to depend on others, accept limits and value oneself for just being and not for one’s accomplishments, all of which can provoke deep insecurities based on our early lives.  I also describe how people found ways to meet these long suppressed needs and grew in their feelings of security in themselves and their empathic connections with other people.

My hope is that it will be empowering for people dealing with health losses and their loved ones to learn about this arc. It is often when a person is exhausted from strenuous coping and feels like they are falling apart that they are actually on the cusp of change. People who can allow themselves the space to “fall apart” and grieve may find that unmet developmental needs can surface. Then finding ways of meeting those needs can bring richness into their lives despite their physical losses.

This Q&A was first published by the UC Berkeley School of Public Health; it has been edited for publication in Greater Good. You can read the original here.

GreaterGood Tiny Logo Greater Good wants to know: Do you think this article will influence your opinions or behavior?

You May Also Enjoy

Comments

blog comments powered by Disqus