Reading Our Minds by Daniel Barron, a psychiatrist and pain management fellow, explores the incorporation of Big Data to improve the practice of psychiatry. The idea of supporting psychiatric assessment with solid data is an appealing one, but many questions come to mind.
I was surprised by an apparent blind spot of the author’s that appeared early on. He writes:
“A recent study showed that Google searches for explicitly suicidal terms were better able to predict completed suicides than conventional self-report measures of suicide risk. Perhaps this is because people who are ‘really gonna do it’ go through the planning and researching (i.e., on Google) of how to kill themselves, but it could be that people are more honest when they approach Google with what’s on their mind.”
Perhaps he hasn’t caught on to the fact that those of us with mental illness are well aware that the doctor who’s asking us questions about suicide has the power to commit us to hospital involuntarily, where our clothes and belongings will be taken away and we may be locked in a seclusion room with nothing to do or think about except how wrong it was to be honest with that damn doctor. That’s not a hypothetical, either; that’s exactly what goes through my mind when I’m contemplating disclosing suicidal ideation to a doctor, because that has happened in the past. Aside from that, though, just imagine if Google had an algorithm that would flag it to emergency services if they thought you were getting a little too close to the edge. I think I’d be motivated to start using Tor. Hello dark web!
A running example through the book was the author’s assessment of a girl he ended up diagnosing with schizophrenia. Her mom reported that she’d had changes in behaviour patterns, social engagement, and internet use, and the author argued it would have been useful to have her browsing history, geolocation data, call/text logs, etc., as this would help to establish her baseline “normal” and what deviated from that.
The results of a number of relevant studies were presented. For example, changes in Twitter behaviour were observed in women who developed postpartum depression. Another study looked at Facebook posts by people with psychotic disorders and noted distinct changes that were seen shortly before people ended up being hospitalized. The changes included more swearing, anger, and references to death.
There were some interesting suggestions for objectively measuring things that are currently evaluated subjectively, which I agree would very much be of benefit to the practice of psychiatry. Speech was one of the examples given. I experience speech impairment as a psychomotor effect of depression, and it could be quite useful to be able to monitor that in a clinical setting.
If you’re wondering about the issue of consent and privacy with all this data, it came up, but it didn’t seem to be treated as much of a barrier. The author writes that he began the book thinking that it would be hard to get patients to agree to data collection, but COVID proved him wrong. As an example, he pointed out that people were willing to download apps that would track geolocation to determine COVID contacts. I’m quite confident in saying that the identifiable data that I’d be willing to give up in the context of a deadly pandemic is not going to be the same as what I’d give up to a psychiatrist.
I think this where another big blind spot comes in. Patients are people. There is a significant power differential between psychiatrist and patient. Involuntary treatment takes away people’s rights for the sake of treatment. Even when treatment is voluntary, decisions are often made by the prescriber alone rather than as part of a collaborative process that supports the agency of individuals with mental illness. Sometimes physicians assume that patients should be able to put up with side effects rather than recognizing the patient’s right to make those choices for themselves. Mental health professionals are in no way immune to stigma; this is borne out both anecdotally and in the research literature.
I could go on, but that’s already a whole lot of context to consider, and it’s disappointing that the author just doesn’t seem to consider it. There’s no indication in the book that the author has sought out feedback from anyone on the patient side of the fence to see how they would feel about the idea of handing over their Google search history to their psychiatrist; perhaps this wasn’t seen as an important part of the process?
It seems like too big an overlook to be accidental that patients don’t appear in this book as people who are empowered to be advocates for themselves, their health care, and their privacy. To assume that patients will readily hand over anything the good doctor wants smacks of paternalism. That’s especially true when no argument has been offered about how all of this Big Data will benefit patients.
As someone who has straddled the patient and mental health professional side of the fence, I say a) back away from my data, and b) I would recommend the author reflect on what that fence looks like for him, and what it might be preventing him from seeing.
Reading Our Minds is available on Amazon.
I received a reviewer copy from the publisher through Netgalley.
You can find my other book reviews here.
This post contains affiliate links, which let you support MH@H at no extra cost to you.