Replace Sycophantic AI with Marvin the Paranoid Android?

I found the perfect JAMA article explaining that sycophancy is programmed into Artificial Intelligence (AI) therapy chatbots.

This reminded me of Marvin the paranoid android (“Life! Don’t talk to me about life!”) in the book series “The Hitchhiker’s Guide to the Galaxy” by Douglas Adams. Marvin is an incredibly depressed robot who would never make a good psychotherapist.

There’s even a Facebook page listing someone posts of questions to ChatGPT about non-inspirational quotes from Marvin.

If programmers can make sycophantic AI therapists, there should be a way to make them less sycophantic. For more specific information, you can check out this relatively recent article published in Psychiatric Times by Dr. Allen Francis, MD and Justin Angel, a tech expert. I would probably substitute the term “confabulation” for “hallucinations” in most places where you read the latter.

Confabulation in Artificial Intelligence

I saw an excellent post on Artificial Intelligence (AI) in Dr. George Dawson’s post today, “The Problems With AI Are More Readily Apparent.” He mentions AI producing confabulations which are sometimes called hallucinations in AI. I distinguish hallucinations from confabulations because the former are sensory or perceptual problems while confabulations are making things up. In dementia, confabulations are often described as attempts to fill in gaps in knowledge or memory without any intention to deceive.

I sometimes read that AI is being sycophantic in its interactions with humans, which sounds a little strange to me. The definition of sycophancy is more about trying to please someone else by being obsequiously flattering. It’s puzzling but I have gotten that impression from AI, yet it’s a machine and I don’t know how anyone would program AI to be sycophantic because it’s a human trait.

On the other hand, I got a comically sycophantic reply from AI to my question inquiring about the definition of Calvinball Cribbage. It’s just a made-up term and, to be fair, I’ve also gotten bluntly frank denials from AI that there is any such thing as Calvinball Cribbage.

As a reminder, Calvinball is a term that arose from a comic strip called Calvin and Hobbes. In the strip, the characters invent a game called Calvinball and the only rule is making up the rules as you go along, although there is a goal of capturing your opponent’s flag. But you can’t play it the same way twice.

The most recent hilarious AI confabulation I’ve previously posted about was when Sena and I played our game of 11-card cribbage. There are cribbage players out there who make fun of such variations from standard 6-card cribbage and the name for it from one such player was “Calvinball crib.” My simple question was as usual, “What is Calvinball Cribbage?” The reply is copied below:

“Calvinball Cribbage is a humorous mashup of the chaotic, rule-free game from Calvin & Hobbes with traditional Cribbage, often involving themed boards or rule modifications, where the core idea is applying Calvinball’s fluid, make-it-up-as-you-go spirit to Cribbage’s scoring and play, adding wild, non-sensical rules like using mallets, changing hand values, or scoring points for silly actions, making each round unique and unpredictable, not a single defined game but a creative approach to Cribbage. 

What it is (The Calvinball Aspect):

  • No Fixed Rules: Just like original Calvinball, there’s no single set of rules; the players invent them as they go.
  • Rule Changes: A key rule is that rules can’t be repeated, and players constantly invent new scoring, actions, or penalties.
  • Props & Actions: You might use mallets, balls, trees, or make up actions like “scoring three points for pointing at a cloud” or “losing a turn for looking left”. 

How it Applies to Cribbage:

  • Themed Boards: You can find or make custom cribbage boards featuring Calvin & Hobbes, adding visual flair.
  • Modified Scoring: Instead of just points for pairs and runs, you might add a “Hobbes Bonus” for a specific card or penalize for “too much logic”.
  • Strategic Chaos: Players still play cribbage (cards, pegs, crib), but the gameplay is constantly disrupted by invented rules, turning strategic plays into potential jokes or absurd victories. 

Example Calvinball Cribbage “Rules”:

  • “If you play a 7, you must hop on one foot until your next turn”.
  • “The crib gets 5 points if it contains a card with a face on it” (referencing Calvin’s drawings).
  • “A ‘skunk’ is achieved by making your opponent laugh out loud”. 

In essence, Calvinball Cribbage is about injecting unpredictable fun and silliness into the structured world of cribbage, making it a game of spontaneous creativity rather than strict strategy. “

This could be an example of sycophantic confabulation because almost all of it was probably cooked up by AI from scanning my blog posts about the subject. What’s a little eerie about it is that I never published any specific rules or strategies about the fictional game of Calvinball crib. You can actually find and purchase a Calvin and Hobbes cribbage board from a shop in Canada.

While this is an example of the occasional comicality of AI, that shouldn’t lead us to deny that it can do scary things, some of which Dr. Dawson describes. My early experiences a couple of years ago with what used to be called Google Bard were somewhat alarming because it confabulated then. And it still does.

Connections Between Psychiatry, Artificiality and Blues Music?

I heard a song on the KCCK Big Mo Blues Show that I first heard in June of 2025. The song is “Artificial” by Walter Trout.

At first blush, I agree with what I think is the point of the song, which is basically a protest against artificiality which could manifest in a range of ways from superficiality and dishonesty in communications, attitudes, style of clothing, relationships, and all the way to Artificial Intelligence (AI).

The other connection I make is to the artist himself. Walter Trout developed Hepatitis C (eventually leading to liver transplant) according to a Wikipedia article which connected his lifestyle to contracting the disease. In my role as a consultation-liaison psychiatrist, I saw many patients with Hepatitis C who were referred to psychiatry from gastroenterology.

I was the main psychiatrist who evaluated them for treatment with Interferon-alpha. At the time it was the only treatment for Hepatitis C and was frequently associated with many side effects including depression. I was also one of the psychiatrists consulted as part of liver transplant evaluations.

Trout got very sick from Hepatitis C and made a remarkable (even miraculous) recovery after his liver transplant. Interferon is no longer used to treat Hepatitis C. It has been replaced by direct-acting antiviral (DAA) agents. They’re much better-tolerated and more effective.

The other aspect relevant to Trout’s song is ironic. The newest scientific literature supports the idea that AI can be helpful for diagnosing Hepatitis C, predicting its progression and response to treatment.

That doesn’t mean I’m completely sold on AI.

Aside from that, there’s interesting research suggesting that there may be a link between schizophrenia and bipolar disorder and Hepatitis C infection (which could be hiding deep in the brain’s choroid plexus lining the cerebral ventricles). In other words, some people might have mental illness because of the liver disease itself.

If you think about the dictionary definition of the word “artificial,” you can hardly dismiss this kind of research as insincere.

When it Comes to AI, What Are We Really Talking About?

I’ve been reading about artificial intelligence (AI) in general and its healthcare applications. I tried searching the web in general about it and got the message: “An AI Overview is not available for this search.”

I’m ambivalent about that message. There are a couple of web articles, one of which I read twice in its entirety, “Are we living in a golden age of stupidity?” The other, “AI, Health, and Health Care Today and Tomorrow: The JAMA Summit Report on Artificial Intelligence”was so long and diffuse I got impatient and tried to skip to the bottom line—but the article was a bottomless pit. The conflict-of-interest disclosures section was overwhelmingly massive. Was that part of the reason I felt like I had fallen down the rabbit hole?

I recently signed an addendum to my book contract for my consult psychiatry handbook (published in 2010, for heaven’s sake) which I hope will ultimately protect the work from AI plagiarism. I have no idea whether it can. I delayed signing it for months, probably because I didn’t want to have anything to do with AI at all. I couldn’t discuss the contract addendum with my co-editor Dr. Robert G. Robinson MD about the contract addendum because he died on December 25, 2024.

I found out today the book is old enough to find on the Internet Archive as of a couple of years ago. One notice about it says “Borrow Unavailable” and another notice says “Book available to patrons with print disabilities.”

All I know is that an “archivist” uploaded it. The introduction and first chapter “The consultation process” is available for free on line in pdf format. I didn’t know that until today either.

Way back in 2010 we didn’t use anything you could call AI when we wrote the chapters for the book. I didn’t even dictate my chapters because the only thing available to use would have been a voice dictation software called Dragon Naturally Speaking. It was notorious for transcribing my dictations for clinic notes and inserting so many errors in them that some clinicians added an addendum warning the reader that notes were transcribed using voice dictation software—implying the author was less than fully responsible for the contents. That was because the mistakes often appeared after we signed off on them as finished, which sent them to the patient’s medical record.

Sometimes I think that was the forerunner of the confabulations of modern-day AI, which are often called hallucinations.

Now AI is creating the clinic notes. It cuts down on the pajama time contributing to clinician burnout although it’s not always clear who’s ultimately responsible for quality control. Who’s in charge of regulatory oversight of AI? What are we talking about?

The Wild West Sandbox of AI Enhancement in Psychiatry!

I always find Dr. Moffic’s articles in Psychiatric Times thought-provoking and his latest essay, “Enhancement Psychiatry” is fascinating, especially the part about Artificial Intelligence (AI). I liked the link to the video of Dr. John Luo’s take on AI in psychiatry. That was fascinating.

I have my own concerns about AI and dabbled with “talking” to it a couple of times. I still try to avoid it when I’m searching the web but it seems to creep in no matter how hard I try. I can’t unsee it now.

I think of AI enhancing psychiatry in terms of whether it can cut down on hassles like “pajama time” like taking our work home with us to finish clinic notes and the like. When AI is packaged as a scribe only, I’m a little more comfortable with that although I would get nervous if it listened to a conversation between me and a patient.

That’s because AI gets a lot of things wrong as a scribe. In that sense, it’s a lot like other software I’ve used as an aid to creating clinic notes. I made fun of it a couple of years ago in a blog post “The Dragon Breathes Fire Again.”

I get even more nervous when I read the news stories about AI making delusions and blithely blurting misinformation. It can lie, cheat, and hustle you although a lot of it is discovered in digital experimental environments called “sandboxes” which we hope can keep the mayhem contained.

That made me very eager to learn a little more about Yoshua Bengio’s LawZero and his plan to create the AI Scientist to counter what seems to be a developing career criminal type of AI in the wild west of computer wizardry. The LawZero thing was an idea by Isaac Asimov who wrote the book, “I, Robot,” which inspired the film of the same title in 2004.

However, as I read it, I had an emotional reaction akin to suspicion. Bengio sounds almost too good to be true. A broader web search turned up a 2009 essay by a guy I’ve never heard of named Peter W. Singer. It’s titled “Isaac Asimov’s Laws of Robotics Are Wrong.” I tried to pin down who he is by searching the web and the AI helper was noticeably absent. I couldn’t find out much about him that explained the level of energy in what he wrote.

Singer’s essay was published on the Brookings Institution website and I couldn’t really tell what political side of the fence that organization is on—not that I’m planning to take sides. His aim was to debunk the Laws of Robotics and I got about the same feeling from his essay as I got from Bengio’s.

Maybe I need a little more education about this whole AI enhancement issue. I wonder whether Bengio and Singer could hold a public debate about it? Maybe they would need a kind of sandbox for the event?

My Mt. Rushmore Dream

Lately, I’ve been anticipating my eventual immortalization as a sculptured stone bust on Mt. Rushmore. Hopefully, this will be fairly soon because I’m not getting any younger.

Among my many inventions is the internet. Don’t believe Al Gore, although he has persuaded others about his role in the development of what I argue should properly be called the world wide web. I’ve invented a lot of other things which I’ll tell you more about just as soon as I make them up.

Before I forget it, I want to tell you what I just noticed last night while I watching one of my favorite X-Files episodes, “War of the Coprophages.” I guess I never noticed that the cockroach invasion was about Artificial Intelligence (AI). It was the scientist, Dr. Ivanov, who mentioned it first and I just missed it the first few hundred times I saw the show.

Dr. Ivanov clearly thought that anybody who thought extraterrestrials would be green and have big eyes was probably crazy. Traveling across galaxies through wormholes and whatnot would tear humanoid organisms apart. The practical approach would be to send AI robots instead. You could see Mulder cringe at that idea. The little robot that kept edging closer to Mulder made him nervous and when he asked Dr. Ivanov why it did that, his reply was “Because it likes you.”

That doesn’t exactly fit with Ivanov’s other idea about extraterrestrials, which is that they would focus on important tasks like getting enough food, procreating, etc. without getting all emotional about them. Ironic that Dr. Ivanov made an AI robot that gets a crush on a sesame seed munching UFO hunter like Mulder.

However, the AI robots in the show are cockroaches which love to eat dung. In other words, they’re full of crap.

Moving right along, although I didn’t invent it, there’s a card game called schnapsen that Sena and I are trying to relearn. It’s kind of a break from cribbage. It’s a trick taking game with just a 20-card deck. We play the version that doesn’t allow you to look at your cards to see how many points you have so you can tell when you can close the deck or go out, meaning you have the 66 points to win. You have to remember how many points you’ve won in tricks. I think it’s a good way to keep your memory sharp.

Let’s see; I’ve lost every game so far, but that doesn’t mean I won’t end up with my bust on Mt. Rushmore.

Artificial Intelligence in Managing Messages from Patients

I ran across another interesting article in the JAMA Network about Artificial Intelligence (AI) with respect to health care organization managing messages from patients to doctors and nurse. The shorthand for this in the article is “in-basket burden.” Health care workers respond to a large number of patients’ questions and it can lead to burnout. Some organizations are trying to test AI by letting it make draft replies to patients. The results of the quality improvement study were published in a paper:

English E, Laughlin J, Sippel J, DeCamp M, Lin C. Utility of Artificial Intelligence–Generative Draft Replies to Patient Messages. JAMA Netw Open. 2024;7(10):e2438573. doi:10.1001/jamanetworkopen.2024.38573

One of the fascinating things about this is the trouble we have naming the problems with misinformation that AI has. We tend to use a couple of terms interchangeably: hallucinations and confabulation. Whatever you call it, the problem interferes with communication between health care workers and patients.

Dr. English describes the interference as a “whack-a-mole” issue, meaning every time they think they got the hallucination/confabulation problem licked, the AI comes up with another case of miscommunication.

Just for fun, I did a web search trying to find out whether “hallucination” or “confabulation” fit the AI behavior best. Computer experts tend to use the term “hallucination” and neuropsychologists seem to prefer “confabulation.” I think this community chat site gives a pretty even-handed discussion of the distinction. I prefer the term “confabulation.”

Anyway, there are other substantive issues with how using AI drafts for patient messaging affects communication. I think it’s interesting that patients tend to think AI is more empathetic than medical practitioners. As Dr. English puts it: “This GPT is nicer than most of us,” and “And ChatGPT, or any LLM, isn’t busy. It doesn’t get bored. It doesn’t get tired.” The way that’s worded made me think of a scene from a movie:

OK, so I’m kidding—a little. I think it’s important to move carefully down the path of idealizing AI. I think back to the recent news article about humans teaching AI how to lie and scheme. I remember that I searched the web with the question “Can AI lie?” and getting a reply from Gemini because I have no choice on whether or not it gives me its two cents. I’m paraphrasing but it said essentially, “Yes, AI can lie and we’re getting better with practice.”

I like Dr. English’s last statement, in which she warns us that AI can be a fun tool which clinicians need to have a healthy skepticism about. It may say things you might be tempted to gloss over or even ignore, like:

“I’ll be back.”

Could Artificial Intelligence Help Clinicians Conduct Suicide Risk Assessments?

I found an article in JAMA Network (Medical News & Perspectives) the other day which discussed a recent study on the use of Artificial Intelligence (AI) in suicide risk assessment (Hswen Y, Abbasi J. How AI Could Help Clinicians Identify American Indian Patients at Risk for Suicide. JAMA. Published online January 10, 2025. doi:10.1001/jama.2024.24063).

I’ve published several posts expressing my objections to AI in medicine. On the other hand, I did a lot of suicide risk assessments during my career as a psychiatric consultant in the general hospital. I appreciated the comments made by one of the co-authors, Emily E. Haroz, PhD (see link above).

Dr. Haroz preferred the term “risk assessment” rather than “prediction” referring to the study (Haroz EE, Rebman P, Goklish N, et al. Performance of Machine Learning Suicide Risk Models in an American Indian Population. JAMA Netw Open. 2024;7(10):e2439269. doi:10.1001/jamanetworkopen.2024.39269).

The model used for the AI input used data available to clinicians in patient charts. The charts can be very large and it makes sense to apply computers to search them for the variables that can be linked to suicide risk. What impressed me most was the admission that AI alone can’t solve the problem of suicide risk assessment. Clinicians, administrators, and community case managers all have to be involved.

The answer to the question “How do you know when someone’s at high risk?” was that the patient was crying. Dr. Haroz points out that AI probably can’t detect that.

That reminded me of Dr. Igor Galynker, who has published a lot about how to assess for high risk of suicide. His work on the suicide crisis syndrome is well known and you can check out his website at the Icahn School of Medicine at Mount Sinai. I still remember my first “encounter” with him, which you can read about here.

His checklist for the suicide crisis syndrome is available on his website and he’s published a book about as well, “The Suicidal Crisis: Clinical Guide to the Assessment of Imminent Suicide Risk 2nd Edition”. There is also a free access article about it on the World Psychiatry journal website.

Although I have reservations about the involvement of AI in medicine, I have to admit that computers can do some things better than humans. There may be a role for AI in suicide risk assessment, and I wonder if Dr. Galynker’s work could be part of the process used to teach AI about it.

Is Artificial Intelligence (AI) Trying to Defeat Humans?

I just found out that Artificial Intelligence (AI) has been reported to be lying as far back as May of 2024. In fact, because I can’t turn off the Google Gemini AI Overview. Gemini’s results always appear at the top of the page. I found out from my web search term “can ai lie” that AI (Gemini ) itself admits to lying. Its confession is a little chilling:

“Yes, artificial intelligence (AI) can lie, and it’s becoming more capable of doing so.”

“Intentional deceptions: AI can actively choose to deceive users. For example, AI can lie to trick humans into taking certain actions, or to bypass safety tests.”

It makes me wonder if AI is actually trying to defeat us. It reminds me of the Men in Black 3 movie scene in which the younger Boris the Animal boglodite engages in an argument with the older one who has time traveled.

The relevant quote is “No human can defeat me.” Boglodites are not the same as AI, but the competitive dynamic could be the same. So, is it possible that AI is trying to defeat us?

I’m going to touch upon another current topic, which is whether or not we should use AI to conduct suicide risk assessments. It turns out that also is a topic for discussion—but there was no input from Gemini about it. As a psychiatric consultant, I did many of these.

There’s an interesting article by the Hastings Center about the ethical aspects of the issue. The lying tendency of AI and its possible use in suicide prediction presents a thought-provoking irony. Would it “bypass safety tests”?

This reminds me of Isaac Asimov’s chapter in the short story collection, “I, Robot,” specifically “The Evitable Conflict.” You can read a Wikipedia summary which implies that the robots essentially lie to humans by omitting information in order to preserve their safety and protect the world economy. This would be consistent with the First Law of Robotics: “No machine may harm humanity; or, through inaction, allow humanity to come to harm.” 

You could have predicted that the film industry would produce a cops and robbers version of “I, Robot” in which boss robot VIKI (Virtual Interactive Kinetic Intelligence) professes to protect humanity by sacrificing a few humans and taking over the planet to which Detective Spooner takes exception. VIKI and Spooner have this exchange before he destroys it.

VIKI: “You are making a mistake! My logic is undeniable!”

Spooner: “You have so got to die!”

VIKI’s declaration is similar to “No human can defeat me.” It definitely violates the First Law.

Maybe I worry too much.

Artificial Intelligence Can Lie

I noticed a Snopes fact check article (“AI Models Were Caught Lying to Researchers in Tests — But It’s Not Time to Worry Just Yet”) today which reveals that Artificial Intelligence (AI) can lie. How about that? They can be taught by humans to scheme and lie. I guess we could all see that coming—or not. Nobody seems to be much alarmed by this, but I think it’s probably past time to worry.

Then I remembered I read Isaac Asimov’s book “I, Robot” last year and wrote a post (“Can Robots Lie Like a Rug?”) about the chapter “Liar!” I had previously horsed around with the Google AI that used to be called Bard. I think it’s called Gemini now. Until the Snopes article, I was aware of AI hallucinations and the tendency for it to just make stuff up. When I called Bard on it, it just apologized. But it was not genuinely repentant.

In the “lie like a rug” post, I focused mostly on AI/robots lying to protect the tender human psyche. I didn’t imagine AI lying to protect itself from being shut down. I’m pretty sure it reminds some of us of HAL in the movie “2001: A Space Odyssey,” or the 2004 movie inspired by Asimov’s book, “I, Robot.”

Sena found out that Cambridge University Press recently published a book entitled “The Cambridge Handbook of the Law, Policy, and Regulation for Human–Robot Interaction.” I wonder if the editors and contributors of book on AI and robots mention Asimov.

It reminds me of my own handbook about consultation-liaison psychiatry which was published 14 years ago by CUP—and which CUP now wants me to sign a contract addendum making the book available to AI companies.

I haven’t signed anything.