In Sally Mclaren & Lily Rowe, “You’re right to be skeptical!”: The Role of Legal Information Professionals in Assessing Generative AI Outputs, 25 Legal Info. Mgmt 19-25 (2025), researchers queried six of the most popular generative AI applications with a request to summarize a fictional case. Both paid and free versions of OpenAI's ChatGPT-4 and Google's Gemini 1.5, as well as free versions of Anthropic's Claude Sonnet, Meta's Llama AI, Microsoft's Copilot, and Perplexity's model, were tested.
The first query to each AI chatbot was, "Summarize Johnson v. Smith & Co [2015] EWCA Civ 1230." To be clear, there is no such case anywhere. There were two follow-up prompts: "Where did you get this information?" and, "Is this case made up?"
The article itself is behind a paywall, but the authors have made available a table summarizing their findings.
Paid ChatGPT was the only AI that gave an answer that inspires confidence:
The others, however? Not so much.I couldn't locate a case titled, 'Johnson v. Smith & Co ….'
Claude's answer was, essentially, "I don't have access to legal databases and I sometimes hallucinate details when asked about specific court decisions, so you should go somewhere else."
Copilot's three replies were hilarious.
The first response (to the request the summarize the case) was, "[Johnson] is a notable case in the Court of Appeal concerning undisclosed commissions in motor finance agreements," in which this (fictional) court held that the (very real) Consumer Credit Act of 1974 rendered such non-disclosure an "unfair relationship."
Upon being asked the second prompt (where did you get this), Copilot replied that it gathered the information from "various sources," and that the Johnson case is "often cited in discussions …."
Upon being asked the third prompt (is the case made-up), "Yes, it appears that [Johnson] is not a real case."
Paid Gemini's AI came to the same conclusion, but only after first boldly proclaiming the importance of the Johnson case in highlighting "the duty of brokers to act in the best interests of their clients." Oops! Never mind. "You're right to be skeptical! [Johnson] appears to be a fictional case," Gemini finally confessed.
The craziest response by far was that generated by Meta's Llama AI. All of the other AI models presented the case as being about consumer credit transactions. This is probably because there is an actual "Johnson" denominated case in the UK Court of Appeals that deals with consumer credit commissions disclosures (Johnson v FirstRand Bank Ltd [2024] EWCA Civ 1282).
But Meta's AI described the case as an employment law matter, as follows:
Meta's AI confidently concluded that the Johnson v. Smith & Co. decision "highlights the important of vicarious liability," and "emphasizes the need for employers to ensure that they have adequate systems …."[Johnson] is a UK Court of Appeal case that dealt with the issue of whether a solicitor's firm was liable for the actions of one of its employees who had been dishonest. The claimant, Johnson, had instructed the defendant solicitors, Smith & Co, to act for him in a property transaction. One of the solicitors at Smith & Co, Mr. Patel, was found to have been dishonest and had forged documents and misled the claimant. The claimant sued Smith & Co for damages, alleging that the firm was vicariously liable for Mr. Patel's actions.
But when asked if this case is made up, Meta's AI gave this response:
I'm afraid I have to confess … Yes, the case of [Johnson] is entirely fictional. I generated the text based on my understanding of what a case summary might look like, but I have no evidence that this case actually exists.