First, I would like to acknowledge that a number of people offered advice on this topic, and they generally know more than I do about philosophy and education. I would name them, but then I would want to have their approval, and I take ultimate responsibility for this text.
On October 22, 2025, I was a co-presenter of a talk at Winona State University to the informal AI Discussion Group along with Professor Malgarzata Plecka. I billed the talk as "The Reverse Turing Test." The basic thesis of the talk was not groundbreaking in the least: if you are trying to discern whether an academic text was generated by AI rather than written by a person, then one strategy is to check the footnotes and references. AI systems will sometimes fabricate references to articles that don't really exist. If a submitted paper has fabricated references, a reviewer can reject it not because it was generated by AI, but because it is a flawed paper. It was probably generated by AI, but that is almost beside the point.
As part of that talk, I digressed onto the philosophical issues pertaining to the situation. Namely, I touched on philosophical theories of Truth. In last month's entry for this blog, I compared "truth as factual accuracy" with "truth as loyalty." A friend of mine who is more familiar with these issues pointed me toward a philosophical position called "The Correspondence Theory of Truth." Does a given statement correspond to a state of affairs in the world?
I sent this paragraph to a philosophy professor:
Some schools of thought deprecate the belief that there is an external, objective, observable world at all, much less one that can be measured. I am reminded of the joke about the drunk at the bar during the philosophy conference who said, "Solipsism works for me, but that’s just one man’s opinion."
That professor pointed out that there are a variety of philosophical schools that don't fall into this dichotomy of "correspondence theory of truth" or solipsism. There are deflationist theories and pragmatist theories which still believe in the external world. Further, one can deny the existence of the external world without being a solipsist, such as in Bishop Berkeley's philosophy of idealism.
Circling back to factual accuracy, though, even if you don't buy a theory similar to the correspondence theory in general, it is neverthesless the case that a research library is part of a scholarly correspondence theory of truth. Even for scholars who cast a jaundiced eye toward the value of telling the truth, they do largely adhere to a different commitment to factual accuracy: within their own scholarly texts, they try to accurately quote and reference other writings.
For example, I have read a book by Melvin Pollner called Mundane Reason. Pollner develops a theory of ethnomethodology, which he then uses to question the existence of the objective, determinate world. Truth is socially constructed, and a community will develop a unitary world of social truth.
Even if I am a scholar who is a follower of Pollner or Bishop Berkeley, I will still adhere to a correspondence theory of truth in my production of my own scholarly texts. If I claim a quotation from one of their texts or a text that discusses them, I will make a citation that will correspond to the location of that statement in the objective, observable universe of texts managed by libraries.
These musings are relevant to a discussion of AI LLMs, because LLMs generally string statements together without worrying whether they are factually accurate statements about the universe of scholarly texts. ChatGPT will often just make crap up. Some commentators politely refer to made up quotes and references as hallucinations. Others rely on Harry Frankfurt’s definition and refer to the act of making up stuff, including references, as bullshitting. Some library researchers have created a metric for testing LLM systems called the Reference Hallucination Score. See citations below.
The scenario that we presented on October 22nd was this: a researcher has claimed to have authored a paper that was submitted to an academic journal. The editors of the journal have a policy of not allowing texts that are generated by AI. The basic justification for detecting AI-generated texts is that they should be identified and rejected. Why should they be rejected, though?
Obviously, if a student is assigned to write a paper without the assistance of AI, and the professor can show that the paper was generated by AI, that is a violation of the assignment. But if the paper is a genuinely brilliant and well-executed paper, perhaps an academic journal should consider publishing it, even if it is generated by an AI LLM. Of course, if the references in the paper are fabricated, then the paper is by definition flawed, regardless of who or what generated it.
During the discussion after the presentation, a professor argued that he uses AI LLMs so extensively that a text under his name is almost always a fusion of his own words and those of the AI LLM, and that he himself might not know which of the two crafted any one passage. The basic motive for detecting AI-generated texts is that human-generated texts are preferred. The basic point of the other professor's claim of blending is that the AI-generated text is superior to the human-generated text and should be utilized. Those two positions are not necessarily opposed to one another.
The core of the problem is whether the human claiming authorship is actually the author of the work. Here I offer a modified definition of author, namely: the human is the author if that human is an authority on the content in the work, and the human takes responsibility for what it says. An AI LLM might have assisted with the wording, or even some of the ideas, but the human should have both a command of the content and editorial control of the final text. After all, human ghost writers have been around for millennia; this is not a new concept.
If a paper has fabricated references, then it is likely that the purported human author left the generation of the text entirely up to the AI system. The human might not understand what the paper says or what it contains. For a paper that is entirely composed by an AI system, the human is not the author of the paper under this definition.
If, on the other hand, the human is fully in control of the subject matter and is using the AI system as a tool to improve the final product, then that case probably falls on the acceptable end of the spectrum. That assertion of authorship, though, would involve actual work by the human.
For instance, if the AI system offers up references, the human would be responsible for actually tracking down those texts (and in the process verifying their existence), reading them, and establishing that they are relevant to the argument. This problem of AI-fabricated references is related to the long-standing problem that arises when human authors borrow citations from earlier human-written texts without verifying them. If the earlier paper has a mistake in the reference, that mistake will be perpetuated by later authors who fail to verify the reference and read the cited work.
You can even couch this discussion within a larger context of education: To me, the point of the university is 1) to educate students, 2) to expand the realm of knowledge through research and scholarship and creative works, and 3) to preserve that realm of knowledge and share it with the broader community.
Educating students can be thought of as the process of building neural networks. However, they are the biological networks made of actual neurons and located between the ears of the students, not neural networks located in massive data centers (which in the industry are sometimes called "multi-layer perceptrons"). If the students (or faculty members) rely solely on an artificial neural network, they are often not building the biologically-based neural network in their own brains, and we are failing to educate them.
A complicating aspect of today’s world is that AI is ubiquitous. One of the things our studentsand our facultywill have to learn is how to delve into AI systems to leverage their own knowledge while still retaining intellectual control and responsibility. ( "Delve" and "leverage" are two words commonly used by AI LLMs when writing proseas are m-dashes.) That authorship requires work, and AI systems make it easy to skip the work and masquerade as the author. Worse, employers may require workers to give up their authorship to increase productivity (that is, increase the quantity of product, not its quality).
NOTE: None of the above text was generated by an AI LLM. It is all the product of the wetware inside my skull, plus the broader collective knowledge held by the community of people with whom I consulted and written texts which are part of my culture (and the product of cultural evolution).
Afterword, December 12, 2025
I shared the above blog entry with my friend Will Dowling, who is retired now from, first, teaching computer science, and then, working in the computer industry. He now teaches an occasional AI course as an adjunct at Temple University. He was the person who suggested that I look into the correspondence theory of truth.
One question he asked when he read my phrase "fabricated references" was: "how would you know whether incorrect references are fabricated, or due to innocent error?" My answer: "You cannot know for certain that a citation was fabricated, but there is a difference between a typo and a citation where none of the details relate to an actual citation. In the paper that Professor Plecka brought to me and my colleagues, NONE of the citations existed in the observable universe of scholarly texts. One citation had for its title a phrase from a quote that the supposed author had made in another, existing article. The AI system probably found that quote and used it when cobbling together its references. Humans can fabricate references, too, but it would be time consuming to do so, and it would be faster to go into a library database and download a number of real citations on the relevant topic."
To my claim that "if there are fabricated references, then the supposed author probably left the writing of the paper entirely to an AI system," Will writes: "I don't buy this. Say I write an article, by myself, no help from AI. It has 10 cited references. I need to fluff up the references section before submission. I go to Claude, and now I have 40 references." My response: Good point.
Will also took issue with my statement: "The basic motive for detecting AI-generated texts is that human-generated texts are preferred." He noted that studies have shown that the average person prefers a text written by an LLM to a text written by a human. That may be true--he cited an article that really exists to back up the point (https://arxiv.org/pdf/2510.13939v3)--but what I meant was that editors of journals and readers who want to learn from other humans might prefer a text that was generated by a human.
Will took issue with my "jaundiced eye" statement. Just because someone does not subscribe to the correspondence theory of truth does not mean that the person does not value telling the truth. My reply: Fair enough. Melvin Pollner might not dislike truth-telling, he just defines truth in a particular way that includes "the politics of experience."
Will didn't agree with my definition of an author. He writes, "Who gets to judge whether you are an authority of the content in the work? This moves what should be unambiguous, author or not?, to the realm of judgment call. I prefer the common (JAMA; The Lancet; Nature; PLOS ONE; etc.) requirement that contributing co-authors must enumerate their contributions including categories such as conceptualization, writing (original draft) and writing (review and editing)."
My response: I agree with Will's basic point. But in libraries, there are instances where a corporation is an author, such as a congressional committee being the author of the text of a hearing. My definition would allow one to argue that an editor at a publishing house is equally the author of a book that the house edited. But in literary studies, there are already debates about the degree to which authorship is shared among a group of people in the text production process, so this is not a new problem. I guess I would retreat back to the point that the author is the entity which takes full responsibility for the text. Whether that person or legal entity knows what is in the text or is an authority on the content, that is a judgment call.
Thanks Will. Your critique is very much welcome and has helped me improve my thinking on this topic.
Bibliography
Aljamaan, Fadi, et al. (2024). "Reference Hallucination Score for Medical Artificial Intelligence Chatbots: Development and Usability Study." JMIR Medical Informatics 12 (1): e54345. doi: 10.2196/54345
Hicks, Michael Townsen, James Humphries, and Joe Slater. (2024). "ChatGPT is Bullshit." Ethics and Information Technology 26: 30. doi: 10.1007/s10676-024-09775-5
Pollner, Melvin. (1987). Mundane Reason. Cambridge UP.