No, AI does not have human-level intelligence

Before we can establish that AI has genuine intelligence, we need to answer a metaphysical question and an epistemological question.

(Image: Immo Wegmann / Unsplash.com)

In an article at Nature, Eddy Keming Chen, Mikhail Belkin, Leon Bergen, and David Danks ask, “Does AI already have human-level intelligence?” and claim that “the evidence is clear” that the answer is Yes. (Though the article is partially pay-walled, a read-only PDF is available here.) But as is typical with bold claims about AI, their arguments are underwhelming, riddled with begged questions and other fallacies.

Defining “intelligence”

Naturally, before we can establish that AI has genuine intelligence, we need to make clear what it would be for it to have intelligence, and how we could go about determining that it has it. The first is a metaphysical question; the second is an epistemological question. Our authors make no serious attempt to answer either one.

Explaining what they mean by “general intelligence,” they write: “A common informal definition of general intelligence, and the starting point of our discussions, is a system that can do almost all cognitive tasks that a human can do.” Expanding on this, they say:

General intelligence is about having sufficient breadth and depth of cognitive abilities, with ‘sufficient’ anchored by paradigm cases. Breadth means abilities across multiple domains–mathematics, language, science, practical reasoning, creative tasks–in contrast to ‘narrow’ intelligences, such as a calculator or a chess-playing program. Depth means strong performance within those domains, not merely superficial engagement.

There are three serious problems here. The first is that it is not illuminating to say that general intelligence entails the ability to carry out cognitive tasks unless our authors have already given us (as they have not) an explanation of what they mean by “cognitive.” Now, in one common usage, “cognitive” tasks are activities of the kind that require intelligence. So, if that is what they mean, then our authors’ definition is circular–they are defining intelligence in terms of cognition, where cognition is (implicitly) defined in terms of intelligence. And as anyone who has taken a logic course knows, one of the fundamental rules of a good definition is that it should not be circular.

Second, another basic rule for definitions familiar from logic is that a good definition should not be too broad. For example, if I defined “gold” as a yellowish metal, this would violate the rule, because it would fail to make clear what distinguishes gold from pyrite. Now, our authors violate this rule as well, because their characterization of intelligence is not precise enough to distinguish genuine intelligence from mere mimicry. They matter-of-factly refer to what calculators and chess-playing programs do as instances of “intelligence.” But it is a commonplace among critics of AI that what calculators and chess-playing programs do is mere mimicry and not genuine intelligence at all, not even of the “narrow” kind.

Of course, our authors would no doubt disagree with the critics. But the point is that a good definition of intelligence should provide us with an independent guide for determining who is right. It should make it clear what it would be to have genuine intelligence as opposed to mere mimicry, so that we could then go on to establish on that basis whether calculators and chess-playing machines actually have it. Imagine a gold miner trying to prove that what he has dug up really is gold rather than pyrite by saying, “I define ‘gold’ as ‘a yellow metal.’ So, the evidence is clear that this is gold!” Obviously, even if what he has really is gold, he cannot establish that it is with that particular definition, because it is so broad that even a mere simulation of gold would meet it. Similarly, our authors cannot claim to have established that AI has genuine intelligence when they are working from a definition so broad that even a simulation would meet it. Even if they were right, their argument could not show that they are right, because given their definition of “intelligence,” it simply begs the question.

A third problem with our authors’ definition is that it violates yet another standard rule for definitions, which is that they should capture the essence of the thing defined. Part of what this involves is leaving out of a definition any reference to features that the thing defined needn’t actually possess. For example, it would be a mistake to define “gold” as a metal used in making jewelry, because even though that is true of gold, it is not essential to gold that it be used in making jewelry. Gold would still have the same nature it has even if human beings had never decided to make jewelry out of it.

But another part of what this rule involves–and the part relevant to our concerns here–is sensitivity to the fact that even features that always exist in things of a certain type are not necessarily part of the essence of the thing. They may instead flow from its essence as “proper accidents” (to use the traditional Scholastic jargon). For example, water is clear and liquid at room temperature, but this isn’t plausibly the essence of water. Rather, these features follow from water’s having the essence it has, which is (either in whole or in part, depending on different views about essence I won’t adjudicate here) a matter of having the chemical composition H2O. And things with a different essence from water might have these same features (as heavy water does, for example). (I discuss the distinction between essence and proper accidents in my book Scholastic Metaphysics, at pp. 230-35.)

Now, our authors violate the rule that a definition should capture a thing’s essence when they define intelligence in terms of a capacity for tasks involving “mathematics, language, science,” and the like. Capacities for mathematics, language, and science certainly follow from our having intelligence, but they are not themselves the essence of intelligence. (That is why they don’t always manifest–though they “flow” from our having intelligence, the flow can be “blocked,” as it were, by immaturity, brain damage, or what have you.)

What would be the essence of intelligence? I would say that it has to do with the interconnected capacities for forming abstract concepts, putting them together into propositions, and reasoning logically from one proposition to another. (See chapter 3 of my book Immortal Souls: A Treatise on Human Nature for a detailed exposition and defense of this traditional conception of intelligence.) It is because we have intelligence in this sense that we are capable of mathematics, language, science, and so on. Those capacities flow or follow from intelligence in this sense.

Of course, our authors may well disagree with this account. But the point for present purposes is that their attempted definition doesn’t reflect any awareness of the need to distinguish essence from proper accidents, and to define intelligence in terms of the former rather than the latter. As in the other ways noted, their account is simply conceptually sloppy.

This sloppiness manifests also in what they say intelligence does not involve. They write:

Intelligence is a functional property that can be realized in different substrates–a point Turing embraced in 1950 by setting aside human biology. Systems demonstrating general intelligence need not replicate human cognitive architecture or understand human cultural references.

This is not wrong, but, without saying more, it is also not very helpful. Suppose I suggested to you that a stone is intelligent and you replied that seems like an absurd claim given that stones can’t speak or reason logically, lack basic knowledge, and so on. And suppose I responded: “True, but remember that systems demonstrating general intelligence need not replicate human cognitive architecture or understand human cultural references.” Presumably, you would not be impressed with this response.

The problem, obviously, is that while genuine intelligence need not always look exactly the way it does in us, it nevertheless is not true that just anything goes. We need some criteria for determining when something departs too far from how intelligence manifests in us to count as genuinely intelligent. And our authors offer no such criteria. They simply assume that, wherever we draw this line, AI will fall on the “genuine intelligence” side of it. But since they merely assume this rather than argue for it, their position once again begs the question.

Detecting intelligence

Having failed to provide a serious definition of intelligence, it is no surprise that they also fail to provide a serious account of how to go about detecting intelligence (since the latter task presupposes the former). There is a lot of hand-waving about “a cascade of evidence,” and gee-whiz references to what LLMs can do. But all of this ultimately boils down to nothing more than a stale appeal to the Turing test. And the problem with the Turing test is that, of its very nature, it cannot distinguish genuine intelligence from a mere clever simulation. Indeed, it deliberately ignores the difference and focuses narrowly on the question of what would lead us to judge a machine to be intelligent, rather than the question of what would make it the case that a machine actually is intelligent.

Since it is the latter question that is at issue here, the Turing test is simply irrelevant. (And as I have argued elsewhere, to make it relevant, the defender of the view that AI is genuinely intelligent will have to appeal to either verificationism or scientism, and the resulting position will be either self-defeating or question-begging.)

But the problem with our authors’ argument is worse than that. It’s not just that they haven’t shown that AI has genuine intelligence. It’s that we already know that it does not have it, and that it amounts to nothing more than a simulation of intelligence. And we know that because mere mimicry is precisely all that computer architectures are designed to do.

Here’s an analogy (which I have developed in more detail elsewhere). The methods employed by entertainers such as David Copperfield, David Blaine, and Penn and Teller are designed to produce effects that merely simulate magic. And no matter how well they work, we know that mere simulation is all they can achieve, because the means they use are in no way preternatural but entirely mundane–sleight of hand, illusions, and so on. Of course, genuine magic is not real in the first place, but that is irrelevant. The point is that even if it were real, it would not be what Copperfield, Blaine, and Penn & Teller are doing, precisely because their methods are not of the type that could produce more than mimicry.

Now, AI operates on an analogous principle. It is designed to produce effects that simulate intelligence by means that don’t require any actual intelligence on the part of the machines themselves. And this is as true of machine learning and related approaches that now dominate AI research as it was of the Turing machine model that dominated the earlier history of AI. To borrow some terminology from John Searle, AI algorithms of whatever kind are sensitive only to the “syntax” of the representations they process rather than their “semantics.” That is to say, they process representations in a way that is sensitive only to their physical properties rather than the meanings that we associate with those physical properties.

Because there is a general correlation between syntax and semantics–for example, the word “love” on a printed page is typically going to be used to express the concept LOVE–the output of a sophisticated algorithm can be made to simulate the sort of thing a human being might write. The correlation is not perfect. For instance, there might be cases where the string of letters “love” is not actually being used to express the concept LOVE, and there might be cases where the concept LOVE is being conveyed but without using the string of letters “love.” This is why AI programs can often be “tripped up” and reveal, through a failure to reflect such nuances, that they don’t actually understand what they are “saying.” Exposure to further data or refinements to an algorithm might work around such problems. But all that yields is a more convincing simulation (just as Penn and Teller and company may come up with new ways of producing ever more convincing illusions). It doesn’t somehow generate a grasp of meaning or semantics on the part of the machine, because its basic structure is in no way sensitive to that in the first place.

Our authors speak as if the dispute over whether AI models are genuinely intelligent or merely simulating intelligence is a matter of “inference to the best explanation”–as if their position and that of the AI skeptic are alternative hypotheses meant to account for the same evidence, and the controversy has to do with which view is more empirically adequate, more parsimonious, and so forth.

But this is as silly as suggesting that the view that Penn and Teller are merely simulating magic rather than producing the real thing is being proposed as the “best explanation” of the “evidence.” It is, of course, nothing of the kind. It is just a simple and straightforward conceptual point about the nature of their methods. And to note that AI produces only a simulation of intelligence rather than the real McCoy is an equally simple and straightforward conceptual point about the nature of its methods.

Interested readers will find a more detailed and academic treatment of these issues in Chapter 9 of Immortal Souls.

(Editor’s note: This essay was posted first, in slightly different form, on the author’s personal site and is reposted here with kind permission.)


If you value the news and views Catholic World Report provides, please consider donating to support our efforts. Your contribution will help us continue to make CWR available to all readers worldwide for free, without a subscription. Thank you for your generosity!

Click here for more information on donating to CWR. Click here to sign up for our newsletter.


About Dr. Edward Feser 60 Articles
Edward Feser is the author of several books on philosophy and morality, including All One in Christ: A Catholic Critique of Racism and Critical Race Theory (Ignatius Press, August 2022), and Five Proofs of the Existence of God and is co-author of By Man Shall His Blood Be Shed: A Catholic Defense of Capital Punishment, both also published by Ignatius Press.

10 Comments

  1. This piece is poorly argued. It’s guilty of the same question-begging that it (rightly( criticizes the Nature authors of. He doesn’t offer any evidence in support of his claim that AI merely simulates intelligence, and offers us no meaningful way of distinguishing between real intelligence and very sophisticated simulations of it. His observation that “AI programs can often be “tripped up and reveal that they don’t actually understand what they are saying” is trivial, as the exact same thing can be said of human beings.

    • I’ll make it simple for you, as someone who worked for a time designing CPUs. Computers possess no intelligence at all and never will. They are simple devices that cannot “think” any more than a light bulb can contemplate the beauty of a work of art it illustrates. Billions of layers of very simple contingent electrical circuits might make the whole device seem very complicated, but it’s not.

      A capacity for value judgements and willingness for authentic love through sacrifice is what distinguishes God’s creatures. on one scale below angels, from everything else. How much closer is a computer to judging innate truth, natural law, the sacredness of life, or the origin of human rights as divine endowments and not political inventions, than a pile of scrap metal in a junkyard? Devaluing judgments count too. Can a computer perform a sinful act or be trapped in mental disturbance. Name the computer that will “decide” to perform an act of emotionally desperate suicide. Name the computer that will destroy itself to save a life.

      • People would not be spending the money on AI that they actually are spending to produce something that experiences “authentic love through self sacrifice” — and, frankly, though that comes close to a discussion of the LIKENESS of God (which is marred by sin), it really has nothing to do with intelligence. No; the claim that AI might achieve human-level intelligence has to do with WHAT PROBLEMS CAN IT SOLVE. It’s obviously already good at some specialized tasks; Mathematica has been able to do calculus better than most math majors for decades, for example.

        Pretty much anything that can be done by a human being acting by routine can be done by AI, at least in principle. You probably have driven to work on “autopilot”; I certainly have. I would not presently trust an AI to drive me to work, but then there are a lot of people I would not trust, either. But it’s not just driving. Most of what we do requires an action of the brain, but not much action of the soul.

        The real issue comes when people try to attribute a soul to AI; some people now even reportedly think that their chatbot is some kind of god.

        This leads us to thorny questions, though, about animal souls and animal intelligence. Animal brains work, more or less, like human brains, and a dog can show “intelligence” comparable to a toddler. For generalized intelligence, I would still favor a dog, let alone a chimp, over the best current AI. But do dogs or chimps have souls, and what does that even mean? Aristotle would say they have nutritive and sensitive souls, but lack rational souls — but Aristotle’s physics was pretty bad, so should we trust him? Most churchmen have, but the Church does not authoritatively teach whether, for example, your dog has an undying soul.

        • “Most of what we do requires an action of the brain, but not much action of the soul.”

          Everything we do requires a value judgment, even a decision to scratch an itch or not scratch an itch. Everything we do is a condition of the soul.

          And if you believe animals do not possess rational intelligence, it only proves you never played touch football with any of my dogs.
          Not to mention the absence of a soul could never know suffering.

          And the ability to process an equation has no connection whatsoever to conceptual thought in addition to the fact that the appliance that processes an equation has no memory of having done so immediately or any time thereafter. Memory in computers is just a rather dumb metaphor.

  2. In other news, Adolf Hitler was not *the* Antichrist. This truth would, however, be little comfort to his millions of victims. He was not *the* Antichrist, but he was antichrist enough.

    Likewise, the sad truth is that many (most?) jobs do not require human-level intelligence. Very, very few require the image and likeness of God.

  3. Excellant!

    I have a modest background in computers and software development. Mr. Baker’s point is spot on. A computer is a Turing machine and Turing machines are deterministic. Ie, given the same input they will always produce the same output.

    All computer programs can be reduced to 3 basic logical operators (AND, OR, NOT) and a way to go to the next instruction. AI is no exception.

    An AI program must also be deterministic because it is implemented on a deterministic machine: software requires hardware. Granted the input to AI is very complex (the whole internet!) and changing constantly. But I maintain that if that input could be completely recorded and somehow played back to the AI it would produce the same results every time. There are two ways for this to not happen: the machine breaks or the AI has access to true random numbers. It is well known that computer programs cannot compute true random numbers; this requires external input. But does fundamentally random behavior really qualify as intelligent?

    Many thanks for a well written and informative article from a perspective in which I am poorly versed.

  4. For an outstanding treatment of this topic, read Richard Cain’s ‘Argument against the metaphysical possibility of artificial computational intelligence.’ It’s online. Quite brilliant.

  5. A.I. is, and always has been, a marketing slogan. These LLMs are essentially patter-matching, next-word guessers. Even the “thinking” models don’t actually think; they compose a stream of pseudo-conscience ramble that predicts-the-next-word as such length that it ends up simulating thought by moving from one point to another in the least thoughtful, and most inefficient, way possible. The A.I., at no point, understands what it is saying. It simply very accurately predicts which words a human being will use one after the other, which it can do because its training data, at this point, contains practically every text ever written by a human being. Anybody who thinks it is intelligent probably also thinks that a dog or a pig is intelligent, simply because it can do tasks of some complexity. Last I heard, when we called somebody “intelligent,” it had to do with their capacity to THINK, and not their capacity to shuffle about and manufacture widgets or enter data; these things, as Edward Feser rightly argues, flow from our intelligence, but are not intelligence itself. We will approach somebody when they say “come here,” which a dog will as well. The dog is not intelligent, because it does not think, comprehend, etc. We, on the other hand, doing that simple task, do think. The A.I., however, never thinks. It plays a glorified game of word association, whether with the user, or with itself in “thinking” mode. So happy that this is being written about here. The older generation in particular simply does not grasp the notion that a sufficiently advance computer can speak and do things without understanding them, and this leads them to believe that A.I. can ever be truly reliable, when, of course, that which does not understand what it is doing, can never be fully reliable. There’s a reason the bubble is popping. We’ve been lied to by marketers, who get louder as IPOs approach.

Leave a Reply

Your email address will not be published.

All comments posted at Catholic World Report are moderated. While vigorous debate is welcome and encouraged, please note that in the interest of maintaining a civilized and helpful level of discussion, comments containing obscene language or personal attacks—or those that are deemed by the editors to be needlessly combative or inflammatory—will not be published. Thank you.


*