“Sing, O Goddess, the anger of Achilles son of Peleus, that brought countless ills upon the Achaeans.” In these opening lines of the Iliad, Homer invokes the Muse. For Homer the act of creating poetry is a divine gift, one that derives from an otherworldly source and is not ultimately reducible to this world. This conception of human creativity as a divine gift pervaded the ancient world, and was also evident among the Hebrews. In Exodus, for instance, we read that God filled the two artisans Bezaleel and Aholiab with wisdom so that they might complete the work of the tabernacle.
The idea that creative activity is a divine gift has largely been lost these days. To ask a cognitive scientist, for instance, what made Mozart a creative genius is unlikely to issue in an appeal to God. If the cognitive scientist embraces neuropsychology, he may suggest that Mozart was blessed with a particularly fortunate collocation of neurons. If he prefers an information processing model of mentality, he may attribute Mozart’s genius to some particularly effective computational modules. If he is taken with Skinner’s behaviorism, he may attribute Mozart’s genius to some particularly effective reinforcement schedules (perhaps imposed early in his life by his father Leopold). And no doubt, in all of these explanations the cognitive scientist will invoke Mozart’s natural genetic endowment. In place of a divine afflatus, the modern cognitive scientist explains human creativity purely in terms of natural processes.
Who’s right, the ancients or the moderns? My own view is that the ancients got it right. An act of creation is always a divine gift and cannot be reduced to purely naturalistic categories. To be sure, creative activity often involves the transformation of natural objects, like the transformation of a slab of marble into Michelangelo’s David. But even when confined to natural objects, creative activity is never naturalistic without remainder. The divine is always present at some level and indispensable.
Invoking the divine to explain an act of creation is, of course, wholly unacceptable to the ruling intellectual elite. Naturalism, the view that nature is the ultimate reality, has become the default position for all serious inquiry among our intellectual elite. From Biblical studies to law to education to science to the arts, inquiry is allowed to proceed only under the supposition that nature is the ultimate reality. Naturalism denies any divine element to the creative act. By contrast, the Christian tradition plainly asserts that God is the ultimate reality and that nature itself is a divine creative act. Within Christian theism, God is primary and fundamental whereas nature is secondary and derivative. Naturalism, by contrast, asserts that nature is primary and fundamental.
Theism and naturalism provide radically different perspectives on the act of creation. Within theism any act of creation is also a divine act. Within naturalism any act of creation emerges from a purely natural substrate — the very minds that create are, within naturalism, the result of a long evolutionary process that itself was not created. The aim of this talk, then, is to present a general account of creation that is faithful to the Christian tradition, that resolutely rejects naturalism, and that engages contemporary developments in science and philosophy.
The Challenge of Naturalism
Why should anyone want to understand the act of creation naturalistically? Naturalism, after all, offers fewer resources than theism. Naturalism simply gives you nature. Theism gives you not only nature, but also God and anything outside of nature that God might have created. The ontology of theism is far richer than that of naturalism. Why, then, settle for less?
Naturalists do not see themselves as settling for less. Instead, they regard theism as saddled with a lot of extraneous entities that serve no useful function. The regulative principle of naturalism is Occam’s razor. Occam’s razor is a principle of parsimony that requires eliminating entities that perform no useful function. Using Occam’s razor, naturalists attempt to slice away the superstitions of the past-and for naturalists the worst superstition of all is God. People used to invoke God to explain all sorts of things for which we now have perfectly good naturalistic explanations. Accordingly, God is a superstition that needs to be excised from our understanding of the world. The naturalists’ dream is to invent a theory of everything that entirely eliminates the need for God (Stephen Hawking is a case in point).
Since naturalists are committed to eliminating God from every domain of inquiry, let us consider how successfully they have eliminated God from the act of creation. Even leaving aside the creation of the world and focusing solely on human acts of creation, do we find that naturalistic categories have fully explained human creativity? Occam’s razor is all fine and well for removing stubble, but while we’re at it let’s make sure we don’t lop off a nose or ear. With respect to human creativity, let’s make sure that in eliminating God the naturalist isn’t giving us a lobotomized account of human creativity. Einstein once remarked that everything should be made as simple as possible but not simpler. In eliminating God from the act of creation, the naturalist needs to make sure that nothing of fundamental importance has been lost. Not only has the naturalist failed to provide this assurance, but there is good reason to think that any account of the creative act that omits God is necessarily incomplete and defective.
What does naturalism have to say about human acts of creation? For the moment let’s bracket the question of creativity and consider simply what it is for a human being to act. Humans are intelligent agents that act with intentions to accomplish certain ends. Although some acts by humans are creative, others are not. Georgia O’Keefe painting an iris is a creative act. Georgia O’Keefe flipping on a light switch is an act but not a creative act. For the moment, therefore, let us focus simply on human agency, leaving aside human creative agency.
How, then, does naturalism make sense of human agency? Although the naturalistic literature that attempts to account for human agency is vast, the naturalist’s options are in fact quite limited. The naturalist’s world is not a mind-first world. Intelligent agency is therefore in no sense prior to or independent of nature. Intelligent agency is neither sui generis nor basic. Intelligent agency is a derivative mode of causation that depends on underlying naturalistic — and therefore unintelligent — causes. Humans agency in particular supervenes on underlying natural processes, which in turn usually are identified with brain function.
It is important to distinguish the naturalist’s understanding of causation from the theist’s. Within theism God is the ultimate reality. Consequently, whenever God acts, there can be nothing outside of God that compels God’s action. God is not a billiard ball that must move when another billiard ball strikes it. God’s actions are free, and though he responds to his creation, he does not do so out of necessity. Within theism, therefore, divine action is not reducible to some more basic mode of causation. Indeed, within theism divine action is the most basic mode of causation since any other mode of causation involves creatures which themselves were created in a divine act.
Now consider naturalism. Within naturalism nature is the ultimate reality. Consequently, whenever something happens in nature, there can be nothing outside of nature that shares responsibility for what happened. Thus, when an event happens in nature, it is either because some other event in nature was responsible for it or because it simply happened, apart from any other determining event. Events therefore happen either because they were caused by other events or because they happened spontaneously. The first of these is usually called “necessity,” the second “chance.” For the naturalist chance and necessity are the fundamental modes of causation. Together they constitute what are called “natural causes.” Naturalism, therefore, seeks to account for intelligent agency in terms of natural causes.
How well have natural causes been able to account for intelligent agency? Cognitive scientists have achieved nothing like a full reduction. The French Enlightenment thinker Pierre Cabanis once remarked: “Les nerfs-voilà tout l’homme” (the nerves — that’s all there is to man). A full reduction of intelligent agency to natural causes would give a complete account of human behavior, intention, and emotion in terms of neural processes. Nothing like this has been achieved. No doubt, neural processes are correlated with behavior, intention, and emotion. Anger presumably is correlated with certain localized brain excitations. But localized brain excitations hardly explain anger any better than do overt behaviors associated with anger — like shouting obscenities.
Because cognitive scientists have yet to effect a full reduction of intelligent agency to natural causes, they speak of intelligent agency as supervening on natural causes. Supervenience is a hierarchical relationship between higher order processes (in this case intelligent agency) and lower order processes (in this case natural causes). What supervenience says is that the relationship between the higher and lower order processes is a one-way street, with the lower determining the higher. To say, for instance, that intelligent agency supervenes on neurophysiology is to say that once all the facts about neurophysiology are in place, all the facts about intelligent agency are determined as well. Supervenience makes no pretense at reductive analysis. It simply asserts that the lower level determines the higher level — how it does it, we don’t know.
Supervenience is therefore an insulating strategy, designed to protect a naturalistic account of intelligent agency until a full reductive explanation is found. Supervenience, though not providing a reduction, tells us that in principle a reduction exists. Given that nothing like a full reductive explanation of intelligent agency is at hand, why should we think that such a reduction is even possible? To be sure, if we knew that naturalism were correct, then supervenience would follow. But naturalism itself is at issue.
Neuroscience, for instance, is nowhere near achieving its ambitions, and that despite its strident rhetoric. Hardcore neuroscientists, for instance, refer disparagingly to the ordinary psychology of beliefs, desires, and emotions as “folk psychology.” The implication is that just as “folk medicine” had to give way to “real medicine,” so “folk psychology” will have to give way to a revamped psychology that is grounded in neuroscience. In place of talking cures that address our beliefs, desires, and emotions, tomorrow’s healers of the soul will manipulate brains states directly and ignore such outdated categories as beliefs, desires, and emotions.
At least so the story goes. Actual neuroscience research has yet to keep pace with its vaulting ambition. That should hardly surprise us. The neurophysiology of our brains is incredibly plastic and has proven notoriously difficult to correlate with intentional states. For instance, Louis Pasteur, despite suffering a cerebral accident, continued to enjoy a flourishing scientific career. When his brain was examined after he died, it was discovered that half the brain had completely atrophied. How does one explain a flourishing intellectual life despite a severely damaged brain if mind and brain coincide?
Or consider a still more striking example. The December 12th, 1980 issue of Science contained an article by Roger Lewin titled “Is Your Brain Really Necessary?” In the article, Lewin reported a case study by John Lorber, a British neurologist and professor at Sheffield University. I quote from the article:
“There’s a young student at this university,” says Lorber, “who has an IQ of 126, has gained a first-class honors degree in mathematics, and is socially completely normal. And yet the boy has virtually no brain.” [Lewin continues:] The student’s physician at the university noticed that the youth had a slightly larger than normal head, and so referred him to Lorber, simply out of interest. “When we did a brain scan on him,” Lorber recalls, “we saw that instead of the normal 4.5-centimeter thickness of brain tissue between the ventricles and the cortical surface, there was just a thin layer of mantle measuring a millimeter or so. His cranium is filled mainly with cerebrospinal fluid.”
Against such anomalies, Cabanis’s dictum, “the nerves — that’s all there is to man,” hardly inspires confidence. Yet as Thomas Kuhn has taught us, a science that is progressing fast and furiously is not about to be derailed by a few anomalies. Neuroscience is a case in point. For all the obstacles it faces in trying to reduce intelligent agency to natural causes, neuroscience persists in the Promethean determination to show that mind does ultimately reduce to neurophysiology. Absent a prior commitment to naturalism, this determination will seem misguided. On the other hand, given a prior commitment to naturalism, this determination is readily understandable.
Understandable yes, obligatory no. Most cognitive scientists do not rest their hopes with neuroscience. Yes, if naturalism is correct, then a reduction of intelligent agency to neurophysiology is in principle possible. The sheer difficulty of even attempting this reduction, both experimental and theoretical, however, leaves many cognitive scientists looking for a more manageable field to invest their energies. As it turns out, the field of choice is computer science, and especially its subdiscipline of artificial intelligence (abbreviated AI). Unlike brains, computers are neat and precise. Also, unlike brains, computers and their programs can be copied and mass-produced. Inasmuch as science thrives on replicability and control, computer science offers tremendous practical advantages over neurological research.
Whereas the goal of neuroscience is to reduce intelligent agency to neurophysiology, the goal of artificial intelligence is to reduce intelligent agency to computer algorithms. Since computers operate deterministically, reducing intelligent agency to computer algorithms would indeed constitute a naturalistic reduction of intelligent agency. Should artificial intelligence succeed in reducing intelligent agency to computation, cognitive scientists would still have the task of showing in what sense brain function is computational (alternatively, Marvin Minsky’s dictum “the mind is a computer made of meat” would still need to be verified). Even so, the reduction of intelligent agency to computation would go a long way toward establishing a purely naturalistic basis for human cognition.
An obvious question now arises: Can computation explain intelligent agency? First off, let’s be clear that no actual computer system has come anywhere near to simulating the full range of capacities we associate with human intelligent agency. Yes, computers can do certain narrowly circumscribed tasks exceedingly well (like play chess). But require a computer to make a decision based on incomplete information and calling for common sense, and the computer will be lost. Perhaps the toughest problem facing artificial intelligence researchers is what’s called the frame problem. The frame problem is getting a computer to find the appropriate frame of reference for solving a problem.
Consider, for instance, the following story: A man enters a bar. The bartender asks, “What can I do for you?” The man responds, “I’d like a glass of water.” The bartender pulls out a gun and shouts, “Get out of here!” The man says “thank you” and leaves. End of story. What is the appropriate frame of reference? No, this isn’t a story by Franz Kafka. The key item of information needed to make sense of this story is this: The man has the hiccups. By going to the bar to get a drink of water, the man hoped to cure his hiccups. The bartender, however, decided on a more radical cure. By terrifying the man with a gun, the bartender cured the man’s hiccups immediately. Cured of his hiccups, the man was grateful and left. Humans are able to understand the appropriate frame of reference for such stories immediately. Computers, on the other hand, haven’t a clue.
Ah, but just wait. Give an army of clever programmers enough time, funding, and computational power, and just see if they don’t solve the frame problem. Naturalists are forever issuing such promissory notes, claiming that a conclusive confirmation of naturalism is right around the corner — just give our scientists a bit more time and money. John Polkinghorne refers to this practice as “promissory materialism.”
Confronted with such promises, what’s a theist to do? To refuse such promissory notes provokes the charge of obscurantism, but to accept them means suspending one’s theism. It is possible to reject promissory materialism without meriting the charge of obscurantism. The point to realize is that a promissory note need only be taken seriously if there is good reason to think that it can be paid. The artificial intelligence community has thus far offered no compelling reason for thinking that it will ever solve the frame problem. Indeed, computers that employ common sense to determine appropriate frames of reference continue utterly to elude computer scientists.
Given the practical difficulties of producing a computer that faithfully models human cognition, the hardcore artificial intelligence advocate can change tactics and argue on theoretical grounds that humans are simply disguised computers. The argument runs something like this. Human beings are finite. Both the space of possible human behaviors and the space of possible sensory inputs are finite. For instance, there are only so many distinguishable word combinations that we can utter and only so many distinguishable sound combinations that can strike our eardrums. When represented mathematically, the total number of human lives that can be distinguished empirically is finite. Now it is an immediate consequence of recursion theory (the mathematical theory that undergirds computer science) that any operations and relations on finite sets are computable. It follows that human beings can be represented computationally. Humans are therefore functionally equivalent to computers. QED.
This argument can be nuanced. For instance, we can introduce a randomizing element into our computations to represent quantum indeterminacy. What’s important here, however, is the gist of the argument. The argument asks us to grant that humans are essentially finite. Once that assumption is granted, recursion theory tells us that everything a finite being does is computable. We may never actually be able to build the machines that render us computable. But in principle we could given enough memory and fast enough processors.
It’s at this point that opponents of computational reductionism usually invoke Gödel’s incompleteness theorem. Gödel’s theorem is said to refute computational reductionism by showing that humans can do things that computers cannot — namely, produce a Gödel sentence. John Lucas made such an argument in the early 1960s, and his argument continues to be modified and revived. Now it is perfectly true that humans can produce Gödel sentences for computational systems external to themselves. But computers can as well be programmed to compute Gödel sentences for computational systems external to themselves. This point is seldom appreciated, but becomes evident from recursion-theoretic proofs of Gödel’s theorem (see, for example, Klaus Weihrauch’s Computability).
The problem, then, is not to find Gödel sentences for computational systems external to oneself. The problem is for an agent to examine oneself as a computational system and therewith produce one’s own Gödel sentence. If human beings are non-computational, then there won’t be any Gödel sentence to be found. If, on the other hand, human beings are computational, then, by Gödel’s theorem, we won’t be able to find our own Gödel sentences. And indeed, we haven’t. Our inability to translate neurophysiology into computation guarantees that we can’t even begin computing our Gödel sentences if indeed we are computational systems. Yes, for a computational system laid out before us we can determine its Gödel sentence. Nevertheless, we don’t have sufficient access to ourselves to lay ourselves out before ourselves and thereby determine our Gödel sentences. It follows that neither Gödel’s theorem nor our ability to prove Gödel’s theorem shows that humans can do things that computers cannot.
Accordingly, Gödel’s theorem fails to refute the argument for computational reductionism based on human finiteness. To recap that argument, humans are finite because the totality of their possible behavioral outputs and possible sensory inputs is finite. Moreover, all operations and relations on finite sets are by recursion theory computable. Hence, humans are computational systems. This is the argument. What are we to make of it? Despite the failure of Gödel’s theorem to block its conclusion, is there a flaw in the argument?
Yes there is. The flaw consists in identifying human beings with their behavioral outputs and sensory inputs. Alternatively, the flaw consists in reducing our humanity to what can be observed and measured. We are more than what can be observed and measured. Once, however, we limit ourselves to what can be observed and measured, we are necessarily in the realm of the finite and therefore computable. We can only make so many observations. We can only take so many measurements. Moreover, our measurements never admit infinite gradations (indeed, there’s always some magnitude below which quantities become empirically indistinguishable). Our empirical selves are therefore essentially finite. It follows that unless our actual selves transcend our empirical selves, our actual selves will be finite as well — and therefore computational.
Roger Penrose understands this problem. In The Emperor’s New Mind and in his more recent Shadows of the Mind, he invokes quantum theory to underwrite a non-computational view of brain and mind. Penrose’s strategy is the same that we saw for Gödel’s theorem: Find something humans can do that computers can’t. There are plenty of mathematical functions that are non-computable. Penrose therefore appeals to quantum processes in the brain whose mathematical characterization employs non-computable functions.
Does quantum theory offer a way out of computational reductionism? I would say no. Non-computable functions are an abstraction. To be non-computable, functions have to operate on infinite sets. The problem, however, is that we have no observational experience of infinite sets or of the non-computable functions defined on them. Yes, the mathematics of quantum theory employs non-computable functions. But when we start plugging in concrete numbers and doing calculations, we are back to finite sets and computable functions.
Granted, we may find it convenient to employ non-computable functions in characterizing some phenomenon. But when we need to say something definite about the phenomenon, we must supply concrete numbers, and suddenly we are back in the realm of the computable. Non-computability exists solely as a mathematical abstraction — a useful abstraction, but an abstraction nonetheless. Precisely because our behavioral outputs and sensory inputs are finite, there is no way to test non-computability against experience. All scientific data are finite, and any mathematical operations we perform on that data are computable. Non-computable functions are therefore always dispensable, however elegant they may appear mathematically.
There is, however, still a deeper problem with Penrose’s program to eliminate computational reductionism. Suppose we could be convinced that there are processes in the brain that are non-computational. For Penrose they are quantum processes, but whatever form they take, as long as they are natural processes, we are still dealing with a naturalistic reduction of mind. Computational reductionism is but one type of naturalistic reductionism — certainly the most extreme, but by no means the only one. Penrose’s program offers to replace computational processes with quantum processes. Quantum processes, however, are as fully naturalistic as computational processes. In offering to account for mind in terms of quantum theory, Penrose is therefore still wedded to a naturalistic reduction of mind and intelligent agency.
It’s time to ask the obvious question: Why should anyone want to make this reduction? Certainly, if we have a prior commitment to naturalism, we will want to make it. But apart from that commitment, why attempt it? As we’ve seen, neurophysiology hasn’t a clue about how to reduce intelligent agency to natural causes (hence its continued retreat to concepts like supervenience, emergence, and hierarchy — concepts which merely cloak ignorance). We’ve also seen that no actual computational systems show any sign of reducing intelligent agency to computation. The argument that we are computational systems because the totality of our possible behavioral outputs and possible sensory inputs is finite holds only if we presuppose that we are nothing more than the sum of those behavioral outputs and sensory inputs. So too, Penrose’s argument that we are naturalistic systems because some well-established naturalistic theory (in this case quantum theory) characterizes our neurophysiology holds only if the theory does indeed accurately characterize our neurophysiology (itself a dubious claim given the frequency with which scientific theories are overturned) and so long as we presuppose that we are nothing more than a system characterized by some naturalistic theory.
Bottom line: The naturalistic reduction of intelligent agency is not the conclusion of an empirically-based evidential argument, but merely a straightforward consequence of presupposing naturalism in the first place. Indeed, the empirical evidence for a naturalistic reduction of intelligent agency is wholly lacking. For instance, nowhere does Penrose write down the Schroedinger equation for someone’s brain, and then show how actual brain states agree with brain states predicted by the Schroedinger equation. Physicists have a hard enough time writing down the Schroedinger equation for systems of a few interacting particles. Imagine the difficulty of writing down the Schroedinger equation for the multi-billion neurons that constitute each of our brains. It ain’t going to happen. Indeed, the only thing these naturalistic reductions of intelligent agency have until recently had in their favor is Occam’s razor. And even this naturalistic mainstay is proving small comfort. Indeed, recent developments in the theory of intelligent design are showing that intelligent agency cannot be reduced to natural causes. Let us now turn to these developments.
The Resurgence of Design
In arguing against computational reductionism, both John Lucas and Roger Penrose attempted to find something humans can do that computers cannot. For Lucas, it was to construct a Gödel sentence. For Penrose, it was finding in neurophysiology a non-computational quantum process. Neither of these refutations succeed against computational reductionism, much less against a general naturalistic reduction of intelligent agency. Nevertheless, the strategy underlying these attempted refutations is sound, namely, to find something intelligent agents can do that natural causes cannot. We don’t have to look far. All of us attribute things to intelligent agents that we wouldn’t dream of attributing to natural causes. For instance, natural causes can throw scrabble pieces on a board, but cannot arrange the pieces into meaningful sentences. To obtain a meaningful arrangement requires an intelligent agent.
This intuition, that natural causes are too stupid to do the things that intelligent agents are capable of, has underlain the design arguments of past centuries. Throughout the centuries theologians have argued that nature exhibits features which nature itself cannot explain, but which instead require an intelligence that transcends nature. From Church fathers like Minucius Felix and Basil the Great (third and fourth centuries) to medieval scholastics like Moses Maimonides and Thomas Aquinas (twelfth and thirteenth centuries) to reformed thinkers like Thomas Reid and Charles Hodge (eighteenth and nineteenth centuries), we find theologians making design arguments, arguing from the data of nature to an intelligence operating over and above nature.
Design arguments are old hat. Indeed, design arguments continue to be a staple of philosophy and religion courses. The most famous of the design arguments is William Paley’s watchmaker argument. According to Paley, if we find a watch in a field, the watch’s adaptation of means to ends (that is, the adaptation of its parts to telling time) ensures that it is the product of an intelligence, and not simply the output of undirected natural processes. So too, the marvelous adaptations of means to ends in organisms, whether at the level of whole organisms, or at the level of various subsystems (Paley focused especially on the mammalian eye), ensure that organisms are the product of an intelligence.
Though intuitively appealing, Paley’s argument had until recently fallen into disuse. This is now changing. In the last five years design has witnessed an explosive resurgence. Scientists are beginning to realize that design can be rigorously formulated as a scientific theory. What has kept design outside the scientific mainstream these last hundred and forty years is the absence of a precise criterion for distinguishing intelligent agency from natural causes. For design to be scientifically tenable, scientists have to be sure they can reliably determine whether something is designed. Johannes Kepler, for instance, thought the craters on the moon were intelligently designed by moon dwellers. We now know that the craters were formed naturally. It’s this fear of falsely attributing something to design only to have it overturned later that has prevented design from entering science proper. With a precise criterion for discriminating intelligently from unintelligently caused objects, scientists are now able to avoid Kepler’s mistake.
Before examining this criterion, I want to offer a brief clarification about the word “design.” I’m using “design” in three distinct senses. First, I use it to denote the scientific theory that distinguishes intelligent agency from natural causes, a theory that increasingly is being referred to as “design theory” or “intelligent design theory” (IDT). Second, I use “design” to denote what it is about intelligently produced objects that enables us to tell that they are intelligently produced and not simply the result of natural causes. When intelligent agents act, they leave behind a characteristic trademark or signature. The scholastics used to refer to the “vestiges of creation.” The Latin vestigium means footprint. It was thought that God, though not physically present, left his footprints throughout creation. Hugh Ross has referred to the “fingerprint of God.” It is “design” in this sense — as a trademark, signature, vestige, or fingerprint — that our criterion for discriminating intelligently from unintelligently caused objects is meant to identify. Lastly, I use “design” to denote intelligent agency itself. Thus, to say that something is designed is to say that an intelligent agent caused it.
Let us now turn to my advertised criterion for discriminating intelligently from unintelligently caused objects. Although a detailed treatment of this criterion is technical and appears in my book The Design Inference, the basic idea is straightforward and easily illustrated. Consider how the radio astronomers in the movie Contact detected an extra-terrestrial intelligence. This movie, which came out last summer and was based on a novel by Carl Sagan, was an enjoyable piece of propaganda for the SETI research program — the Search for Extra-Terrestrial Intelligence. To make the movie interesting, the SETI researchers had to find an extra-terrestrial intelligence (the actual SETI program has yet to be so fortunate).
How, then, did the SETI researchers in Contact find an extra-terrestrial intelligence? To increase their chances of finding an extra-terrestrial intelligence, SETI researchers have to monitor millions of radio signals from outer space. Many natural objects in space produce radio waves. Looking for signs of design among all these naturally produced radio signals is like looking for a needle in a haystack. To sift through the haystack, SETI researchers run the signals they monitor through computers programmed with pattern-matchers. So long as a signal doesn’t match one of the pre-set patterns, it will pass through the pattern-matching sieve. If, on the other hand, it does match one of those patterns, then, depending on the pattern matched, the SETI researchers may have cause for celebration.
The SETI researchers in Contact did find a signal worthy of celebration, namely the sequence of prime numbers from 2 to 101, represented as a series of beats and pauses (2 = beat-beat-pause; 3 = beat-beat-beat-pause; 5 = beat-beat-beat-beat-beat-pause; etc.). The SETI researchers in Contact took this signal as decisive confirmation of an extra-terrestrial intelligence. What is it about this signal that warrants us inferring design? Whenever we infer design, we must establish two things — complexity and specification. Complexity ensures that the object in question is not so simple that it can readily be explained by natural causes. Specification ensures that this object exhibits the type of pattern that is the signature of intelligence.
To see why complexity is crucial for inferring design, consider what would have happened if the SETI researchers had simply witnessed a single prime number — say the number 2 represented by two beats followed by a pause. It is a sure bet that no SETI researcher, if confronted with this three-bit sequence (beat-beat-pause), is going to contact the science editor at the New York Times, hold a press conference, and announce that an extra-terrestrial intelligence has been discovered. No headline is going to read, “Aliens Master the Prime Number Two!”
The problem is that two beats followed by a pause is too short a sequence (that is, has too little complexity) to establish that an extra-terrestrial intelligence with knowledge of prime numbers produced it. A randomly beating radio source might by chance just happen to output the sequence beat-beat-pause. The sequence of 1126 beats and pauses required to represent the prime numbers from 2 to 101, however, is a different story. Here the sequence is sufficiently long (that is, has enough complexity) to confirm that an extra-terrestrial intelligence could have produced it.
Even so, complexity by itself isn’t enough to eliminate natural causes and detect design. If I flip a coin 1000 times, I’ll participate in a highly complex (or what amounts to the same thing, highly improbable) event. Indeed, the sequence I end up flipping will be one of 10300 possible sequences. This sequence, however, won’t trigger a design inference. Though complex, it won’t exhibit a pattern characteristic of intelligence. In contrast, consider the sequence of prime numbers from 2 to 101. Not only is this sequence complex, but it also constitutes a pattern characteristic of intelligence. The SETI researcher who in the movie Contact first noticed the sequence of prime numbers put it this way: “This isn’t noise, this has structure.”
What makes a pattern characteristic of intelligence and therefore suitable for detecting design? The basic intuition distinguishing patterns that alternately succeed or fail to detect design is easily motivated. Consider the case of an archer. Suppose an archer stands fifty meters from a large wall with bow and arrow in hand. The wall, let’s say, is sufficiently large that the archer cannot help but hit it. Now suppose each time the archer shoots an arrow at the wall, the archer paints a target around the arrow so that the arrow sits squarely in the bull’s-eye. What can be concluded from this scenario? Absolutely nothing about the archer’s ability as an archer. Yes, a pattern is being matched; but it is a pattern fixed only after the arrow has been shot. The pattern is thus purely ad hoc.
But suppose instead the archer paints a fixed target on the wall and then shoots at it. Suppose the archer shoots a hundred arrows, and each time hits a perfect bull’s-eye. What can be concluded from this second scenario? Confronted with this second scenario we are obligated to infer that here is a world-class archer, one whose shots cannot legitimately be referred to luck, but rather must be referred to the archer’s skill and mastery. Skill and mastery are of course instances of design.
The type of pattern where the archer fixes a target first and then shoots at it is common to statistics, where it is known as setting a rejection region prior to an experiment. In statistics, if the outcome of an experiment falls within a rejection region, the chance hypothesis supposedly responsible for the outcome is rejected. Now a little reflection makes clear that a pattern need not be given prior to an event to eliminate chance and implicate design. Consider, for instance, a cryptographic text that encodes a message. Initially it looks like a random sequence of letters. Initially we lack any pattern for rejecting natural causes and inferring design. But as soon as someone gives us the cryptographic key for deciphering the text, we see the hidden message. The cryptographic key provides the pattern we need for detecting design. Moreover, unlike the patterns of statistics, it is given after the fact.
Patterns therefore divide into two types, those that in the presence of complexity warrant a design inference and those that despite the presence of complexity do not warrant a design inference. The first type of pattern I call a specification, the second a fabrication. Specifications are the non-ad hoc patterns that can legitimately be used to eliminate natural causes and detect design. In contrast, fabrications are the ad hoc patterns that cannot legitimately be used to detect design. The distinction between specifications and fabrications can be made with full statistical rigor.
Complexity and specification together yield a criterion for detecting design. I call it the complexity-specification criterion. According to this criterion, we reliably detect design in something whenever it is both complex and specified. To see why the complexity-specification criterion is exactly the right instrument for detecting design, we need to understand what it is about intelligent agents that makes them detectable in the first place. The principal characteristic of intelligent agency is choice. Whenever an intelligent agent acts, it chooses from a range of competing possibilities.
This is true not just of humans, but of animals as well as of extra-terrestrial intelligences. A rat navigating a maze must choose whether to go right or left at various points in the maze. When SETI researchers attempt to discover intelligence in the extra-terrestrial radio transmissions they are monitoring, they assume an extra-terrestrial intelligence could have chosen any number of possible radio transmissions, and then attempt to match the transmissions they observe with certain patterns as opposed to others. Whenever a human being utters meaningful speech, a choice is made from a range of possible sound-combinations that might have been uttered. Intelligent agency always entails discrimination, choosing certain things, ruling out others.
Given this characterization of intelligent agency, the crucial question is how to recognize it. Intelligent agents act by making a choice. How then do we recognize that an intelligent agent has made a choice? A bottle of ink spills accidentally onto a sheet of paper; someone takes a fountain pen and writes a message on a sheet of paper. In both instances ink is applied to paper. In both instances one among an almost infinite set of possibilities is realized. In both instances a contingency is actualized and others are ruled out. Yet in one instance we ascribe agency, in the other chance.
What is the relevant difference? Not only do we need to observe that a contingency was actualized, but we ourselves need also to be able to specify that contingency. The contingency must conform to an independently given pattern, and we must be able independently to formulate that pattern. A random ink blot is unspecifiable; a message written with ink on paper is specifiable. Ludwig Wittgenstein in Culture and Value made essentially the same point: “We tend to take the speech of a Chinese for inarticulate gurgling. Someone who understands Chinese will recognize language in what he hears. Similarly I often cannot discern the humanity in man.”
In hearing a Chinese utterance, someone who understands Chinese not only recognizes that one from a range of all possible utterances was actualized, but is also able to specify the utterance as coherent Chinese speech. Contrast this with someone who does not understand Chinese. In hearing a Chinese utterance, someone who does not understand Chinese also recognizes that one from a range of possible utterances was actualized, but this time, because lacking the ability to understand Chinese, is unable to specify the utterance as coherent speech.
To someone who does not understand Chinese, the utterance will appear gibberish. Gibberish — the utterance of nonsense syllables uninterpretable within any natural language — always actualizes one utterance from the range of possible utterances. Nevertheless, gibberish, by corresponding to nothing we can understand in any language, also cannot be specified. As a result, gibberish is never taken for intelligent communication, but always for what Wittgenstein calls “inarticulate gurgling.”
This actualizing of one among several competing possibilities, ruling out the rest, and specifying the one that was actualized encapsulates how we recognize intelligent agency, or equivalently, how we detect design. Experimental psychologists who study animal learning and behavior have known this all along. To learn a task an animal must acquire the ability to actualize behaviors suitable for the task as well as the ability to rule out behaviors unsuitable for the task. Moreover, for a psychologist to recognize that an animal has learned a task, it is necessary not only to observe the animal making the appropriate discrimination, but also to specify this discrimination.
Thus to recognize whether a rat has successfully learned how to traverse a maze, a psychologist must first specify which sequence of right and left turns conducts the rat out of the maze. No doubt, a rat randomly wandering a maze also discriminates a sequence of right and left turns. But by randomly wandering the maze, the rat gives no indication that it can discriminate the appropriate sequence of right and left turns for exiting the maze. Consequently, the psychologist studying the rat will have no reason to think the rat has learned how to traverse the maze.
Only if the rat executes the sequence of right and left turns specified by the psychologist will the psychologist recognize that the rat has learned how to traverse the maze. Now it is precisely the learned behaviors we regard as intelligent in animals. Hence it is no surprise that the same scheme for recognizing animal learning recurs for recognizing intelligent agency generally, to wit: actualizing one among several competing possibilities, ruling out the others, and specifying the one chosen.
Note that complexity is implicit here as well. To see this, consider again a rat traversing a maze, but now take a very simple maze in which two right turns conduct the rat out of the maze. How will a psychologist studying the rat determine whether it has learned to exit the maze. Just putting the rat in the maze will not be enough. Because the maze is so simple, the rat could by chance just happen to take two right turns, and thereby exit the maze. The psychologist will therefore be uncertain whether the rat actually learned to exit this maze, or whether the rat just got lucky.
But contrast this now with a complicated maze in which a rat must take just the right sequence of left and right turns to exit the maze. Suppose the rat must take one hundred appropriate right and left turns, and that any mistake will prevent the rat from exiting the maze. A psychologist who sees the rat take no erroneous turns and in short order exit the maze will be convinced that the rat has indeed learned how to exit the maze, and that this was not dumb luck.
This general scheme for recognizing intelligent agency is but a thinly disguised form of the complexity-specification criterion. In general, to recognize intelligent agency we must observe a choice among competing possibilities, note which possibilities were not chosen, and then be able to specify the possibility that was chosen. What’s more, the competing possibilities that were ruled out must be live possibilities, and sufficiently numerous so that specifying the possibility that was chosen cannot be attributed to chance. In terms of complexity, this is just another way of saying that the range of possibilities is complex.
All the elements in this general scheme for recognizing intelligent agency (that is, choosing, ruling out, and specifying) find their counterpart in the complexity-specification criterion. It follows that this criterion formalizes what we have been doing right along when we recognize intelligent agency. The complexity-specification criterion pinpoints what we need to be looking for when we detect design.
The implications of the complexity-specification criterion are profound, not just for science, but also for philosophy and theology. The power of this criterion resides in its generality. It would be one thing if the criterion only detected human agency. But as we’ve seen, it detects animal and extra-terrestrial agency as well. Nor is it limited to intelligent agents that belong to the physical world. The fine-tuning of the universe, about which cosmologists make such a to-do, is both complex and specified and readily yields design. So too, Michael Behe’s irreducibly complex biochemical systems readily yield design. The complexity-specification criterion demonstrates that design pervades cosmology and biology. Moreover, it is a transcendent design, not reducible to the physical world. Indeed, no intelligent agent who is strictly physical could have presided over the origin of the universe or the origin of life.
Unlike design arguments of the past, the claim that transcendent design pervades the universe is no longer a strictly philosophical or theological claim. It is also a fully scientific claim. There exists a reliable criterion for detecting design — the complexity-specification criterion. This criterion detects design strictly from observational features of the world. Moreover, it belongs to probability and complexity theory, not to metaphysics and theology. And although it cannot achieve logical demonstration, it is capable of achieving statistical justification so compelling as to demand assent. When applied to the fine-tuning of the universe and the complex, information-rich structures of biology, it demonstrates a design external to the universe. In other words, the complexity-specification criterion demonstrates transcendent design.
This is not an argument from ignorance. Just as physicists reject perpetual motion machines because of what they know about the inherent constraints on energy and matter, so too design theorists reject any naturalistic reduction of specified complexity because of what they know about the inherent constraints on natural causes. Natural causes are too stupid to keep pace with intelligent causes. We’ve suspected this all along. Intelligent design theory provides a rigorous scientific demonstration of this longstanding intuition. Let me stress, the complexity-specification criterion is not a principle that comes to us demanding our unexamined acceptance — it is not an article of faith. Rather, it is the outcome of a careful and sustained argument about the precise interrelationships between necessity, chance, and design (for the details, please refer to my monograph The Design Inference).
Demonstrating transcendent design in the universe is a scientific inference, not a philosophical speculation. Once we understand the role of the complexity-specification criterion in warranting this inference, several things follow immediately: (1) Intelligent agency is logically prior to natural causation and cannot be reduced to it. (2) Intelligent agency is fully capable of making itself known against the backdrop of natural causes. (3) Any science that systematically ignores design is incomplete and defective. (4) Methodological naturalism, the view that science must confine itself solely to natural causes, far from assisting scientific inquiry actually stifles it. (5) The scientific picture of the world championed since the Enlightenment is not just wrong but massively wrong. Indeed, entire fields of inquiry, especially in the human sciences, will need to be rethought from the ground up in terms of intelligent design.
The Creation of the World
I want now to take stock and consider where we are in our study of the act of creation. In the phrase “act of creation,” so far I have focused principally on the first part of that phrase — the “act” part, or what I’ve also been calling “intelligent agency.” I have devoted much of my talk till now to contrasting intelligent agency with natural causes. In particular, I have argued that no empirical evidence supports the reduction of intelligent agency to natural causes. I have also argued that no good philosophical arguments support that reduction. Indeed, those arguments that do are circular, presupposing the very naturalism they are supposed to underwrite. My strongest argument against the sufficiency of natural causes to account for intelligent agency, however, comes from the complexity-specification criterion. This empirically-based criterion reliably discriminates intelligent agency from natural causes. Moreover, when applied to cosmology and biology, it demonstrates not only the incompleteness of natural causes, but also the presence of transcendent design.
Now, within Christian theology there is one and only one way to make sense of transcendent design, and that is as a divine act of creation. I want therefore next to focus on divine creation, and specifically on the creation of the world. My aim is to use divine creation as a lens for understanding intelligent agency generally. God’s act of creating the world is the prototype for all intelligent agency (creative or not). Indeed, all intelligent agency takes its cue from the creation of the world. How so? God’s act of creating the world makes possible all of God’s subsequent interactions with the world, as well as all subsequent actions by creatures within the world. God’s act of creating the world is thus the prime instance of intelligent agency.
Let us therefore turn to the creation of the world as treated in Scripture. The first thing that strikes us is the mode of creation. God speaks and things happen. There is something singularly appropriate about this mode of creation. Any act of creation is the concretization of an intention by an intelligent agent. Now in our experience, the concretization of an intention can occur in any number of ways. Sculptors concretize their intentions by chipping away at stone; musicians by writing notes on lined sheets of paper; engineers by drawing up blueprints; etc. But in the final analysis, all concretizations of intentions can be subsumed under language. For instance, a precise enough set of instructions in a natural language will tell the sculptor how to form the statue, the musician how to record the notes, and the engineer how to draw up the blueprints. In this way language becomes the universal medium for concretizing intentions.
In treating language as the universal medium for concretizing intentions, we must be careful not to construe language in a narrowly linguistic sense (for example, as symbol strings manipulated by rules of grammar). The language that proceeds from God’s mouth in the act of creation is not some linguistic convention. Rather, as John’s Gospel informs us, it is the divine Logos, the Word that in Christ was made flesh, and through whom all things were created. This divine Logos subsists in himself and is under no compulsion to create. For the divine Logos to be active in creation, God must speak the divine Logos. This act of speaking always imposes a self-limitation on the divine Logos. There is a clear analogy here with human language. Just as every English utterance rules out those statements in the English language that were not uttered, so every divine spoken word rules out those possibilities in the divine Logos that were not spoken. Moreover, just as no human speaker of English ever exhausts the English language, so God in creating through the divine spoken word never exhausts the divine Logos.
Because the divine spoken word always imposes a self-limitation on the divine Logos, the two notions need to be distinguished. We therefore distinguish Logos with a capital “L” (that is, the divine Logos) from logos with a small “l” (that is, the divine spoken word). Lacking a capitalization convention, the Greek New Testament employs logos in both senses. Thus in John’s Gospel we read that “the Logos was made flesh and dwelt among us.” Here the reference is to the divine Logos who incarnated himself in Jesus of Nazareth. On the other hand, in the First Epistle of Peter we read that we are born again “by the logos of God.” Here the reference is to the divine spoken word that calls to salvation God’s elect.
Because God is the God of truth, the divine spoken word always reflects the divine Logos. At the same time, because the divine spoken word always constitutes a self-limitation, it can never comprehend the divine Logos. Furthermore, because creation is a divine spoken word, it follows that creation can never comprehend the divine Logos either. This is why idolatry-worshipping the creation rather than the Creator-is so completely backwards, for it assigns ultimate value to something that is inherently incapable of achieving ultimate value. Creation, especially a fallen creation, can at best reflect God’s glory. Idolatry, on the other hand, contends that creation fully comprehends God’s glory. Idolatry turns the creation into the ultimate reality. We’ve seen this before. It’s called naturalism. No doubt, contemporary scientific naturalism is a lot more sophisticated than pagan fertility cults, but the difference is superficial. Naturalism is idolatry by another name.
We need at all costs to resist naturalistic construals of logos (whether logos with a capital “L” or a small “l”). Because naturalism has become so embedded in our thinking, we tend to think of words and language as purely contextual, local, and historically contingent. On the assumption of naturalism, humans are the product of a blind evolutionary process that initially was devoid not only of humans but also of any living thing whatsoever. It follows that human language must derive from an evolutionary process that initially was devoid of language. Within naturalism, just as life emerges from non-life, so language emerges from the absence of language.
Now it’s certainly true that human languages are changing, living entities — one has only to compare the King James version of the Bible with more recent translations into English to see how much our language has changed in the last 400 years. Words change their meanings over time. Grammar changes over time. Even logic and rhetoric change over time. What’s more, human language is conventional. What a word means depends on convention and can be changed by convention. For instance, there is nothing intrinsic to the word “automobile” demanding that it denote a car. If we go with its Latin etymology, we might just as well have applied “automobile” to human beings, who are after all “self-propelling.” There is nothing sacred about the linguistic form that a word assumes. For instance, “gift” in English means a present, in German it means poison, and in French it means nothing at all. And of course, words only make sense within the context of broader units of discourse like whole narratives.
For Christian theism, however, language is never purely conventional. To be sure, the assignment of meaning to a linguistic entity is conventional. Meaning itself, however, transcends convention. As soon as we stipulate our language conventions, words assume meanings and are no longer free to mean anything an interpreter chooses. The deconstructionist claim that “texts are indeterminable and inevitably yield multiple, irreducibly diverse interpretations” and that “there can be no criteria for preferring one reading to another” is therefore false. This is not to preclude that texts can operate at multiple levels of meaning and interpretation. It is, however, to say that texts are anchored to their meaning and not free to float about indiscriminately.
Deconstruction’s error traces directly to naturalism. Within naturalism, there is no transcendent realm of meaning to which our linguistic entities are capable of attaching. As a result, there is nothing to keep our linguistic usage in check save pragmatic considerations, which are always contextual, local, and historically contingent. The watchword for pragmatism is expedience, not truth. Once expedience dictates meaning, linguistic entities are capable of meaning anything. Not all naturalists are happy with this conclusion. Philosophers like John Searle and D. M. Armstrong try simultaneously to maintain an objective realm of meaning and a commitment to naturalism. They want desperately to find something more than pragmatic considerations to keep our linguistic usage in check. Insofar as they pull it off, however, they are tacitly appealing to a transcendent realm of meaning (take, for instance, Armstrong’s appeal to universals). As Alvin Plantinga has convincingly argued, objective truth and meaning have no legitimate place within a pure naturalism. Deconstruction, for all its faults, has this in its favor: it is consistent in its application of naturalism to the study of language.
By contrast, logos resists all naturalistic reductions. This becomes evident as soon as we understand what logos meant to the ancient Greeks. For the Greeks logos was never simply a linguistic entity. Today when we think “word,” we often think a string of symbols written on a sheet of paper. This is not what the Greeks meant by logos. Logos was a far richer concept for the Greeks. Consider the following meanings of logos from Liddell and Scott’s Greek-English Lexicon:
- the word by which the inward thought is expressed (speech)
- the inward thought or reason itself (reason)
- reflection, deliberation (choice)
- calculation, reckoning (mathematics)
- account, consideration, regard (inquiry, -ology)
- relation, proportion, analogy (harmony, balance)
- a reasonable ground, a condition (evidence, truth)
Logos is therefore an exceedingly rich notion encompassing the entire life of the mind.
The etymology of logos is revealing. Logos derives from the root l-e-g. This root appears in the Greek verb lego, which in the New Testament typically means “to speak.” Yet the primitive meaning of lego is to lay; from thence it came to mean to pick up and gather; then to select and put together; and hence to select and put together words, and therefore to speak. As Marvin Vincent remarks in his New Testament word studies: “logos is a collecting or collection both of things in the mind, and of words by which they are expressed. It therefore signifies both the outward form by which the inward thought is expressed, and the inward thought itself, the Latin oratio and ratio: compare the Italian ragionare, ‘to think’ and ‘to speak’.”
The root l-e-g has several variants. We’ve already seen it as l-o-g in logos. But it also occurs as l-e-c in intellect and l-i-g in intelligent. This should give us pause. The word intelligent actually comes from the Latin rather than from the Greek. It derives from two Latin words, the preposition inter, meaning between, and the Latin (not Greek) verb lego, meaning to choose or select. The Latin lego stayed closer to its Indo-European root meaning than its Greek cognate, which came to refer explicitly to speech. According to its etymology, intelligence therefore consists in choosing between.
We’ve seen this connection between intelligence and choice before, namely, in the complexity-specification criterion. Specified complexity is precisely how we recognize that an intelligent agent has made a choice. It follows that the etymology of the word intelligent parallels the formal analysis of intelligent agency inherent in the complexity-specification criterion. The appropriateness of the phrase intelligent design now becomes apparent as well. Intelligent design is a scientific research program that seeks to understand intelligent agency by investigating specified complexity. But specified complexity is the characteristic trademark of choice. It follows that intelligent design is a thoroughly apt phrase, signifying that design is inferred precisely because an intelligent agent has done what only an intelligent agent can do, namely, make a choice.
If intelligent design is a thoroughly apt phrase, the same cannot be said for the phrase natural selection. The second word in this phrase, selection, is of course a synonym for choice. Indeed, the l-e-c in selection is a variant of the l-e-g that in the Latin lego means to choose or select, and that also appears as l-i-g in intelligence. Natural selection is therefore an oxymoron. It attributes the power to choose, which properly belongs only to intelligent agents, to natural causes, which inherently lack the power to choose. Richard Dawkins’s concept of the blind watchmaker follows the same pattern, negating with blind what is affirmed in watchmaker. That’s why Dawkins opens his book The Blind Watchmaker with the statement: “Biology is the study of complicated things that give the appearance of having been designed for a purpose.” Natural selection and blind watchmakers don’t yield actual design, but only the appearance of design.
Having considered the role of logos in creating the world, I want next to consider its role in rendering the world intelligible. To say that God through the divine Logos acts as an intelligent agent to create the world is only half the story. Yes, there is a deep and fundamental connection between God as divine Logos and God as intelligent agent-indeed, the very words logos and intelligence derive from the same Indo-European root. The world, however, is more than simply the product of an intelligent agent. In addition, the world is intelligible.
We see this in the very first entity that God creates — light. With the creation of light, the world becomes a place that is conceptualizable, and to which values can properly be assigned. To be sure, as God increasingly orders the world through the process of creation, the number of things that can be conceptualized increases, and the values assigned to things become refined. But even with light for now the only created entity, it is possible to conceptualize light, distinguish it from darkness, and assign a positive value to light, calling it good. The world is thus not merely a place where God’s intentions are fulfilled, but also a place where God’s intentions are intelligible. Moreover, that intelligibility is as much moral and aesthetic as it is scientific.
God, in speaking the divine Logos, not only creates the world but also renders it intelligible. This view of creation has far reaching consequences. For instance, the fact — value distinction dissolves opposite God’s act of creation — indeed, what is and what ought to be unite in God’s original intention at creation. Consider too Einstein’s celebrated dictum about the comprehensibility of the world. Einstein claimed: “The most incomprehensible thing about the world is that it is comprehensible.” This statement, so widely regarded as a profound insight, is actually a sad commentary on naturalism. Within naturalism the intelligibility of the world must always remain a mystery. Within theism, on the other hand, anything other than an intelligible world would constitute a mystery.
God speaks the divine Logos to create the world, and thereby renders the world intelligible. This fact is absolutely crucial to how we understand human language, and especially human language about God. Human language is a divine gift for helping us to understand the world, and by understanding the world to understand God himself. This is not to say that we ever comprehend God, as in achieving fixed, final, and exhaustive knowledge of God. But human language does enable us to express accurate claims about God and the world. It is vitally important for the Christian to understand this point. Human language is not an evolutionary refinement of grunts and stammers formerly uttered by some putative apelike ancestors. We are creatures made in the divine image. Human language is therefore a divine gift that mirrors the divine Logos.
Consider what this conception of language does to the charge that biblical language is hopelessly anthropomorphic. We continue to have conferences in the United States with titles like “Reimagining God.” The idea behind such titles is that all our references to God are human constructions and can be changed as human needs require new constructions. Certain feminist theologians, for instance, object to referring to God as father. God as father, we are told, is an outdated patriarchal way of depicting God that, given contemporary concerns, needs to be changed. “Father,” we are told is a metaphor co-opted from human experience and pressed into theological service. No. No. No. This view of theological language is hopeless and destroys the Christian faith.
The concept father is not an anthropomorphism, nor is referring to God as father metaphorical. All instances of fatherhood reflect the fatherhood of God. It’s not that we are taking human fatherhood and idealizing it into a divine father image à la Ludwig Feuerbach or Sigmund Freud. Father is not an anthropomorphism at all. It’s not that we are committing an anthropomorphism by referring to God as father. Rather, we are committing a “theomorphism” by referring to human beings as fathers. We are never using the word “father” as accurately as when we attribute it to God. As soon as we apply “father” to human beings, our language becomes analogical and derivative.
We see this readily in Scripture. Jesus enjoins us to call no one father except God. Certainly Jesus is not telling us never to refer to any human being as “father.” All of us have human fathers, and they deserve that designation. Indeed, the Fifth Commandment tells us explicitly to honor our human fathers. But human fathers reflect a more profound reality, namely, the fatherhood of God. Or consider how Jesus responds to a rich, young ruler who addresses him as “good master.” Jesus shoots back, “Why do you call me good? There is no one good except God.” Goodness properly applies to God. It’s not an anthropomorphism to call God good. The goodness we attribute to God is not an idealized human goodness. God defines goodness. When we speak of human goodness, it is only as subordinate to the divine goodness.
This view, that human language is a divine gift for understanding the world and therewith God, is powerfully liberating. No longer do we live in a Platonic world of shadows from which we must escape if we are to perceive the divine light. No longer do we live in a Kantian world of phenomena that bars access to noumena. No longer do we live in a naturalistic world devoid of transcendence. Rather, the world and everything in it becomes a sacrament, radiating God’s glory. Moreover, our language is capable of celebrating that glory by speaking truly about what God has wrought in creation.
The view that creation proceeds through a divine spoken word has profound implications not just for the study of human language, but also for the study of human knowledge, or what philosophers call epistemology. For naturalism, epistemology’s primary problem is unraveling Einstein’s dictum: “The most incomprehensible thing about the world is that it is comprehensible.” How is it that we can have any knowledge at all? Within naturalism there is no solution to this riddle. Theism, on the other hand, faces an entirely different problematic. For theism the problem is not how we can have knowledge, but why our knowledge is so prone to error and distortion. The Judeo-Christian tradition attributes the problem of error to the fall. At the heart of the fall is alienation. Beings are no longer properly in communion with other beings. We lie to ourselves. We lie to others. And others lie to us. Appearance and reality are out of sync. The problem of epistemology within the Judeo-Christian tradition isn’t to establish that we have knowledge, but instead to root out the distortions that try to overthrow our knowledge.
On the view that creation proceeds through a divine spoken word, not only does naturalistic epistemology have to go by the board, but so does naturalistic ontology. Ontology asks what are the fundamental constituents of reality. According to naturalism (and I’m thinking here specifically of the scientific naturalism that currently dominates Western thought), the world is fundamentally an interacting system of mindless entities (be they particles, strings, fields, or whatever). Mind therefore becomes an emergent property of suitably arranged mindless entities. Naturalistic ontology is all backwards. If creation and everything in it proceeds through a divine spoken word, then the entities that are created don’t suddenly fall silent at the moment of creation. Rather they continue to speak.
I look at a blade of grass and it speaks to me. In the light of the sun, it tells me that it is green. If I touch it, it tells me that it has a certain texture. It communicates something else to a chinch bug intent on devouring it. It communicates something else still to a particle physicist intent on reducing it to its particulate constituents. Which is not to say that the blade of grass does not communicate things about the particles that constitute it. But the blade of grass is more than any arrangement of particles and is capable of communicating more than is inherent in any such arrangement. Indeed, its reality derives not from its particulate constituents, but from its capacity to communicate with other entities in creation and ultimately with God himself.
The problem of being now receives a straightforward solution: To be is to be in communion, first with God and then with the rest of creation. It follows that the fundamental science, indeed the science that needs to ground all other sciences, is communication theory, and not, as is widely supposed an atomistic, reductionist, and mechanistic science of particles or other mindless entities, which then need to be built up to ever greater orders of complexity by equally mindless principles of association, known typically as natural laws. Communication theory’s object of study is not particles, but the information that passes between entities. Information in turn is just another name for logos. This is an information-rich universe. The problem with mechanistic science is that it has no resources for recognizing and understanding information. Communication theory is only now coming into its own. A crucial development along the way has been the complexity-specification criterion. Indeed, specified complexity is precisely what’s needed to recognize information.
Information — the information that God speaks to create the world, the information that continually proceeds from God in sustaining the world and acting in it, and the information that passes between God’s creatures — this is the bridge that connects transcendence and immanence. All of this information is mediated through the divine Logos, who is before all things and by whom all things consist (Colossians 1:17). The crucial breakthrough of the intelligent design movement has been to show that this great theological truth — that God acts in the world by dispersing information — also has scientific content. All information, whether divinely inputted or transmitted between creatures, is in principle capable of being detected via the complexity-specification criterion. Examples abound:
- The fine-tuning of the universe and irreducibly complex biochemical systems are instances of specified complexity, and signal information inputted into the universe by God at its creation.
- Predictive prophecies in Scripture are instances of specified complexity, and signal information inputted by God as part of his sovereign activity within creation.
- Language communication between humans is an instance of specified complexity, and signals information transmitted from one human to another.
The positivist science of this and the last century was incapable of coming to terms with information. The science of the new millennium will not be able to avoid it. Indeed, we already live in an information age.
Creativity, Divine and Human
In closing this talk, I want to ask an obvious question: Why create? Why does God create? Why do we create? Although creation is always an intelligent act, it is much more than an intelligent act. The impulse behind creation is always to offer oneself as a gift. Creation is a gift. What’s more, it is a gift of the most important thing we possess — ourselves. Indeed, creation is the means by which a creator — divine, human, or otherwise — gives oneself in self-revelation. Creation is not the neurotic, forced self-revelation offered on the psychoanalyst’s couch. Nor is it the facile self-revelation of idle chatter. It is the self-revelation of labor and sacrifice. Creation always incurs a cost. Creation invests the creator’s life in the thing created. When God creates humans, he breathes into them the breath of life — God’s own life. At the end of the six days of creation God is tired — he has to rest. Creation is exhausting work. It is drawing oneself out of oneself and then imprinting oneself on the other.
Consider, for instance, the painter Vincent van Gogh. You can read all the biographies you want about him, but through it all van Gogh will still not have revealed himself to you. For van Gogh to reveal himself to you, you need to look at his paintings. As the Greek Orthodox theologian Christos Yannaras writes: “We know the person of van Gogh, what is unique, distinct and unrepeatable in his existence, only when we see his paintings. There we meet a reason (logos) which is his only and we separate him from every other painter. When we have seen enough pictures by van Gogh and then encounter one more, then we say right away: This is van Gogh. We distinguish immediately the otherness of his personal reason, the uniqueness of his creative expression.”
The difference between the arts and the sciences now becomes clear. When I see a painting by van Gogh, I know immediately that it is his. But when I come across a mathematical theorem or scientific insight, I cannot decide who was responsible for it unless I am told. The world is God’s creation, and scientists in understanding the world are simply retracing God’s thoughts. Scientists are not creators but discoverers. True, they may formulate concepts that assist them in describing the world. But even such concepts do not bear the clear imprint of their formulators. Concepts like energy, inertia, and entropy give no clue about who formulated them. Hermann Weyl and John von Neumann were both equally qualified to formulate quantum mechanics in terms of Hilbert spaces. That von Neumann, and not Weyl, made the formulation is now an accident of history. There’s nothing in the formulation that explicitly identifies von Neumann. Contrast this with a painting by van Gogh. It cannot be confused with a Monet.
The impulse to create and thereby give oneself in self-revelation need not be grand, but can be quite humble. A homemaker arranging a floral decoration engages in a creative act. The important thing about the act of creation is that it reveal the creator. The act of creation always bears the signature of the creator. It is a sad legacy of modern technology, and especially the production line, that most of the objects we buy no longer reveal their maker. Mass production is inimical to true creation. Yes, the objects we buy carry brand names, but in fact they are largely anonymous. We can tell very little about their maker. Compare this with God’s creation of the world. Not one tree is identical with another. Not one face matches another. Indeed, a single hair on your head is unique — there was never one exactly like it, nor will there ever be another to match it.
The creation of the world by God is the most magnificent of all acts of creation. It, along with humanity’s redemption through Jesus Christ, are the two key instances of God’s self-revelation. The revelation of God in creation is typically called general revelation whereas the revelation of God in redemption is typically called special revelation. Consequently, theologians sometimes speak of two books, the Book of Nature, which is God’s self-revelation in creation, and the Book of Scripture, which is God’s self-revelation in redemption. If you want to know who God is, you need to know God through both creation and redemption. According to Scripture, the angels praise God chiefly for two things: God’s creation of the world and God’s redemption of the world through Jesus Christ. Let us follow the angels’ example.