Seb and I discussed the history and contemporary status of the natural sciences and their relationship to philosophy and metaphysics.
a talk delivered for the Philosophy, Cosmology, and Consciousness program at CIIS.edu on Friday, January 29th, 2021.
I’m teaching for Schumacher College again, this time online. This course focuses on two towering exemplars of the organic approach to science, the German poet and naturalist Johann Wolfgang von Goethe (1749-1832) and the British mathematician and philosopher Alfred North Whitehead (1861-1947).
The course will run via live video conference on Saturday mornings (PST) for six weeks beginning in late January. Visit the Schumacher College website to register (before Jan 10, 2021). Here’s a short interview I did to introduce the foci of the course:
The course begins in the late eighteenth century by setting out the revolutionary cultural, philosophical, and scientific context within which Goethe developed his participatory understanding of Nature. Goethe is still primarily known as a poet, but students will come to see how the rise of Newton’s clockwork vision of the cosmos and the development of Kant’s nascent theory of living organization led Goethe (with help from the German Idealist Friedrich Schelling) to imagine a more organic and relational way of doing science. The course then turns to explore Goethe’s novel approach to the study of light and colour, geology, plant metamorphosis, and animal morphology.
During the nineteenth century, Goethe’s participatory way of doing natural science was largely forgotten, especially in the English-speaking world. Modern physics and biology followed Descartes and Newton’s lead by becoming increasingly mechanistic, while organic ways of thinking were dismissed as childish pre-modern holdovers. But at the turn of the twentieth century, physics underwent a series of revolutions that upset the mechanistic world-picture. It was the relativistic and quantum paradigm shifts that brought Whitehead out of mathematics and into metaphysics and cosmology. The course examines the reasons for the breakdown of the mechanistic view of Nature and unpacks Whitehead’s organic alternative, placing him alongside Goethe and Schelling as part of a legacy of participatory thinkers.
The course culminates in an exploration of organic science in our own day, looking at the enduring influence of participatory thinking in physics, biology, and spirituality. Students will be invited to reimagine the scientific world view in the context of an ensouled universe.
This course is designed for students of intellectual history who are fascinated by subversive streams of thought that have not yet been given their due. Some background in the history of European philosophy and science will be helpful, but the lecturer will attempt to make the ideas accessible to everyone.
Recommended reading prior to course start date:
1) The Theory of Knowledge Implicit in Goethe’s World View By Rudolf Steiner (77 pages, available free online)
2) Physics of the World-Soul: Whitehead’s Adventure in Cosmology By Matthew Segall (130 pages, available free online)
Thanks to Bruce Alderman at The Integral Stage for moderating.
My lecture in two parts introducing German Idealism (focusing on Kant, Fichte, Schelling, Goethe, Hegel)
I’ve just finished drafting this article, which will hopefully be featured in a special issue of the Journal of Philosophy, Theology, and the Sciences focused on panpsychism. It still needs plenty of editing, but I’m sharing it here for those who want a sneak peak. Criticisms and suggestions definitely welcome.
Title: “Physicalism and Its Discontents: A Study in Whitehead’s Panexperientialist Alternative”
Abstract (150-200 words): This paper brings Alfred North Whitehead’s Philosophy of Organism into conversation with the recent panpsychist turn in analytic philosophy of mind. Whitehead’s unabashedly metaphysical project broadly aligns with recent critiques of reductive physicalism and the turn toward a conception of experience as basic to Nature. This paper thus examines physicalism’s dominant strategies for explaining consciousness, including eliminativism, epiphenomenalism, and emergentism, and concludes that the panpsychist alternative is superior. However, Whitehead’s process-relational panexperientialism diverges in crucial respects from the dominant substance-property variants of panpsychism. I argue that Whitehead’s version avoids many of the conceptual difficulties plaguing the latter and that it thus represents a more formidable alternative to standard physicalism.
Key words: panpsychism, panexperientialism, physicalism, emergence, experience, consciousness, process philosophy
The skull-crackingly hard problem concerning the place of consciousness in the physical universe has led an increasing number of analytic philosophers of mind to take seriously the panpsychist alternative to standard physicalism. Nonetheless, Brüntrup and Jaskolla note in their editors’ introduction to Panpsychism: Contemporary Perspectives that the usual response to the doctrine remains “an incredulous stare” (2017, 2). Perhaps the most forceful dismissal to date comes from Colin McGinn, who in a reply to Galen Strawson rejects panpsychism as “a comforting piece of utter balderdash” that only stoned hippies could believe (McGinn 2006, 93).
But an explanation for the emergence of consciousness in the universe known to physics has thus far proven elusive. Fundamental philosophical questions remain to be answered before the criteria for such a scientific explanation can even be established. For example, is consciousness essentially ‘real’ or ‘illusory’? That is, does it “have truck with the totality of things by reason of its sheer actuality,” as Alfred North Whitehead suspected (1929, 15), or is it a peripheral accident, a mere epiphenomenon emergent from blindly churning physiochemical processes that are otherwise well understood by natural science? Does consciousness evolve, and if so, does it intelligently influence the behavior of the organisms instantiating it? These questions are not merely theoretical or academic. They cut to the very core who and what we are, shaping our sense of what it means to be human.
Despite the initial incredulity it provokes, this paper argues that panpsychism—specifically Whitehead’s process-relational, panexperiential version—provides a viable alternative to scientific materialism while also avoiding the philosophical excesses of dualism and idealism. Strange as it may sound to modern ears, panpsychism has a long and rich history stretching back to the origins of Western philosophy. Heraclitus opposed Parmenides’ vision of unchanging Being with the doctrine that ‘everything flows’ (panta rhea). Heraclitus understood the universe to be “an ever-living fire” (pyr aeizoon), making him not only the first recorded process philosopher but the first panpsychist, as well (Skrbina 2005, 29). Even in the early modern period, thinkers like Giordano Bruno and Gottfried Leibniz, often lauded for their important contributions to the emergence of the scientific worldview, continued to uphold some version of the doctrine. “Lucretius tells us what an atom looks like to others,” writes Whitehead, “and Leibniz tells us how an atom is feeling about itself” ( 1967, 132). Skeptics may be tempted to excuse Bruno and Leibniz’s panpsychist eccentricity as an unthought residue of pre-modern animism. Once enlightened by the findings of contemporary physics and biology, surely these luminaries would happily have dispensed with the ‘primitive’ notion that atoms can feel? Perhaps not. What, after all, are we to make of Whitehead, another mathematical and philosophical genius who critiqued scientific materialism and arrived at his own variety of panpsychism not despite but because of the findings of contemporary physics and biology?
“There persists…[a] fixed scientific cosmology which presupposes the ultimate fact of an irreducible brute matter…spread throughout space in a flux of configurations. In itself such a material is senseless, valueless, purposeless. It just does what it does do, following a fixed routine imposed by external relations which do not spring from the nature of its being…[This] is an assumption which I shall challenge as being entirely unsuited to the scientific situation at which we have now arrived” (Whitehead  1967, 17).
This paper brings Whitehead’s “Philosophy of Organism” ( 1978) into conversation with the recent panpsychist turn in analytic philosophy of mind. Whitehead’s unabashedly metaphysical project broadly aligns with recent critiques of reductive physicalism and the turn toward a conception of experience as basic to Nature. However, Whitehead’s process-relational panexperientialism diverges in crucial respects from the dominant substance-property variants of panpsychism. I argue that Whitehead’s version avoids many of the conceptual difficulties plaguing the latter and that it thus represents a more formidable alternative to standard physicalism.
1. Why not Whitehead?: A Brief Historical Excursus
“Urge & urge & urge,
Always the procreant urge of the world.
Out of the dimness opposite equals advance, always substance & increase, always sex,
Always a knit of identity, always distinction, always a breed of life.”
—Whitman (“Song of Myself”)
Understanding Whitehead’s process-relational approach to panpsychism (or panexperientialism, as David Ray Griffin has renamed it [Griffin 2008, 78]) first requires a bit of historical contextualization. While Whitehead’s early work with Bertrand Russell on the logical foundations of mathematics is widely acknowledged by analytic philosophers as seminal to the emergence of their school of thought, Whitehead’s later metaphysical speculations are for the most part either ignored or ridiculed. For example, W. V. Quine traveled to Harvard in the mid-1920s to study with the coauthor of the Principia Mathematica. After attending the lectures that became Science and the Modern World (1925), Quine acknowledged “a vivid sense of being in the presence of the great” but went on to admit that the notes he took were mostly full of doodles. “What [Whitehead] said,” Quine reports, “had little evident bearing on the problems that I recognized” (Quine 1985, 83). Another student of Whitehead’s at Harvard, Donald Davidson, was initially transfixed by his ideas, but later reflected that his encounter with Whitehead “set [him] back philosophically for years” by confirming his youthful “inclination to think that doing philosophy was like writing poetry” (Davidson 1999, 14). Not everyone was quite as sour on Whitehead’s speculations at Harvard. Ernest Nagel credited Whitehead with being one of the first to realize and attempt to address the metaphysical problems that were becoming “acutely pressing in the special sciences,” praising him for his “[sensitivity] to the advances of recent science as well to the ancient tradition that philosophy is the systematic study of being” (E. Nagel 1954, 154). But Nagel also noted “the severe abuse of language to which Whitehead is partial” (ibid.), a familiar (if not entirely fair) refrain among those who attempt to read him for the first time.
To round out this historical excursus, let us return to Nagel’s point about the special sciences. By the mid-1920s, the new quantum and relativity theories had already succeeded in demolishing the old mechanical philosophy of Nature by transforming matter into energy and merging space and time together with gravity. The classical explanations of Nature offered by a once confident scientific materialism no longer made any sense. A second scientific revolution was afoot. At the same time, Ludwig Wittgenstein led the logical positivists in a revolt against the excesses of British idealism by blowing up the bridge purporting to connect the metaphysical speculations of philosophers with the ultimate nature of things: “Whereof one cannot speak, thereof one must be silent” (Wittgenstein 1922, 189). The physicists struggling to come to terms with the strange ontological implications of their discoveries could henceforth expect no help from philosophers. Whitehead’s own pathbreaking work on the application of mathematics to physics made him especially sensitive to Einstein’s relativistic revolution; he was also well aware of the concurrently unfolding quantum revolution. His sensitivity to the metaphysical earthquake underway in the physical sciences awakened Whitehead from the dogmatic slumber of the mechanistic paradigm. “What is the sense of talking about a mechanical explanation,” Whitehead asked, “when you do not know what you mean by mechanics?” (1925, 16). His Philosophy of Organism is a protest against the lifeless Nature imagined by Descartes, Galileo, and Newton, and a rejection of the narrow linguistic analysis and sterile logical positivism of his philosophical contemporaries. His is an attempt to make natural science philosophical again by asking whether physical causes and motions need be so violently segregated from the conscious reasons and emotions by which we apprehend them.
In Process & Reality: An Essay in Cosmology ( 1978), Whitehead aims for nothing less than the construction of an organic system of the universe that not only brings quantum and relativity theories into coherence, but gathers up scientific truths, aesthetic feelings, and religious values into an integral vision of reality. It is true that Whitehead found it necessary to invent many new turns of phrase to accomplish this feat. He thus contrasts his speculative philosophical method with that of the “critical school” (Whitehead  1968, 173), which for my purposes can easily be identified with the then just emerging analytic school of thought. This school assumes that humanity “has consciously entertained all the fundamental ideas which are applicable to its experience” and that “human language, in single words or in phrases, explicitly expresses these ideas” (ibid.). The critical or analytic school, Whitehead continues, “confines itself to verbal analysis within the limits of the dictionary” (ibid.). In contrast, Whitehead’s speculative method “appeals to direct insight, and endeavors to indicate its meanings by further appeal to situations which promote such specific insights. It then enlarges the dictionary” (ibid.). Whitehead credits analytic philosophy for its “delicate accuracy of expression,” but marks the main “divergence between the schools [as] the quarrel between safety and adventure” (ibid.).
Davidson worried about the adventurous Whitehead’s attempted alliance between speculative philosophy and mystical poetry. Both, according to Whitehead, make “reference to form beyond the direct meanings of words.” He continues: “If you like to phrase it so, philosophy is mystical. For mysticism is direct insight into depths as yet unspoken” (ibid., 174). Davidson’s complaint may be short-sighted, however, especially once one has acknowledged the profound metaphysical problems that after nearly a century of careful analysis continue to plague not only the physical sciences but the philosophy of mind, as well. Hamlet was right: “There are more things in heaven and earth…”
While getting to the bottom of Whitehead’s chilly reception among analytic philosophers is not the aim of this paper, a few conjectures can be offered. After a celebrated first career as a mathematician, Whitehead’s untimely entry into philosophy in the mid-1920s can be read as heralding the more recent return to metaphysics both in the analytic and Continental traditions. Philosophers are finally catching up to the problems Whitehead was pointing out nearly a century ago. Perhaps it is just because his cosmological ideas initially emerged in the wrong season that they have remained buried in the snow. In addition to the unfortunate timing, Whitehead’s lack of easy classification is probably another reason for his neglect. Neither an analytic philosopher, nor a phenomenologist, Whitehead’s approach generally confounds partisans of both schools. That said, his process-relational philosophy has been creatively taken up by a number of friendly thinkers on the Continent (initially Henri Bergson (1999, 47), later Gilles Deleuze ( 1994, 284-285;  1993, 76ff), and most recently Isabelle Stengers (2011) and Bruno Latour (2005). Whitehead’s thought also featured prominently in the Speculative Realism movement that swept through Continental philosophy beginning in late 2010 (Bryant et al. 2011; Harman 2018). He is perhaps best situated within the American pragmatist tradition stemming from Charles Sanders Peirce, William James, and John Dewey, though even here the shoe pinches. Dewey is the only one who lived long enough to respond to Whitehead’s philosophy, which he praises for its organicism and experiential point of departure but criticizes for its mathematical residues (Schilpp 1941). In the end it must be admitted that Whitehead’s process-relational philosophy is singular in its aims and conclusions. Any attempt to pigeonhole his thought into a school inevitably trivializes it. Of course, Whitehead himself generated a school, but there exist plenty of wild Whiteheadians who avoid any established orthodoxies, like Deleuze, Stengers, and Latour, or Randall Auxier and Gary Herstein (2017).
Finally, there is the issue of Whitehead’s inclusion of reformed conceptions of teleology and God in his cosmological scheme. For many philosophers and natural scientists, this rules out in advance any serious engagement with his ideas. Daniel J. Nicholson and John Dupré, for example, claim that the theological baggage of Whitehead’s process philosophy is a “liability” for thinkers with a naturalistic aim (2018, 7). But a closer look at Whitehead’s process-relational reformulations of purpose and divinity may reveal to those who rushed to dismiss them that Whitehead shares many of their criticisms of traditional natural theology. By the time God and teleology return from Whitehead’s adventure in cosmology, the former is no longer an omnipotent Creator but a creature of Creativity suffering with the rest of us, and the latter is no longer an eternal design imposed from beyond the world but an aesthetic lure immanent in the experience of each and every actual occasion in the world, whether that experience belongs to Shakespeare or “to the most trivial puff of existence in far-off empty space” (Whitehead  1978, 28).
My hope is that this paper brings Whitehead out of cold storage and at least thaws his ideas enough to get those unfamiliar with his Philosophy of Organism to consider the alternative it represents, not only to physicalism, but to dualism and idealism, as well. Despite Quine’s first impression, it may turn out that Whitehead has much to say about the problems faced by contemporary analytic philosophers, especially those who, against all odds, now find themselves affirming the panpsychist heresy.
2. The Place of Consciousness in a Physical Universe
Serious conceptual difficulties await any philosopher attempting to understand the place of consciousness in the physical universe. David Chalmers’ well-known “hard problem of consciousness” (1995) is perhaps the most oft cited formulation of the impasse, but the basic problem goes back to Rene Descartes’ argument that a real distinction exists between a thinking or mental substance and an extended or material substance ( 1982, 21]. While many contemporary physicists would be quick to dismiss Descartes’ idea of an immaterial soul as unscientific, his correlate idea of extended matter continues to shape the scientific imaginary of Nature as something explainable without remainder in purely mathematical terms. While Descartes faced the difficult problem of accounting for the relationship between two entirely autonomous substances, contemporary physicalists face what is arguably an even harder problem: how can extended matter in motion ever give rise to the seemingly interior experience of conscious thought and emotion? As Galen Strawson has pointed out, even if this “seeming” experience ends up being some sort of illusion, the seeming itself still demands an explanation: “any such illusion is already and necessarily an actual instance of the thing said to be an illusion” (Strawson 2018).
Let us run through the various metaphysical options at play for those affirming standard physicalism, by which I mean any variation on the ontology that posits that the final real things (whether particles, fields, or some other mode of existence yet to be discovered by science) are passively enduring objects entirely devoid of subjective enjoyment and aim. When addressing the place of consciousness in Nature, physicalists generally draw upon three basic explanatory strategies: eliminativism, epiphenomenalism, and emergentism. Many physicalists, in order to side-step patent absurdities, end up tacitly sliding back and forth between two or more of these positions in the course of their explanations of consciousness. Unfortunately, there is little consistency in how these terms are defined in existing literature, hence the need to offer accounts of each position as they are considered for the purposes of this paper.
Eliminativism tries to deny the reality of consciousness outright, arguing that our folk psychological intuitions and self-reports about it are hopelessly misguided and need to be replaced by more mature neurophysiological or computational accounts. While Paul and Patricia Churchland are perhaps the most prominent contemporary defenders of this position (P. S. Churchland 1986; P. M. Churchland 1988), its origins can be traced back to Wilfred Sellars (1956) and Quine (1960). Quine’s reflections on the matter are especially relevant. He raises the question of whether eliminativism truly “repudiates” conscious experiences as factually mistaken, or whether it is meant as a theory identifying such experiences with physiological facts (Quine 1960, 265). He decides that there is no real distinction to be made in this case between explanation and identification. If the elimination of consciousness in favor of physiological processes is the same as the identification of consciousness with correlated physiological processes, all the sudden eliminativism starts to sound a lot like panpsychism, with the crucial qualification that the panpsychist refuses to grant brain matter any special ontological status, as though it instantiated experiential capacities not found to some degree in all physical processes. In Whitehead’s terms: “There’s nothing in the real world which is merely an inert fact. Every reality is there for feeling: it promotes feeling; and it is felt” (Whitehead  1978, 310). In other words, if Quine’s reading is right, Whitehead is also an eliminativist about that sort of consciousness that is imagined to be something extra in addition to physical processes.
More recently, a quasi-transcendental version of eliminativism has been defended under the label of “illusionism” (Frankish 2016). The idea is that we suffer inextricably from what Daniel Dennett calls a “user-illusion” (Dennett 2017, 222). There is really no one home inside, but because we are constitutively blind to the neural basis of our user-illusion, we cannot help but keep knocking on the door. The answer to all our knocking comes only as a bunch of mouth-squeaks signifying nothing (other than more squeaks). We are just a bunch of neurons and chemistry playing out an evolutionary algorithm. “We’re all zombies” (Dennett 2004, 67). Despite his critics, Dennett denies that his version of physicalism is eliminativist (Dennett 2017, 224). His philosophy is a good example of the way the most inventive physicalists end up combining aspects of multiple positions, sliding from eliminativism for questions of ontology to emergentism when it’s a question of the practical functionality of conscious will (Dennett 2003).
Hard core eliminativists like the Churchlands, or like the speculative realist philosopher Ray Brassier (2007), can at least be credited with bitting the materialist bullet by accepting that any physicalism worthy of the name leaves absolutely no room in the universe for anything like what most people mean by consciousness. For Brassier, eliminativism is not just a promising neuroscientific theory of consciousness but a tremendous opportunity for speculative philosophy. Philosophers, rather than acting as “a sop to the pathetic twinge of human self-esteem” by continuing to seek the restoration of a meaningful connection between human consciousness and the cosmic processes that generate it, should instead follow the logic of eliminativism to its admittedly nihilistic conclusions (Brassier 2007, xi). Even if attempts to restore meaning succeed in increasing our quality of life, Brassier still calls upon self-respecting philosophers to reject them, since “thinking has interests that do not coincide with those of living” (ibid.). The eliminativist position can be criticized as self-refuting, since it denies in theory what, short of suicide, one cannot deny in practice (though even the act suicide implies a conscious decision to kill oneself). How can one claim to hold to the view of eliminative materialism if the capacity for holding meaningful views of anything is precisely what the position purports to be eliminating? Brassier responds to the performative contradiction criticism by pointing out that the eliminativist project entails a rejection and replacement of the folk psychological view of ‘views’ or ‘beliefs’ assumed by its critics. Following Paul Churchland, Brassier reduces the propositional meanings and sentential beliefs of folk psychology to the “dynamics and kinematics” of neural activation patterns in the brain (Brassier 2007, 12, 15-17). What it is to hold a particular view (e.g., “Eliminativism is true”) is just for the relevant neural pathways to fire.
While panpsychism may initially affront the common sense of modern Western adults, eliminativism is an even bigger stretch. Of course, the common sense folk psychology of a particular era cannot be given the privileged position of determining metaphysical reality. Whitehead’s process-relational panexperientialism entails a radical revisioning of our common sense understandings of consciousness and propositional meaning. But it does not deny conscious experience outright. Philosophy can reform common sense without eliminating the very possibility of a meaningful life. According to Whitehead, “As we think, we live” (Whitehead,  1968, 63). Thinking is, after all, as natural to the life of a conscious organism as eating or breathing. If our philosophy cannot in the end be squared with the “overpowering deliverances” (Whitehead  1978, 50) of experience and the “concrete affairs of life” (Whitehead  1967, 80), it is a good sign that we have made a wrong turn somewhere in our abstract reasoning. This, at least, is how a pragmatic radical empiricist like Whitehead addresses the matter: “Metaphysics is nothing but the description of the generalities which apply to all the details of practice” (Whitehead  1978, 13).
Epiphenomenalism claims there is room enough for consciousness to be somehow excreted by the brain, but only as a semi-transparent ghost or “inert spectator” (James 1890, 129) with no causal influence over the goings-on of the body or its proximal environment. As formulated most famously by Thomas Huxley, epiphenomenalism is the view that consciousness is “completely without any power…as the steam-whistle which accompanies the work of a locomotive engine is without influence upon its machinery” (Huxley 1875, 62). Epiphenomenalists at least acknowledge the irreducibility of our direct intuition of conscious experience. But assuming a broadly naturalistic and thus evolutionary framework (as Huxley and most contemporary defenders of the doctrine claim to) rules out accounts of epiphenomenal consciousness as sealed off from but nonetheless perfectly correlated with physical processes via a “pre-established harmony” (e.g., Leibniz). Any naturalistic account must explain the causal nexus between mental and physical processes, even if the causal relationships are said to move in only one direction, i.e., physical causes determining an epiphenomenal steam-whistle. Given the requirements of naturalism, the problem with epiphenomenalism is that it is incomprehensible how such a complex ghost-like consciousness could ever have evolved if it serves no function at all for the organism it haunts. If consciousness plays no active role in shaping an organism’s behavior, it cannot be selected for and thus has no role in biological adaptation (T. Nagel 2012, 44ff). As James argued more than a century ago, it is an absurd abuse of scientistic reasoning to assert in the same breath that, while consciousness exists, “all those manners of existence which make it seem relevant to our outward life are mere meaningless coincidences, inexplicable parts of the general and intimate irrationality of this disjointed world” (James 1879, 21). Not only is the view epiphenomenalist view incoherent, the opposed view, that consciousness to varying degrees depending on cerebral complexity “[exerts] a constant pressure in the direction of survival,” grants further plausibility to the Darwinian evolutionary story: “It is, in fact, hard to see how without an effective superintending ideal the evolution of so unstable an organ as the mammalian cerebrum can have proceeded at all” (ibid., 16).
The neuroscientist Michael Graziano attempts to avoid this problem with epiphenomenalism by redefining conscious awareness in neuroscientific terms as “attention” (Graziano 2019). While focusing on the ‘phenomenal properties’ of conscious awareness gives philosophers the impression that subjective experience is some sort of extra ethereal or nonphysical essence (e.g., private ‘qualia’), what Graziano calls an “attention schema” has been scientifically measured in brain-based computational terms (ibid.). The attention schema is the brain’s way of internally modeling certain aspects of its own activity, and our reports and claims about our own consciousness appear to correlate with it (ibid., 101). Graziano thus slides away from the hard problem of consciousness to ask a different question: what sort of neural computations allow us to make claims about supposedly conscious experiences? “In this theory,” writes Graziano, “the ghost in the machine, the consciousness inside us, is a topic of discussion among us only because our intuitions are informed by an attention schema, with its incomplete account of attention” (ibid., 103). While a supposedly ethereal essence would have no way of altering the behavior of an organism, the attention schema serves an adaptive function by monitoring, predicting, and controlling the brain’s attentional resources (ibid., 101). It performs this function in a purely physical way without the influence of any extra-physical consciousness.
While a Whiteheadian approach has its own reasons for being critical of the search for ethereal ‘phenomenal properties’ or private ‘qualia’ (see sections 3 and 4 below), Graziano’s neuroscientific slight of hand gets us no closer to understanding the place of consciousness in the physical world. To start with, consciousness is not merely “a topic of discussion” and cannot be reduced to the sentential claims we make about ourselves and our experience. Whatever else it is, conscious experience of oneself in a world is an immediately intuited concrete fact, not just a linguistic report about or computational model of a fact. Graziano admits he isn’t offering a philosophical answer for how consciousness arises in the brain, but he also implies that his properly scientific approach forces us to accept that “there is no meaningful answer to the question” (ibid., 97). We are just “a biological machine that claims to have a hard problem” (ibid., 96). We are brain networks running a linguistic program whose only power is that it can make claims about itself, statements about what it believes is going on and what its own and other people’s intentions are. These beliefs, claims, and intentions have no bearing on what is actually going on inside the skull or beyond it, since their meanings are epiphenomenal to computations in the brain and the motion of matter through spacetime.
A broader assumption baked into Graziano’s approach is that “the brain is an information processing device” (ibid., 95). This is stated as though it were a truth that neuroscience has discovered, but it is hardly that. It is a theoretical paradigm and a research program, that is, a framework for studying the brain as if it were a computer, not a fact about what the brain is. Other neuroscientists and philosophers of mind reject the computational approach and instead study brain activity from an enactive and embodied perspective (Varela et al.  2016, 44ff; Thompson 2007, 51ff). From an enactive perspective, speaking in terms of decontextualized and disembodied ‘information processing’ going on inside the skull neglects the extent to which meaningful information presupposes an experiential horizon within which it can be interpreted. Evan Thompson extends Gregory Bateson’s claim that “information is a difference that makes a difference” (Bateson  2000, p. 315), adding that information “is the making of a difference that makes a difference to somebody somewhere” (Thompson 2007, 57). Informational meaning is thus embedded not only in the complex dynamics of an experience-imbued brain, but in the sensorimotor networks of the body, and even extends out into the surrounding environment with which the organism is structurally coupled and has co-evolved.
Emergentism claims that consciousness suddenly appears in the universe whenever matter manages to arrange itself into the appropriate dynamical shapes. Some say a simple form of consciousness emerged with the first living cells (‘biopsychism’), while others claim these cells had to blindly organize themselves into large packs of neurons called brains before the light of consciousness could flicker on (‘cerebropsychism’). Still others insist that it was necessary for these brains to become sufficiently entangled in the symbolic network of a language before full-blown consciousness could explode onto the scene (‘anthropopsychism’).
There are weak and strong versions of emergence (Brogaard 2016, 131ff). The higher level capacities of a weakly emergent consciousness are at least in principle deducible from and thus in fact causally reducible to its lower level constituents. Once cognitive neuroscience discovers the relevant underlying brain mechanisms, complicated as they may be, the mystery of consciousness will be understood to have been only an artifact of our limited knowledge. Weak emergence thus presents an epistemological puzzle for physicalism to solve, rather than an ontological impasse forcing it to re-examine its premises. Of course, if weak emergentists do solve the engineering problem of how the brain makes the mind, it is difficult to see how they will avoid sliding back into epiphenomenalism.
Strongly emergent conceptions, in contrast, affirm the ontological novelty of consciousness above and beyond its physical components, even granting it downward causal influence upon the body and surrounding environment. Such a view at least refuses to explain away the evident facts and overpowering deliverances of conscious thought and intention, facts that law, politics, morality, religion, and practical life in general require; facts that even the endeavor to produce scientific knowledge itself necessarily presupposes, for what else is knowledge but a mode of consciousness? As Whitehead quipped, “Scientists animated by the purpose of proving they are purposeless constitute an interesting subject for study” (Whitehead  1958, 16). But unless it can explain how meaning and purpose arise out of mass and energy, strong emergentism lands us right back where Descartes left us nearly four centuries ago, with irreducible mind on one side, brute matter on the other, and no rational account of how they might relate to one another. Focusing on the gradual development of mental capacities from bacterial chemotaxis to Shakespearean poetry over the course of billions of years of biological evolution is an obvious strategy for narrowing this gap. But merely saying ‘evolution did it’ doesn’t cut it, since it wasn’t Darwinian evolution that gave rise to cellular life. Darwin’s theory of speciation by natural selection presupposes self-producing and reproducing organisms, it does not explain them. In Thompson’s terms, “natural selection is an emergent consequence of autopoiesis, not its cause”(2007, 212).
On the other hand, there is a wider definition of evolution than that assigned by Darwin. Whitehead was convinced that evolution had relevance for not just biology but all the sciences, including physics and cosmology. He imaginatively generalized Darwin’s theory such that evolution by the reproductive inheritance of variations checked by environmental pressure became evolution by the rhythmic propagation, or vibratory reiteration, of actual occasions along historical routes, whereby a particular occasion’s conformal physical prehensions of past actualities (=the inheritance of efficient causes) are integrated with its novel conceptual prehensions of future possibilities (=the formal causes of variation) into some emergent enduring pattern of experiential value. Whitehead argued that materialism could not survive its encounter with evolutionary theory., since the former implies merely the “purposeless and unprogressive” rearrangement of externally related substances and their accidental properties, while “the whole point of the modern doctrine is the evolution of the complex organisms from antecedent states of less complex organisms” (Whitehead  1967, 101). “The doctrine,” Whitehead continues, “cries aloud for a conception of organism as fundamental for nature” (ibid.).
Information theoretic accounts of the gap between matter and life provide some hope for a pathway forward, but without incarnating information into the meaningful horizon of experience enacted by living organisms, research programs seeking to analogize brain activity to computation end up having to conceive of information processing as some sort of quasi-conscious homunculus hovering above the neurochemistry of the brain and steering it around. For example, neuroscientists regularly describe information processing in the brain as “goal relevant,” “selective,” and “sensitive” (Sy et al. 2015, 122), all terms implying intentionality and purposefulness, even though the presuppositions of mechanistic biology upon which computational neuroscience rests says such powers are impossible. Luckily, taking information seriously does not require “assuming that abstract properties have physical potency,” as Terrence Deacon put it (Deacon 2012, 192).
Deacon is a strong emergentist who tries to dispel the homunculus and de-etherealize information by describing it not as an extra essence added to the physical but in terms of the “absential” features of an incomplete Nature:
“A counterintuitive figure/background reversal, focusing on what is absent rather than present, offers a means to repair some of the serious inadequacies in our conceptions of matter, order, life, work, information, representation, and even consciousness and conceptions of value” (Deacon 2011, 44).
Information is just what is absent from physically present matter. It is not involved in the push and pull of causal efficacy, but instead ‘constrains’ these physical interactions, acting as a formal and final cause that ratchets physics (thermodynamics) up a contragrade organizational gradient into chemistry (morphodynamics), biology (teleodynamics), and eventually full-blown conscious thought (intentionality). Like the enactivists, Deacon limits information processing to the living world, denying ententionality to the physical and chemical realms. He grants morphodynamic systems the ability to ‘fall up’ negentropic gradients of complexity toward the telic informational processes of living semiosis, but rejects the idea of any aim or value or elán implanted in matter prior to the emergence of life. Telos is added later and not baked in. Not the creative evolution of organisms, but vacuous bits of matter with no internal values…hurrying through space” (Whitehead  1968, 158) are fundamental for Nature.
It is here that the panpsychist integration of physics and experience goes further toward the naturalization of information by making sign interpretation, or in Whitehead’s terms, ‘prehension,’ an intrinsic part of cosmogenesis from the get go. Deacon criticizes Whitehead for projecting “micro humunculi” down to the level of quantum events, arguing that his panexperientialism obfuscates the need for an explanation of “why the [characteristics] of physical processes associated with life and mind [differ] so radically from those associated with the rest of physics and chemistry” (Deacon 2012, 79). Deacon admits that Whitehead in fact does offer an explanation for these differences in terms of the organizational complexity of enduring ‘societies’ of actual occasions of experience that emerge in the course of evolution. “Yet, if specific organizational complexity is what matters, then little explanatory significance is added by the assumption that some level of micro intentionality was suffused throughout all the component processes” (Deacon 2012, 78). While Deacon’s approach succeeds in narrowing the distance between physical causality and conscious intentionality, an explanatory gap still remains. Whitehead’s wager is that this gap is extreme enough to require fully undoing modern science’s “bifurcation of Nature” (Whitehead 1920, 30) by affirming that feeling or prehension is as intrinsic to natural processes as causality. Indeed, Whitehead’s experiential concept of prehension is meant to account for the very possibility of causal relation as such (Whitehead  1968, 164-165): prehension is what allows the real potentiality of the objectified past to pass back into the subjective immediacy of a new actual occasion of experience. Prehension is akin to the ‘information processing’ of computationalists, only it avoids the vagaries of their epiphenomenalism by rendering the detection of form as a process of feeling, thus embodying information in an experiential horizon. While his Philosophy of Organism does grant some degree of mentality to even the simplest of actual occasions, Whitehead’s panexperientialism doesn’t add anything extra to the natural world we find ourselves within: “the operation of mentality is primarily to be conceived as a diversion of the flow of energy” (Whitehead  1968, 168). In other words, mentality is an absential constraint upon energy’s otherwise entropic tendency. Were this entropic tendency the final word in Nature’s becoming, we would not be here to regret the fact. Whitehead is thus attempting to render the true nature of the physical universe transparent to us as the ongoing aesthetic achievement of a vast nexus of experiential occasions: “these unities of existence, these occasions of experience, are the really real things which in their collective unity compose the evolving universe, ever plunging into the creative advance” (Whitehead  1968, 151). Quarks, photons, protons, electrons, neutrons and the like appear to be our most ancient ancestors, close to the “primate organisms” (Whitehead  1967, 132) of our cosmic ecology. Out of their co-evolution emerged atoms, stars, and galaxies, all examples of the complex social achievements of actual occasions.. The evolution of these physical organisms proves that Nature’s capacity for emergent value and organizational complexity long predates the arrival of biological cells. These particle and astronomical organisms may be minimally or maximally conscious. The point is that at whatever scale it occurs, information processing is an experiential process, with the intensity of experience depending on the degree of integration of prehended data achieved by any given society of occasions.
3. The Physics of Experience: Avoiding Inflationary and Deflationary Accounts of Consciousness
“The doctrine I am maintaining is that neither physical nature nor life can be understood unless we fuse them together as essential factors in the composition of ‘really real’ things whose interconnections and individual characters constitute the universe.”
—Whitehead ( 1968, 150)
If physicalists are willing to take seriously the idea that human beings might not really be conscious, perhaps they can grant that it is no more absurd to entertain the possibility that stars and galaxies have minds. If Whitehead’s panexperientialist alternative turns out to have philosophical advantages over scientific materialism, perhaps we can learn to live with its mind expanding implications. After all, if materialism is true, we aren’t really alive, anyway. Another advantage of panexperientialism is that it can help philosophy avoid the excesses of Absolute idealism by not expanding mind too much.
Whitehead’s panexperientialism is an attempt to take consciousness at face value without unduly inflating or deflating its significance in the universe. The most inflationary accounts tend toward Absolute idealism, while the most deflationary tend toward eliminative materialism. The Kantian transcendental or critical approach views consciousness (with its categories of understanding and forms of intuition) as an a priori condition for knowledge of anything, including the physical world. It is thus an important compromise position, holding materialism at bay by preventing us from ever knowing anything about a mind-independent reality, while also checking the mind’s tendency to declare itself the ground of being. Kant admitted that via introspection we can only ever access an ‘empirical me,’ but he nonetheless posited a ‘transcendental I’ or Ego as the necessary correlate of everything thought or experienced, whether in myself (temporal intuition) or outside (spatial intuition). Kant’s transcendental Ego is no longer a clear and distinct substantial reality, as Descartes had imagined when he declared “I am a thing that thinks” (Descartes  1996, 24). So what is it? From James’ radically empirical perspective, the Kantian Ego “is simply nothing: as ineffectual and windy an abortion as Philosophy can show,” for if it be granted any other status, given Kant’s transcendental premises, there is little to prevent the Fichtean and Hegelian move to “call it the First Principle of Philosophy, to spell its name in capitals and pronounce it with adoration, to act, in short, as if [we are] going up in a balloon whenever the notion of it [crosses our] mind” (James 1890, 365). The Kantian compromise is thus an inherently unstable position. It saves mind from ever being reduced to matter, but at the cost of leaving us in total ignorance regarding the transcendental ground of our own consciousness and the substantial reality of Nature. Philosophers are left poised in a vulnerable state of metaphysical indecision, only a moderate dose of nitrous oxide away from floating into the mania of Absolute idealism, and only a mildly depressive mood away from succumbing to eliminative materialism. Might Whitehead’s “organic realism” (Whitehead  1978, 309) put philosophy on more solid experiential ground?
Presented with the general panpsychist hypothesis of a “pervasive perhaps ubiquitous” (Seager 2016, 229) subjectivity inherent in Nature, the first thing the incredulous tend to ask is whether the view entails that stones are conscious, or that tables and chairs stand at attention before us contemplating existence, or that spoons enjoy the flavor of the tea they stir. Few panpsychist philosophers actually uphold such views about stones and human artifacts, at least not without all the necessary qualifications (alchemists and astrologers notwithstanding). The proper panpsychist response to the skepticism of physicalists about the extent of mind’s reach into Nature is to ask whether it is really possible for them to conceive of their own consciousness as an illusion. For if the computational model of mind is true and experience contributes nothing to the functioning of the brain, if our consciousness is really just a complex set of what William Seager calls “bare recognitional capacities” evolutionarily elaborated “into a rich but delusive system of beliefs,” then when it comes down to it we human beings “are actually no more conscious than rocks” (Seager 2016, 231).
Which is more believable? That you and I are no more ‘alive’ than a pile of stones? That we and the stones are merely finite appearances in the eternal substance of the Absolute? Or that stones are more ‘alive’ than we think? From the perspective of Whitehead’s panexperiential organic realism, deflationary materialism and inflationary idealism are equally out of line. What, after all, does contemporary physics tell us about the materiality of a stone?: “[Vanished] from the field of ultimate scientific conceptions is the notion of vacuous material existence with passive endurance, with primary individual attributes, and with accidental adventures”; in short, physics has “[displaced] the notion of static stuff by the notion of fluent energy” (Whitehead  1978, 309). Stones, understood scientifically, are thus more like attenuated energy events whose relative stability is the effect of reiterated vibratory patterns of activity. For Whitehead, “the energetic activity considered in physics is the emotional intensity entertained in life” (Whitehead  1968, 168), though of course the emotional intensity realized in a stone is quite negligible due to the lack of any evolved organization for channeling and amplifying its scattered feelings into the more or less unified consciousness evident in animals. The physicist may retort that these patterns are merely mathematical equations and that we have no scientific basis for attributing experience or anything else concrete to the activity they describe. Indeed, many panpsychists are happy to admit that physics tells us only about the abstract aspects of matter and thus “can’t characterize the intrinsic nonstructural nature of concrete reality in any respect at all” (Strawson 2016, 85). In that case, it turns out ‘matter’ is among the most abstract ideas ever imagined by human minds. But in Whitehead’s way of thinking, this “divergence of the formulae about nature from the appearance of nature has robbed the formulae of any explanatory character” (Whitehead  1968, 154). Energetic activity is not just a mathematical abstraction but an abstract description of something real: “Nature is full-blooded. Real facts are happening” (Whitehead  1968, 144). Further, unlike some panpsychist readings of Russell’s neutral monism (Russell 1927), Whitehead’s process-relational rendering doesn’t claim experience is a ‘primary attribute’ or ‘intrinsic property’ of matter. This is because in Whitehead’s view, physics has moved beyond the substantialist view of matter, and talk of essential or accidental properties only made sense given such an ontology. The twentieth-century quantum and relativistic revolutions in physics dispensed with the ideas of “simple location” (Whitehead  1967, 51) and “nature at an instant” (Whitehead  1968, 145). There are no simply located, instantaneously present material particles or configurations of material particles, just as there are no simply located, instantaneous experiential states or properties. Both energy and experience are activities with fuzzy boundaries, and our panpsychist ontology should reflect this fact. Yet the substance-property ontology is difficult to shake, even for the physicists who know very well that it no longer captures what their equations are describing. The substance-property mode of thought is pervasive in Western philosophy. Descartes, so critical of Aristotle for other reasons, is fully infected by it, and many contemporary analytic philosophers who similarly consider their thinking to be free of any unexamined tradition nonetheless continue to construe reality in its terms. This mode of thought comes naturally since it is woven into the subject-predicate grammar of most of our languages. It is no surprise that Whitehead’s process-relational alternative is at first difficult to grasp.
While there was an “essential distinction between [substantial] matter at an instant and the agitations of experience,” with this conception of matter having been swept away, a door is opened to analogies between energetic activity and concrete experience (Whitehead  1968, 115). Experiences, like energy vectors, are intrinsically process-relational in that they always involve transition beyond themselves: they manifest in a “specious present” (Whitehead  1967, 104) as a tension between the actualized facts of an inherited past and the potential forms of an anticipated future. Whitehead turns to our own lived bodies for a more concrete characterization of physical process, since it is the human body that “provides our closest experience of the interplay of actualities in nature” (Whitehead  1968, 115). In addition to the grammar of our language, our visual experience of the immediately presented world reinforces the scientifically mistaken idea that reality is composed of substances with qualities. The grey stone is one of Whitehead’s favorite examples: ancient Greek philosophers perceived “the grey stone” and from that simple observation “evolved the generalization that the actual world can be conceived as a collection of primary substances qualified by universal qualities” (Whitehead  1978, 158). Modern natural philosophers beginning with Galileo elaborated this ontology into a conveniently bifurcated system of primary objective quantities (mass, velocity, dimensionality, etc.) and secondary subjective qualities (color, taste, value, etc.). Descartes’ mind/body dualism finished the job. Thenceforward it is not the stone that is grey, but the private quale of the perceiving subject that is grey. The stone itself is just an extensional lump obeying the fixed laws of gravity and chemical decay. Scrubbing Nature clean of all qualitative residues and tucking them safely away within conscious subjects allowed modern science to make truly remarkable progress explaining those aspects of Nature amenable to precise measurement and mathematical description (Goff 2017b, 12-14). But after a few hundred years of world-transforming progress, this powerful methodology still finds itself embarrassed by the hard problem. Consciousness appears to be “a strange intrusion into an otherwise well-behaved world” (Seager 2016, 234), though of course, it can hardly be said to have intruded if it was the methodology of modern science itself that initially excluded it from the physical world. Limited to the precise measurements afforded by strict sense-perception and to mathematical modeling, science finds no enjoyment, aim, or creativity in Nature, “it finds mere rules of succession” (Whitehead  1968, 154). But this is because, by design, science deals with only half the evidence of human experience.
In addition to the relatively superficial affordances of sense-perception granted us by the five outward facing senses, what Whitehead calls “perception in the mode of presentational immediacy” (Whitehead  1978, 121), he also describes a more primordial form of bodily experience or “sense-reception” (ibid., 113-114) referred to as “perception in the mode of causal efficacy” (ibid., 120). It is this latter form of human experience that modern science has all but ignored. When our eyes are functioning normally, they are transparent to the world. Nonetheless, it is evidently true that we see with our eyes. Causal efficacy is the feeling of our eyes blinking when we pull back the curtains and the sunlight floods onto our face. Presentational immediacy is the view of the meadow out the window after our eyes adjust. While presentational immediacy grants us perception of the grey stone as a geometrically projected patch of color, causal efficacy grants us perception of the grey stone’s weight when we pick it up in our hand, of the way this weight influences the muscle fibers and nerve endings in our arm as, “by channels of transmission and of enhancement” (ibid., 119), its ‘weightiness’ is delivered to the presiding occasions of the brain wherein we consciously feel it. “It is the accepted doctrine in physical science,” Whitehead tells us,
“that a living body is to be interpreted according to what is known of other sections of the physical universe. This is a sound doctrine, but it is double-edged. For it carries with it the converse deduction that other sections of the physical universe are to be interpreted in accordance with what we know of the human body” (ibid.).
Modern physics tells us that “the quiet extensive stone” is more complex than it at first appears to be. Were we able to apprehend the stone in a more direct way than that afforded by visual perception, it would reveal itself as a “society of separate molecules in violent agitation” (Whitehead  1978, 78). Picking up the stone grants us no more insight into its inner life, but the feeling of its weight in our hand grants us a clue with profound metaphysical implications. Our consciousness is not separate from but “intimately entwined in bodily life” (Whitehead  1968, 21). We consciously feel the stone because the human body, acting as an experiential amplifier, transmits the stone’s energetic activity along coordinated routes of actual occasions, accruing interpretive enhancements along the way, until the activity achieves final integration in a central occasion of experience. “The human body is thus achieving on a scale of concentrated efficiency a type of social organization, which with every gradation of efficiency constitutes the orderliness” found in the wider universe (Whitehead  1978, 119). Transmission of feelings within the body can thus be understood as analogous to the transmission of energy occurring in the rest of Nature. The body, after all, is part of and continuous with the rest of the external world, “just as much part of nature as anything else there—a river, or a mountain, or a cloud” (Whitehead  1968, 21).
Those seeking a truly naturalistic account of consciousness needn’t rush to deflationary explanations, whether eliminativist, epiphenomenalist, or emergentist. Such deflationary accounts would be understandable if the only alternatives available were dualism or idealism. Panpsychism, especially Whitehead’s panexperiential version, provides another option. It avoids the metaphysical travesty of dualism, the inflationary conjecture of idealism that “nature is mere appearance and mind is the sole reality,” and the deflationary conjecture of materialism that “physical nature is the sole reality and mind is an epiphenomenon” (Whitehead  1968, 150). It begins its explanation of consciousness modestly by examining our intimate feelings of bodily inheritance, and it concludes that these feelings provide a clue as to the functioning of energy in the rest of Nature. The conclusion may seem strange at first, but the philosophical payoff might just be worth it.
4. The Combination and Decomposition Problems for Panpsychism and Cosmopsychism: Bugs, or Features for Whitehead?
The philosophical payoff of panpsychism is that it dissolves the hard problem of consciousness, giving experience its proper place in Nature without undermining the scientific image of the universe. Indeed, panpsychism may have important advantages over materialism for interpreting contemporary physical cosmology (Segall 2018). But substance-property panpsychists have their own problem to deal with: the combination problem. Does Whitehead’s process-relational approach help solve it?
The solution to James’ original statement of the combination problem is already in James’ own statement: there is a 101st feeling, a “totally new fact,” and “the 100 original feelings might, by a curious physical law, be a signal for its creation, when they came together” (James 1890, 160). Whitehead’s process-relational ontology, in particular his genetic account of mutually sensitive prehensions (Whitehead  1978, 235ff), is an attempt to make good on James’ psychological insight by building it out into a coherent cosmological scheme.
Whitehead is neither a micropsychist nor a cosmopsychist exclusively. He tries to have it both ways. There is a universal soul, a psyche of the cosmos, a God of this world, and there are countless creatures creating in concert with it. Creativity transcends both, it is the source of all evolving parts, wholes, bodies, and souls. For Whitehead the combination problem becomes a logic of concrescence, a way of thinking change as more than just the rearrangement of pre-existing parts or the fragmentation of a pre-existing whole but as a genuine becoming, as an “emergent evolution” or “creative advance” (Whitehead  1978, 21, 30, 229) where neither wholes nor parts pre-exist their relations. Whitehead’s account of process is an account of combination and decomposition, of conjunction and disjunction. Process means the growing together of many objects into one subject, and the perishing of that subject back into many as a superject: “The many become one, and are increased by one” (ibid., 21). Concrescence is a cumulative process and not merely an additive one.
5. The Wonder Remains
“Philosophy begins in wonder. And, at the end, when philosophy has done its best, the wonder remains. There have been added, however, some grasp of the immensity of things, some purification of emotion by understanding.”
—Whitehead ( 1968, 168-169)
Whitehead is thus clearly an emergentist rather than constitutive panpsychist (Goff 2017a, 114). But it’s not that human consciousness is breaking the laws of physics, it’s that Nature’s ‘laws’ are queerer than our mechanical models let on. Like Deacon (2012) with his absential constraints in an incomplete Nature, Whitehead’s knowledge of mathematical physics led him to reject the causal closure of physics. Laws are habits emergent from the social activity of actual occasions of experience, not divine decrees from heaven imposed upon dead matter. But unlike Deacon, Whitehead goes further by granting life and mind some subtle congress with things from the beginning of time. Indeed, without life and mind Nature would have no time to evolve. The laws of physics are indifferent to life, mind, and time, so the show would be over before it even began.
Human consciousness is the achievement of the human body. The human body is the organizational achievement of a nexus of experiential occasions stretching back billions of years through the evolution of life on Earth, the birth of our Sun and planetary system, and the fusion of quarks into baryons, back even to the birth of God (Whitehead  1978, 348). Consciousness is human physics. Our philosophical conceptions, moral decisions, aesthetic creations, and religious concerns are not violations of the laws of physics (which are really statistical habits, anyway), no more so than the emergence of stars and galaxies was a violation of particle physics, or the emergence of cellular life was a violation of geology. “[Nature] is never complete. It is always passing beyond itself” (Whitehead  1978, 289).
Tononi, Giulio. Melanie Boly, Marcello Massimini, Christof Koch. 2016. “Integrated information theory: from consciousness to its physical substrate.” Nature Reviews Neuroscience 17:450–461.
Auxier, Randall. Gary Herstein. 2017. The Quantum of Explanation: Whitehead’s Radical Empiricism. New York: Routledge.
Bateson, Gregory.  2000. Steps to an Ecology of Mind. Chicaco: University of Chicago Press.
Bergson, Henri. 1999. Duration and Simultaneity: With Reference to Einstein’s Theory. Translated by Leon Jacobson. Manchester: Clinamen Press Ltd.
Brassier, Ray. 2007. Nihil Unbound: Enlightenment and Extinction. New York: Palgrave.
Brogaard, Berit. 2016. “In Search of Mentons: Panpsychism, Physicalism, and the Missing Link.” In Panpsychism: Contemporary Perspectives. Edited by Godehard Bruntrup and Ludwig Jaskolla. Oxford: Oxford University Press.
Bryant, Levi. Graham Harman. Nick Srnicek (eds.). 2011. The Speculative Turn: Continental Materialism and Realism. Melbourne, Australia: re.press.
Chalmers, D. J. 1995. “Facing up to the problem of consciousness.” In Journal of Consciousness Studies 2:200-19.
Churchland, Paul M. 1998. Matter and Consciousness. Cambridge, MA: MIT Press.
Churchland, Patricia S. 1986. Neurophilosophy: Toward a Unified Science of the Mind/Brain. Cambridge, MA: MIT Press.
Davidson, Donald. 1999. The Library of Living Philosophers: The Philosophy of Donald Davidson, edited by Lewis Edwin Hahn. Illonois: Open Court.
Deacon, Terrence. Incomplete Nature: How Mind Emerged from Matter. W. W. Norton: New York, 2012.
Deleuze, Gilles. Difference and Repetition. Translated by Paul Patton. Columbia University Press.  1994.
______. The Fold: Leibniz and the Baroque. University of Minnesota Press.  1993.
Dennett, Daniel. 2003. Freedom Evolves. New York: Viking.
_____. 2004. “‘Epiphenomenal’ Qualia?” In There’s Something About Mary: Essays on Phenomenal Consciousness and Frank Jackson’s Knowledge Argument, edited by P. Ludlow, Y. Nagasawa, and D. Stoljar. Cambridge, MA: MIT Press.
_____. 2017. From Bacteria to Bach and Back: The Evolution of Minds. New York: W. W. Norton.
Descartes, Rene.  1996. Meditations on First Philosophy. Cambridge, UK: Cambridge University Press.
_____.  1982. Principles of Philosophy. Translated by Valentine Rodger Miller and Reese P. Miller. London: Kluwar Academic Publishers.
Dewey, John. 1941. “The Philosophy of Whitehead.” In The Library of Living Philosophers: The Philosophy of Alfred North Whitehead, edited by Paul Arthur Schilpp, 641-700. Illinois: Open Court.
Frankish, Keith. 2016. “Illusionism as a Theory of Consciousness.” In Journal of Consciousness Studies, 23:11–39.
Goff, Philip. 2017a. “Panpsychism.” In The Blackwell Companion to Consciousness (2nd Edition), edited by Susan Schneider and Max Velmans. Malden, MA: John Wiley and Sons, LTD.
_____. 2017b. Consciousness and Fundamental Reality. Oxford: Oxford University Press.
Graziano, Michael S. A. 2019. “We Are Machines That Claim to Be Conscious.” In Journal of Consciousness Studies, 26:95-104.
Griffin, David Ray. 2008. Unsnarling the World-Knot: Consciousness, Freedom, and the Mind-Body Problem. Eugene, OR: Wipf and Stock Publishers.
Harman, Graham. 2018. Speculative Realism: An Introduction. Cambridge, UK: Polity Press.
Huxley, Thomas H. 1875. “On the Hypothesis that Animals are Automata.” In The Eclectic Magazine of Foreign Literature, Science, and Art, Volume 84. New York: Leavitt, Trow, & Company.
James, William. 1897. “Are We Automata?” In Mind, 4, 13:1-22.
_____. 1890. Principles of Psychology, Vol 1. New York: Henry Holt and Co.
Latour, Bruno. 2005. “What is Given in Experience?” In Boundary 2, 32: 222-237.
McGinn, Colin. 2006. “Hard Questions.” In: Consciousness and Its Place in Nature: Does Physicalism Entail Panpsychism?, edited by Anthony Freeman, 90–99. Exeter: Imprint Academic.
Nagel, Ernest. 1954. Sovereign Reason: And Other Studies in the Philosophy of Science. New York: The Free Press.
Nagel, Thomas. 2012. Mind and Cosmos: Why the Materialist Neo-Darwinian Conception of Nature is Almost Certainly False. Oxford: Oxford University Press.
Nicholson, Daniel J. John Dupré. 2018. Everything Flows: Towards a Processual Philosophy of Biology. Oxford: Oxford University Press.
Quine, Willard Van Orman. 1960. Word and Object. Cambridge, MA: MIT Press.
_____. 1985. The Time of My Life: An Autobiography. Cambridge, MA: MIT Press.
Russell, Bertrand. 1927. The Analysis of Matter. London: Kegan Paul, Trench, Trubner.
Seager, William. 2016. “Panpsychist Infusion.” In Panpsychism: Contemporary Perspectives, edited by Godehard Bruntrup and Ludwig Jaskolla. Oxford: Oxford University Press.
Segall, Matthew T. 2018. Physics of the World-Soul: Whitehead’s Adventure in Cosmology. Berkeley: Lulu Press.
Sellars, Wilfred. 1956. “Empiricism and the Philosophy of Mind.” In The Foundations of Science and the Concepts of Psychology and Psychoanalysis: Minnesota Studies in the Philosophy of Science (Volume 1), edited by H. Feigl and M. Scriven, 253–329. Minneapolis, MN: University of Minnesota Press.
Skrbina, David. 2005. Panpsychism in the West. Cambridge, MA: MIT Press.
Strawson, Galen (2017). Panpsychism: Contemporary Perspectives. Oxford University Press.
_____. 2018. “The Consciousness Deniers.” In The New York Review of Books.
Stengers, Isabelle. 2011. Thinking With Whitehead: A Free and Wild Creation of Concepts. Cambridge, MA: Harvard.
Sy, Jocelyn L.. James C. Elliot. Barry Giesbrecht. 2015. “Post-Perceptual Processing During the Attentional Blink is Modulated by Inter-Trial Task Expectances.” In Attention, predictions and expectations, and their violation: attentional control in the human brain, edited by Simone Vossel, Joy J. Geng, Karl J. Friston. Frontiers Media.
Thompson, Evan. 2007. Mind in Life: Biology, Phenomenology, and the Sciences of Mind. Cambridge, MA: Harvard University Press.
Varela, Francisco. Evan Thompson. Eleanor Rosch.  2016. The Embodied Mind. Cambridge, MA: MIT Press.
Whitehead, Alfred North. 1920. The Concept of Nature. Cambridge, MA: Cambridge University Press.
_____.  1978. Process and Reality. New York: The Free Press.
_____.  1967. Adventures of Ideas. New York: The Free Press.
_____.  1968. Modes of Thought. New York: The Free Press.
_____.  1967. Science and the Modern World. New York: The Free Press.
Wittgenstein, Ludwig. 1922. Tractatus Logico-Philosophicus. London: Harcourt, Brace, Incorporated.
Here’s the recording of a lecture that Becca Tarnas and I delivered last night for the Atlanta Astrological Society.
Here are some relevant links if you want a more in depth discussion on some of what I mention in this lecture:
Notes from a talk I gave at CIIS this past March titled “Politics and Pluralism in the Anthropocene”
Here’s the video of the whole panel:
Foucault on Hegel:
“[T]ruly to escape Hegel involves an exact appreciation of the price we have to pay to detach ourselves from him. It assumes that we are aware of the extent to which Hegel, insidiously perhaps, is close to us; it implies a knowledge, in that which permits us to think against Hegel, of that which remains Hegelian. We have to determine the extent to which our anti-Hegelianism is possibly one of his tricks directed against us, at the end of which he stands, motionless, waiting for us.” (Discourse on Language, Inaugural Lecture at the Collège de France, 1970-1971. tr. A. M. Sheridan Smith)
Begin with Hegel’s claim to have achieved absolute knowledge of Spirit, and to at least foreseeing the becoming concrete of this Spirit in the historical, social, and ethical life of the human community.
Marx read Hegel upside down, but still read Hegel. He was a materialist, but a dialectical materialist who recognized the potential of the human spirit, and this potential’s degradation and alienation from itself at the hands of capitalism. Marx tried to shake human beings awake, out of their slumber, out of the false consciousness that commodifies labor, life, and value.
It is not easy to do better than Hegel and Marx in terms of understanding, diagnosing, and prescribing action to overcome the contradictory situation in which humans find themselves as neoliberal capitalist subjects. But the dawning realization that we live in the time of the Anthropocene is fundamentally changing our situation. We can no longer talk about the nonhuman world, about what used to be called “Nature,” as though it was something separate from us, some kind of inert background or stage upon which human history progresses. As good dialecticians, Hegel and Marx fully recognized this entanglement of the human and the physical world, but they did so in a rather anthropocentric way that still presupposed and celebrated the idea of mastering nature. The Anthropocene signals, yes, the end of history, but also the beginning of (or at least the beginning of human recognition of) what Latour refers to as geostory. From Latour’s point of view, Hegel would never have expected our current situation, where Spirit, after its millennial march of dialectical progress, suddenly finds itself at risk of being suffocated, sublated, by carbon dioxide. As Latour describes it, the ecological crisis is pushing us into a profound mutation in our relation to the world. When the world as it has been known to the Western metaphysical project ends, we are left not with no world, but with many worlds. For Latour, politics is the composition of common worlds through the negotiation of differences. Political negotiation cannot be undertaken with the presupposition that unity has somehow already been achieved. If politics fails, we are left with a war of the worlds. A pluralistic politics asks us to forgo the desire for the premature unification of the world, to accept that “the world” has ended and diplomatic negotiation is the only viable way of “worlding.” Ours is always a world-in-process, and any unity we do achieve is fragile and must be continually re-affirmed and maintained.
Latour has been deeply influenced by William James. James positioned his ontological pluralism against Hegel and Marx’s dialectical monisms. William James was appreciative of Hegel, but certainly he was an counter-Hegelian thinker. As far as Marx goes, James was too American to ever fully reject at least the individualist spirit of capitalism, even if he was suspicious of capitalism’s larger cultural impact and its relation to American imperialism. In a letter to H. G. Wells in 1906, for example, James lamented “the moral flabbiness born of the exclusive worship of the bitch-goddess SUCCESS.” James thought worship of success, by which he meant money, was “our national disease.” James championed the individual, but an individual who is sympathetic to meeting and being transformed by novel differences, whose selfhood is leaky and perforated by human and nonhuman otherness, whose identity is always in-the-making and open to question and revision.
James on excess: “Every smallest state of consciousness, concretely taken, overflows its own definition. Only concepts are self-identical; only ‘reason’ deals with closed equations; nature is but a name for excess; every point in her opens out and runs into the more; and the only question, with reference to any point we may be considering, is how far into the rest of nature we may have to go in order to get entirely beyond its overflow. In the pulse of inner life immediately present now in each of us is a little past, a little future, a little awareness of our own body, of each other’s persons, of the sublimities we are trying to talk about, of earth’s geography and the direction of history, of truth and error, of good and bad, and of who knows how much more? Feeling, however dimly and subconsciously, all these things, your pulse of inner life is continuous with them, belongs to them and they to it. You can’t identify it with either one of them rather than with the others, for if you let it develop into no matter which of those directions, what it develops into will look back on it and say, ‘That was the original germ of me.’” (A Pluralistic Universe)
James leans strongly in the direction of particular, unique, once-occurrent individuals (even if he does not see individuals as autopoietic, but as sympoietic). In contrast, some historical performances of communism have leaned in the other direction, toward some abstract conception of communal will, and when individuals stood in the way of this abstract will, as we saw in Stalin’s Soviet Union and Mao’s China, they were crushed. Our capitalist society claims to prize the individual highest, but it is corporate individuality that we really cherish. We human individuals are mere cogs in the labor machine, and Earth is a store of raw materials and a garbage heap. So either way, in either situation, capitalism or communism, human and nonhuman communities and individuals are in trouble.
Our challenge today, in the Anthropocene, is to think individuality and community concretely, to think relation, difference, and particularity concretely. Normally thinking seeks out universals, essences, substances, and to the extent that the Western metaphysical project has sought out universals, essences, and substances that failed to align with the particular contours of the sensory, social worlds that we inhabit, it has done great violence those worlds. As a result of the failure of our ideas and concepts to cohere with reality—that is, to sympoietically relate to the communities of actual organisms composing the living planet—these concepts have functioned to destroy them. Humans, whether we like it or not, are in community with these organisms, our worlds overlap and perforate one another; we touch interior to interior, my inside bleeding into your inside bleeding into all nonhuman insides. But our subjectivities do not just add up or sum to some seamless Globe-like Mind. Gaia, Latour is constantly reminding his readers, is not a Globe! To the extent that we are all internally related to one another, we form a network of of entangled, overlapping perspectives, where each perspective is still unique and once-occurrent, novel; and yet each is also related to what has come before and will be related to what comes next. We are individuals-in-communion, communities whose wholeness subscends the individuals who compose them. Subscendence is a concept developed by Timothy Morton to refer to the way that wholes, like Gaia, are actually less than the sum of their parts. He calls this “implosive holism” and contrasts it with “explosive holism,” the sort of holism that led Stalin to murder millions of individuals for the sake of the Soviet Union, or that leads some environmentalists to emphasize saving species or even the whole planet without paying enough attention to individual organisms (a species doesn’t feel pain; only individual organisms feel pain, etc.).
So the question becomes, how do we think pluralism, difference, and diversity concretely, and not abstractly. Because when we think particular identities or individuals abstractly, we do violence to them, we try to universalize them in an overly abstract way without being sensitive to their unique contours. This is a form of reductionism. We can reduce individuals “up” to the whole, or reduce them “down” to their parts. Pluralism is trying to find a middle path between both forms of reductionism: It seeks a “strung-along” sort of holism (as James put it), not a global or continuous holism where each thing is connected to everything else in exactly the same way. Instead, as Donna Haraway puts it, “Nothing is connected to everything” even though “everything is connected to something.”
Thinking pluralism concretely means stepping out of a sense of exclusively human society, out of the self-enclosed social bubble that used to insulate us from any access whatsoever to something called Nature, or “the environment” standing in wait “over there” for science to objectify into knowledge or for the economy to commodify into money. Thinking pluralism concretely means stepping outside of the monetary monism of contemporary capitalism, where all value is reduced to exchange value in the human marketplace, to instead become part of a democracy of fellow creatures, as Whitehead puts it, where values pervade the biosphere, and “Nature” is no longer just a realm of inert, law-abiding facts but of creative, expressive agencies. Thinking pluralism concretely means walking out of the old Copernican universe, forgetting the mastery-seeking knowledge supplied by the monotheistic gaze of Science, in order to inhabit a new cosmos composed of infinitely many perspectives, more a pluriverse than a universe.
“The doctrine of relativity affects every branch of natural science, not excluding the biological sciences. . . . Relativity, in the form of novel formulae relating time and space, first developed in connection with electromagnetism. . . . Einstein then proceeded to show its bearing on the formulae for gravitation. It so happens therefore that owing to the circumstances of its origin a very general doctrine is linked with two special applications.”
–Whitehead (The Principle of Relativity, 3).
One of the biggest surprises for me upon reading Auxier and Herstein’s book The Quantum of Explanation was learning that Whitehead’s theory of extension (or “mereotopology” as it has come to be called) has been taken up by computer scientists working in the field of robotic vision (see for example the work of Ian Pratt-Hartmann).
“It is a widely acknowledged fact in this sub-discipline that Alfred North Whitehead’s work on extension is foundational for their enterprise. Our experience has been that Whitehead scholars are simply astounded to learn of this fact. Yet we should have expected and even predicted such a connection” (QE 90).
Guilty as charged. While I think I got things mostly right in section 3.2 of my dissertation (“From Geometric Conditions of Possibility to Genetic Conditions of Actuality”), the promising application of Whitehead’s topological scheme to robotic vision certainly brings this aspect of his project into sharper focus for me. As a radical empiricist, Whitehead was searching for a formal account of our concrete experience of projectively related extensa. We are finite creatures with limited sensory organs and processing capacity. We do not experience the world of spatial relations in terms of infinitesimal points or the geometrical schemes built up from such points. Rather, what we encounter in our immediate experiential field are the intuitive whole-part relational structures formalized by non-metrical projective geometry.
Following Einstein’s articulation of the special and general theories of relativity (in 1905 and 1916, respectively), and his problematic “mono-metric” identification of a 4-D geometrical model with physical space-time*, Whitehead pursued his theory of extension with renewed urgency. Somehow, the uniformity of spatial geometry had to be preserved, else scientific measurement would become impossible. Einstein did not appear to realize that allowing the contingent warping of space by massive objects undermined the fundamental logical requirements of measurement: that space have a necessary and universal structure (or, as Auxier and Herstein put it, “we must have a standard unit of spatial comparison for conjugacy…and standard(s) of spatial projection” so as to bring this unit into comparison with whatever we are trying to measure [QE 102]). By collapsing the difference between physical space and his favored geometrical scheme, Einstein made the structure of spatial geometry contingent upon randomly arrayed masses.
“We must know the complete distribution of matter and energy in the universe prior to knowing its geometry. But we must have a comprehensive grasp of this geometry in order to discover this distribution. As Whitehead pointed out, with General Relativity as our theory of space and gravity, we are saddled with a situation where we must first know everything before we can know anything” (QE 104).
Einstein’s “mono-metric” model has been one of the most successful in the history of science. But because of the unexpected observations of the rotational velocity of galaxies and of cosmic inflation rates, its theoretical supremacy has begun to be seriously questioned. Some astrophysicists have attempted to save the theory by inventing “dark matter” and “dark energy” to explain the missing mass that would bring observations back into agreement with Einstein’s theory. Auxier and Herstein refer to these inventions as “an especially unhappy piece of nonsense” (QE 20). I’m sympathetic, but I wouldn’t go quite that far. To my mind, these invented entities are akin to the epicycles of Ptolemaic astronomy. In other words, these exotic and invisible forms of mass/energy (which supposedly compose ~96% of the universe) are postulated ad hoc in an attempt to “save the appearances” (as ancient astronomers used to say). Ancient astronomers were tasked by Plato with explaining the seemingly erratic motion of the planets in terms of a theoretical model composed only of uniform circular motions. When new planetary observations conflicted with the model, more circles were added (epicycles) to bring the model back into alignment with appearances. One view of science is that it is just about refining existing theoretical presuppositions to fit new observations, gradually approaching a perfect identity between model and reality. In this sense, the addition of epicycles to match observations could continue indefinitely. After all, Ptolemy’s geocentric model was more accurate than Copernicus’ heliocentric model (which itself still required epicycles until Kepler and Newton updated his math). The geocentric model is still accurate enough that modern planetarium projectors (invented in the 1920s by a company in Jena, Germany) continue to utilize it, reproducing Ptolemy’s deferents and epicycles with their internal gears and motors (see also).
But as Karl Popper taught us, scientific theories must be subject to empirical falsification. The eternal circular orbits of Ptolemy’s model fall out of phase with the long-term evolution of planetary orbits, while the (updated) heliocentric model accommodates this evolution well. As Thomas Kuhn, another great philosopher of science, taught us, the history of science is not just about the gradual refinement of old theories to fit new observations in an asymptotic convergence of model to reality; rather, this history is also characterized by periods of revolutionary crisis as aging paradigms are supplanted by deeper, wider, more elegant and inclusive explanatory perspectives. Einstein’s genius was to bring the reigning Newtonian theory of gravity into alignment with Maxwell’s theory of electromagnetism. A deeper theory of space was born. But in a sense, despite many other successful observational predictions, empirical falsification is exactly what happened to Einstein’s gravitational theory when it failed to accurately predict the observed rotational velocity of galaxies. However, because this darling model had made a number of other accurate predictions, and because no widely accepted alternative paradigm was on hand, astrophysicists decided to fudge the numbers by inventing new free parameters, new epicycles, to bring the theory back into alignment with observations. Appearances were thereby saved, but at the cost of conjuring into existence an entire universe (or 96% of one, at least) of cold and dark, that is, unobservable, matter/energy.
Even though he did formulate a “bimetric” alternative in 1922 (QE 109), Whitehead’s problem is not with Einstein’s model. This isn’t a “scientists have been wrong before, so why should we trust them now?” argument. Science is about modeling. In some sense, scientific models are always wrong. That’s the name of the game, after all: build a model and throw it against reality until it breaks. Then study why it broke until you find a new model that doesn’t break as quickly. Gradually, more robust, inclusive models emerge. Rather, Whitehead’s problem is with the philosophically naive “model-centrism” that leads scientists to equate their favored model with reality in a dogmatically literalistic way. We should never assume the reigning physical models of the universe offer a final account of the way things are (especially when today’s two most successful models, relativity and quantum theory, remain irreconcilable). Science is not ontology: science is a method of inquiry involving the making and breaking of toy models.
The dogmatic equation of a favored geometrical model with physical reality not only undermined the logical basis of measurement, it led Einstein to dismiss our concrete experience of an irreversible flow of time as nothing more than a “stubbornly persistent illusion.” This is Whitehead’s famous “fallacy of misplaced concreteness” writ large. Einstein’s unquestioned commitment to the classical “spectator theory of knowledge” prevented him from grasping the profoundly relational implications of his new theory of space. He upheld the old Galilean-Cartesian view of a bifurcated Nature, construing our consciousness as somehow external to a cosmos that we can only ever confusedly experience. Whitehead offers an alternative, fully relational epistemology and ontology that re-embeds experience in the cosmos: we are creative participants in a cosmogenetic relational nexus.
Instead of rushing to eliminate experience from our understanding of a relativistic (or relational) reality, Whitehead carefully examined the hidden epistemic presuppositions and metaphysical requirements of Einstein’s more specific application of relativity to the physics of light and gravitation. The result of his examination was eventually assembled in Process and Reality as the fourth category of explanation, a truly general principle of relativity: “it belongs to the nature of a ‘being’ that it is a potential for every ‘becoming'” (PR 22). Obviously, the importance of Whitehead’s fourth category of explanation (of which there are 26 others) can only be understood within the total gestalt of his categoreal scheme (which includes the category of the ultimate: Creativity; eight categories of existence, among which the most important are eternal objects and actual occasions; and nine categories of obligation). Whitehead’s categoreal scheme is laid out in Part I of Process and Reality as something like an opening credit roll listing the conceptual dramatis personae who, in Part II, will take the stage to exemplify their adequacy. But I’m not going to run through the whole dress rehearsal right now (for a helpful exegesis of Whitehead’s first four categories of explanation, see pgs. 108-110 of QE). Suffice it to say that Whitehead’s principle of relativity expresses the truth that everything co-exists in a web of relatedness, whether actually or potentially.
Auxier and Herstein:
“This is the principle that Einstein and his devotees have abandoned: not the mathematical expression of their physical model; that model is itself only an application of what has become the standard dogma of orthodox cosmology, with its narrowly defined approach to the interpretation of a truncated representation of experience. Rather, physical cosmology has left behind the full principle of relativity and its unqualified commitment to the incurable relatedness of the real. That abandonment comes in the truncation of experience at the root of their largely unexpressed theory of experience [i.e., the theory of the bifurcation of Nature]. For one cannot have a universal principle of relativity—applicable to all that is real—unless one takes experience in its real, relational totality. Experience—both actual and potential—is exactly the kind of reality that falls under the principle of relativity. One cannot take the metaphysical principle of relativity seriously unless one is a radical empiricist” (QE 110).
In The Quantum of Explanation, Auxier and Herstein have brilliantly succeeded in elucidating the features of a radically empirical cosmology. As Whitehead reminds us early and often in Process and Reality, the purpose of philosophy is not to explain away the existence of the concrete by reduction to the abstract, but to explain the emergence of abstraction from concretion. The proper questions are: how does concrete fact participate in general form and how are general forms exemplified in concrete facts?
For a longer discussion of Whitehead’s radical empiricism a.k.a. relational realism, see my essay “Retrieving Realism: A Whiteheadian Wager.”
*It has been brought to my attention that the matter of whether Einstein thought the physics of gravitation is reducible to the geometry of space-time is not so clear cut. See for example: “Why Einstein did not believe that general relativity geometrizes gravity” by Lehmkuhl. The research continues…
Auditors are welcome, though space is limited. Email me at firstname.lastname@example.org for more information.
One of our core texts in this course will be my Physics of the World-Soul (a new third edition soon to be published).
I’ll be teaching another short course at Schumacher College in the UK the week of April 22nd-26th, 2019.
Here’s a link if you’re interested in registering:
Here’s what I’ll be teaching on:
“The Evolution of Consciousness and the Cosmological Imagination”
This week-long course will trace the evolution of consciousness in the West from ancient Greece through to the present. The goal is twofold: to understand the historical process whereby humanity severed itself from a meaningful universe and to re-ignite the cosmological imagination allowing us to reconnect to the soul of the world. The course begins by exploring Plato’s cosmology and theory of participation and moves on to consider the Scientific Revolution and the Romantic reaction to it. It concludes with a study of several contemporary efforts to re-enchant the cosmos by grounding human consciousness back in the more-than-human creative process responsible for generating it. In addition to Plato, the course draws upon the archetypal astronomy of Johannes Kepler, the Naturphilosophie of Goethe and Schelling, the nature poetry of Coleridge and Wordsworth, the esoteric philosophy of Rudolf Steiner and Owen Barfield, the process philosophy of Alfred North Whitehead, and the contemporary participatory theory of Jorge Ferrer.
*featured image above by Jakob Boehme
A Slovakian visual artist, András Cséfalvay, recently invited me to submit a video for inclusion in his upcoming exhibition in Prague focused on the cultural significance of Pluto (my video is embedded below). Back in 2006, Pluto was demoted from its planetary status by the International Astronomical Union. Following the flyby of NASA’s New Horizons spacecraft in 2015, the scientific and popular controversy over Pluto’s classification was reignited in part because Pluto proved to be more lively (i.e., geologically active) than astronomers had assumed.
Shortly after I accepted Cséfalvay’s invitation, a group of planetary scientists led by Philip Metzger (a physicist at my alma mater the University of Central Florida) published a paper that wades right into the center of the conflict. According to Metzger, “The IAU definition would say that the fundamental object of planetary science, the planet, is supposed to be defined on the basis of a concept that nobody [no planetary scientist] uses in their research.”
Pluto finds itself caught in the middle of a clash of paradigms: many (not all*) astronomers stand on one side arguing that the defining characteristic of a planet is that it clears its own orbit of other objects (Pluto does not), while on the other side planetologists like Metzger classify planets based on their spherical shape.
Metzger explains: “It turns out [sphericality] is an important milestone in the evolution of a planetary body, because apparently when it happens, it initiates active geology in the body.”
Metzger goes on to say that the IAU definition is too sloppy, since if taken literally, there would be no planets at all in our solar system (none of the bodies orbiting our Sun fully clears its own orbit).
So what is Pluto? Scientifically speaking, I think the planetary scientists have come up with a better classificatory scheme. As a process thinker, I agree with them that the best way to understand the essence of a planet is in terms of its evolutionary history. But my interest in this debate is more philosophical. I think about this controversy in the context of an interplay between the ontologies of multiple paradigms. For astronomers, Pluto is a mere “dwarf planet”; for planetologists, Pluto is a geologically active planet; and for astrologers, Pluto is Hades, Lord of the Underworld, the archetypal power of death and rebirth.
Having been influenced by the work of Bruno Latour (in this case, see especially An Inquiry into Modes of Existence), I see the philosopher’s role as akin to that of a diplomat. I ask: is it possible to translate between a plurality of paradigms and to avoid the need to collapse our view of Pluto into Newton’s single vision? Can Pluto be a telescopically-enhanced point of light in the sky, a geologically active planetary body, and King of Hell all at once?
I also think about this debate as it relates the transcendental conditions of knowledge. For Kant, a table of twelve categories and our fixed intuitions of space and time delimits what we can know. The mind structures a priori everything we are capable of knowing about Nature. In 2006, the International Astronomical Union acted as a sort of institutionalized enforcer of transcendental limits, establishing the classificatory rules that the rest of the community of knowledge producing scientists is supposed to obey. Archetypal astrologers transmute the transcendental approach even more radically, replacing Kant’s twelve categories with the ten planetary archetypes (the Sun and Moon are included along with Mercury through Pluto). These cosmically incarnate archetypal powers condition each individual knower, stamping each of us with a unique planetary signature at the moment of our emergence from the womb. The participatory epistemology underlying the archetypal cosmological paradigm implies new conditions of experiential access to reality. Our knowing is mediated not just by mental categories, but by archetypal powers inhabiting Nature as much as mind.
Metzger’s et al.’s recent scientific paper is titled “The Reclassification of Asteroids from Planets to Non-Planets.” Here’s the abstract:
It is often claimed that asteroids’ sharing of orbits is the reason they were re-classified from planets to non-planets. A critical review of the literature from the 19th Century to the present shows this is factually incorrect. The literature shows the term asteroid was broadly recognized as a subset of planet for 150 years. On-going discovery of asteroids resulted in a de facto stretching of the concept of planet to include the ever-smaller bodies. Scientists found utility in this taxonomic identification as it provided categories needed to argue for the leading hypothesis of planet formation, Laplace’s nebular hypothesis. In the 1950s, developments in planet formation theory found it no longer useful to maintain taxonomic identification between asteroids and planets, Ceres being the primary exception. At approximately the same time, there was a flood of publications on the geophysical nature of asteroids showing them to be geophysically different than the large planets. This is when the terminology in asteroid publications calling them planets abruptly plunged from a high level of usage where it had hovered during the period 1801 – 1957 to a low level that held constant thereafter. This marks the point where the community effectively formed consensus that asteroids should be taxonomically distinct from planets. The evidence demonstrates this consensus formed on the basis of geophysical differences between asteroids and planets, not the sharing of orbits. We suggest attempts to build consensus around planetary taxonomy not rely on the non-scientific process of voting, but rather through precedent set in scientific literature and discourse, by which perspectives evolve with additional observations and information, just as they did in the case of asteroids.
It struck me that this line of inquiry may have profound implications for the future of astrological theory and practice, specifically the way we understand the difference between the ten planetary archetypes and the indefinite number of asteroidal archetypes. Does the unique geophysical history underlying planet formation correlate with a uniquely potent and living archetypal signature (that of a planetary god or goddess), such that astroids and dwarf planets (i.e., non-spherical bodies) must be treated more as underdeveloped demigods or shattered spirits? My limited exposure to astrologers who foreground asteroids suggests they would bristle at the idea of them being less archetypally significant than planets.
Or, if Pluto is a dwarf planet or an asteroid, perhaps that says something profound about the evolutionary power of these chaotically orbiting fragments of rock and ice. They are reminders of the violent history of our solar system, of the fact that tremendous destruction (i.e., an entire eon composed of nothing but mega-collisions between orbiting bodies, appropriately referred to by geologists as the Hadean) prepares the way for the miraculous emergence of more or less orderly living worlds.
In any event, this whole dispute between astronomers and planetary scientists about the status of Pluto has me wondering what experts in a third and for too long marginalized paradigm, astrology, can contribute to the conversation.
Here’s the video I submitted to Cséfalvay for his Prague exhibition:
*For example, Harvard astronomer Owen Gingerich, the Chair of the IAU committee that voted to demote Pluto, disagreed with his own committee on this issue.