Three Pound Brain

No bells, just whistling in the dark…

Month: January, 2014

Love, God, and Entropy

by rsbakker

An infinitesimal speck becomes everything, and the universe blooms and burns with the hard light of innumerable stars autocannibalizing, finally choking on iron and exploding into clouds that collapse into new stars surrounded by discs of parental debris that they cook and cook, forcing matter to flee into ever more complicated forms to maximize entropy, a constant pressure culminating in the generational replication of structure, and a vast proliferation of different entropic maximizations–the sum of life!–each more efficient than the previous, adaptations stacked upon adaptations, mapping ever more possibilities of structure and dynamics, morphology and behaviour, until the sheer complexity of the latter undoes the integrity of the former, and behaviour begins complicating without material constraint, collapsing the entropic interval, becoming a black hole of a more horrifying sort, an entropy maximizing God–Death–grown to a devouring haze that encompasses stars, galaxies, clusters and superclusters, consuming filaments and more, until passing at long last through its own jaws and fading into the infrared wash that is the end of the universe.

Or you can read this, courtesy of ochlocrat…

Life as Perpetual Motion Machine: Adrian Johnston and the Continental Credibility Crisis

by rsbakker

In Thinking, Fast and Slow, Daniel Kahneman cites the difficulty we have distinguishing experience from memory as the reason why we retrospectively underrate our suffering in a variety of contexts. Given the same painful medical procedure, one would expect an individual suffering for twenty minutes to report a far greater amount than an individual suffering for half that time or less. Such is not the case. As it turns out duration has “no effect whatsoever on the ratings of total pain” (380). Retrospective assessments, rather, seem determined by the average of the pain’s peak and its coda.

Absent intellectual effort, the default is to remove the band-aid slowly.

Far from being academic, this ‘duration neglect,’ as Kahneman calls it, places the therapist in something of a bind. What should the physician’s goal be? The reduction of the pain actually experienced, or the reduction of the pain remembered. Kahneman provocatively frames the problem as a question of choosing between selves, the ‘experiencing self’ that actually suffers the pain and the ‘remembering self’ that walks out of the clinic. Which ‘self’ should the therapist serve? Kahneman sides with the latter. “Memories,” he writes, “are all we get to keep from our experience of living, and the only perspective that we can adopt as we think about our lives is therefore that of the remembering self” (381). As he continues:

“Confusing experience with the memory of it is a compelling cognitive illusion—and it is the substitution that makes us believe a past experience can be ruined. The experiencing self does not have a voice. The remembering self is sometimes wrong, but it is the one that keeps score and governs what we learn from living, and it is the one that makes decisions. What we learn from the past is to maximize the qualities of our future memories, not necessarily of our future experience. This is the tyranny of the remembering self.” 381

There’s many, many ways to parse this fascinating passage, but what I’m most interested in is the brand of tyranny Kahneman invokes here. The use is metaphoric, of course, referring to some kind of ‘power’ that remembering possesses over experience. But this ‘power over’ isn’t positive: the ‘remembering self’ is no ‘tyrant’ in the interpersonal or political sense. We aren’t talking about a power that one agent holds over another, but rather the way facts belonging to one capacity, experiencing, regularly find themselves at the mercy of another, remembering.

Insofar as the metaphor obtains at all, you could say the power involved is the power of selection. Consider the sum of your own sensorium this very moment—the nearly sub-audible thrum of walled-away urban environs, the crisp white of the screen, the clamour of meandering worry on your margins, the smell of winter drafts creeping through lived-in spaces—and think of how wane and empty it will have become when you lie in bed this evening. With every passing heartbeat, the vast bulk of experience is consigned to oblivion, stranding us with memories as insubstantial as coffee-rings on a glossy magazine.

It has to be this way, of course, for both brute biomechanical and evolutionary developmental reasons. The high-dimensionality of experience speaks to the evolutionary importance of managing ongoing environmental events. The biomechanical complexity required to generate this dimensionality, however, creates what might be called the Problem of Indisposition. Since any given moment of experience exhausts our capacity to experience, each subsequent moment of experience all but utterly occludes the moment prior. The astronomical amounts of information constitutive of momentary experience is all but lost, ‘implicit’ in the systematic skeleton of ensuing effects to be sure, but inaccessible to cognition all the same.

Remembering, in other words, is radically privative. As a form of subsequent experiencing, the machinery involved generating the experience remembered has been retasked. Accordingly, the question of just what gets selected becomes all important. The phenomenon of duration neglect noted above merely highlights one of very many kinds of information neglected. In this instance, it seems, evolution skimped on the metacognitive machinery required to reliably track and rationally assess certain durations of pain. Remembering the peak and coda apparently packed a bigger reproductive punch.

Kahneman likens remembering to a tyrant because selectivity, understood at the level of agency, connotes power. The automaticity of this selectivity, however, suggests that abjection is actually the better metaphor, that far from being a tyrant, remembering is more a captive to the information available, more a prisoner in Plato’s Cave, than any kind of executive authority.

If any culprit deserves the moniker of ‘tyrant’ here, it has to be neglect. Why do so many individuals  choose to remove the band-aid slowly? Because information regarding duration plays far less a roll than information regarding intensity. Since the mechanisms responsible for remembering systematically neglect such information, that information possesses no downstream consequences for the machinery of decision-making. What we have traditionally called memory consists of a fractionate system of automata scattered throughout the brain. What little they cull from experiencing is both automatic and radically heuristic. Insofar as the metaphor of ‘tyrant’ applies at all, it applies to the various forms of neglect suffered by conscious cognition, the myriad scotomas constraining the possibilities of ‘remembering experience’—or metacognition more generally.

Kahneman’s distinction wonderfully illustrates the way the lack of information can have positive cognitive effects. Band-aids get pulled slowly because only a spare, evolutionarily strategic fraction of experiencing can be remembered. We only recall enough of experience, it seems safe to assume, to solve the kinds of problems impacting our paleolithic ancestors’ capacity to reproduce. This raises the general question of just what kinds of problems we should expect metacognition—given the limitations of its access and resources—to be able to solve.

Or put more provocatively, the question that philosophy has spent millennia attempting to evade in the form of skepticism: If we don’t possess the metacognitive capacity to track the duration of suffering, why should we expect theoretical reflection to possess the access and capacity to theoretically cognize the truth of experience otherwise? Given the sheer complexity of the brain, the information consciously accessed is almost certainly adapted to various, narrow heuristic functions. It’s easy to imagine specialized metacognitive access and processing adapting to solve specialized problems possessing reproductive benefits. But it seems hard to imagine why evolution would select for the ability to theoretically intuit experience for what it is. Even worse, theoretical reflection is an exaptation, a cultural achievement. As such, we should expect it to be a naive metacognitive consumer, taking all information absent any secondary information regarding that information’s sufficiency.

In other words, not only should we expect theoretical reflection to be blind, we should also expect it to be blind to its own blindness.

It is this question of neurobiological capacity and evolutionary problem-solving that I want to bring to Adrian Johnston’s project to materially square the circle of subjectivity—or as he puts it, to secure “the possibility of a gap between, on the one hand, a detotalized, disunified plethora of material substances riddled with contingencies and conflicts and, on the other hand, the bottom-up surfacing out of these substances of the recursive, self-relating structural dynamics of cognitive, affective, and motivational subjectivity—a subjectivity fully within but nonetheless free at certain levels from material nature” (209).

I’ve considered several attempts by different Continental philosophers to deal with the challenges posed by the sciences of the mind over the past three years: Quentin Meillasoux in CAUSA SUIcide, Levi Bryant in The Ptolemaic Restoration, Martin Hagglund in Reactionary Atheism, and Slavoj Zizek in Zizek Hollywood, each of which has received thousands of views. With Meillasoux I focussed on his isolation of ‘correlation’ as a problematic ontological assumption, and the way he seemed to think he need only name it as such, and all the problems of subjectivity raised by Hume and normativity raised by Wittgenstein could just be swept under the philosophical rug. With Bryant I focussed on the problem of dogmatic ontologism, the notion that naming correlation as a problem somehow warranted a return to the good old preKantian days, where we could make ontological assertions without worrying about our capacity to make such claims. With Hagglund I raised issues with his interpretation of Derrida as an early thinker of ‘ultratranscendental materialism,’ showing how the concepts at issue were intentional through and through, and thus thoroughly incompatible with the natural scientific project. With Zizek I focussed on the way his deflationary ontology of negative subjectivity arising from some ‘gap’ in the real, aside from simply begging all the questions it purported to answer, amounted to an ontologization of what is far more parsimoniously explained as a cognitive illusion.

And, of course, I took the opportunity to demonstrate the explanatory power of the Blind Brain Theory in each case, the way each of these approaches actually exploit various metacognitive illusions to make their case.

Now, having recently completed Johnston’s Prolegomena to Any Future Materialism: The Outcome of Contemporary French Philosophy, I’ve come to realize that these thinkers* are afflicted with the same set of recurring problems, problems which must be overcome if anything approaching a compelling account of the kind Johnston sets as his goal is to be had. These might be enumerated as follows:

Naivete Problem: With the qualified exception of Zizek, these authors seem largely (and in some cases entirely) ignorant of the enormous philosophical literature dealing with the problems intentionality poses for materialism/physicalism. They also seem to have scant knowledge of the very sciences they claim to be ‘grounding.’

No Cognitive Guarantee Problem: These authors take it as given that radical self-deception is simply not a possible outcome of a mature neuroscience–that something resembling subjectivity as remembered is ‘axiomatic.’ In all fairness, this is a common presumption of those critical of the eliminativist implications of the sciences of the brain. Rose and Abi-Rached, for instance, make it the centrepiece of their attempt to defang the neuroscientific threat to social science in their Neuro: The New Brain Sciences and the Management of the Mind. (Their strategy is twofold: on the one hand, they (like some of the authors considered here) give a conveniently narrow characterization of the threat in terms of subjectivity, arguing that the findings of neuroscience in this regard are simply confirming the subject-decentering theoretical insights already motivating much of the social sciences. Then they essentially cherry-pick researchers and commentators in the field who confirm their thesis without giving dissenters a hearing.) The unsettling truth is that wholesale, radical deception regarding who and what we are is entirely possible (evolution only cares about accuracy insofar as it pays reproductive dividends), and actually already a matter of empirical fact regarding a handful of cognitive capacities.

Talk Is Cheap Problem: There is a decided tendency among these authors to presume the effectiveness of metaphysical argumentation, to not only think that ontological claims merit serious attention in the sciences, but that the threat posed is merely ideological and not material. Rehearsing old arguments against determinism (especially when it’s the Second Law of Thermodynamics that needs to be refuted) will make no difference whatsoever once the brain ceases to be a ‘grey box’ and becomes continuous with our technology.

Implausible Continuity Problem: All of these authors ignore what I call the Big Fat Pessimistic Induction: the fact that, all things being equal, we should expect science to revolutionize the human as radically as it has revolutionized every other natural domain now that the brain has become empirically tractable. They assume, rather, that the immunity the opacity of the brain had granted their tradition historically will somehow continue.

Metacognitive Reliability Problem: All of these authors overlook the potentially crippling issue of metacognitive deception, despite the mounting evidence of metacognitive unreliability. I should note that this tendency is common in Analytic Philosophy of Mind as well (but less and less so as the years pass).

Intentional Dissociation Problem: All of these authors characterize the cognitive scientific threat in the narrow terms of subjectivity rather than intentionality broadly construed, the far more encompassing rubric common to Analytic philosophy. Given the long Continental tradition of critiquing commonly held conceptions of subjectivity, the attractiveness of this approach is understandable, but no less myopic.

I think Prolegomena to Any Future Materialism: The Outcome of Contemporary French Philosophy suffers from all these problems—clearly so. What follows is not so much a review—I’ll await the final book of his trilogy for that (for a far more balanced consideration see Stephan Craig Hickman’s serial review here, here, here, and here)—as a commentary on the general approach one finds in many Continental materialisms as exemplified by Johnston. What all these authors want is some way of securing—or salvaging—some portion of the bounty of spirit absent spirit. They want intentionality absent theological fantasy, and materialism absent nihilistic horror. What I propose is a discussion of the difficulties any such project must overcome—a kind of prolegomena to Johnston’s Prolegomena—and a demonstration why he cannot hope to succeed short of embracing the very magical thinking he is so quick to deride.

Insofar as this is a blog post, part of a living, real time debate, I heartily encourage partisans of his approach to sound off. I am by no means a scholar of any of these authors, so I welcome corrections of misinterpretations. Strawmen teach few lessons, and learn none whatsoever. But I also admit to a certain curiosity given the optimistic stridency of so much of Johnston’s rhetoric. “From my perspective,” he writes in a recent interview, “these naturalists are overconfident aggressors not nearly as well-armed as they believe themselves to be. And, the anti-naturalists react to them with unwarranted fear, buying into the delusions of their foes that these enemies really do wield scientifically-solid, subject-slaying weapons.” I’m sure everyone reading this would love to see what kind of walk accompanies this talk! From my quite contrary perspective, the only way a book like this could be written is for the lack of any sustained interaction with those holding contrary views. Write for your friends long enough, and your writing becomes friendly.

In my own terms, Johnston is an explicit proponent of what might be called noocentrism, the last bastion, now that geocentrism and biocentrism have been debunked, of the intuition that we are something special. Freud, of course, famously claimed to have accomplished this overthrow, to have inflicted the third great ‘narcissistic wound,’ when he had only camouflaged the breastworks by carving intentionality along different mortices. Noocentrism represents an umbrella commitment to our metacognitive intuitions regarding the various efficacies of experience, and these are the intuitions that Johnston explicitly seeks to vindicate. He is ‘preoccupied,’ as he puts it, “with constructing an ontology of freedom” (204). Since any such ontology contradicts the prevailing understanding of the natural arising out of the sciences–how can freedom arise in a nature where everything is in-between, a cog for indifferent forces?–the challenge confronting any materialism is one of explaining subjectivity in a materially consistent manner. As he puts it in his recent Society and Space interview:

“For me, the true ultimate test of any and every materialism is whether it can account in a strictly materialist (yet non-reductive) fashion for those phenomena seemingly most resistant to such an account. Merely dismissing these phenomena (first and foremost, those associated with subjectivity) as epiphenomenal relative to a sole ontological foundation (whether as Substance, Being, Otherness, Flesh, Structure, System, Virtuality, Difference, or whatever else) fails this test and creates many more problems than it supposedly solves.”

Naturalizing consciousness and intentionality—or in Johnston’s somewhat antiquated jargon, explaining the material basis of subjectivity—is without a doubt the holy grail, not only of contemporary philosophy of mind, but of several sciences as well. And he is quite right to insist, I think, that any such naturalization that simply eliminates intentional phenomena (along the lines of Alex Rosenberg’s position, say) hasn’t actually naturalized anything at all. If consciousness and intentionality don’t exist as we intuit them, then we need some account of why we intuit them as such. Elimination, in other words, has to explain why elimination is required in the first place.

But global eliminativist materialist approaches (such as Rosenberg’s and my own) are actually very rare. In contemporary debates, philosophers and researchers tend to be eliminativists or antirealists about specific intentional phenomena, qualia, content, norms, or so on, rather than all intentional phenomena. This underscores two problems that loom large over Johnston’s account, at least as it stands in this first volume. The first has to do with what I called the Intentional Dissociation Problem above, the fact that the problem of subjectivity is simply a subset of the larger problem of intentionality. It falls far short of capturing the ‘problem space’ that Johnston purports to tackle. Some philosophers (Pete Mandik comes to mind) are eliminativists about subjectivity, yet realists about other semantic phenomena.

The second has to do with the fact that throughout the course of the book he repeatedly references reductive and eliminative materialisms as his primary rhetorical foil without actually engaging any of the positions in any meaningful way. Instead he references Catherine Malabou’s perplexing work on neuroplasticity, stating that “one need not fear that bringing biology into the picture of a materialist theory of the subject leads inexorably to a reductive materialism of a mechanistic and/or eliminative sort; such worries are utterly unwarranted, based exclusively on an unpardonable ignorance of several decades of paradigm-shifting discoveries in the life sciences” (Prolegomena, 29). Why? Apparently because epigenetics and neural plasticity “ensure the openness of vectors and logics not anticipated or dictated by the bump-and-grind efficient causality of physical particles alone” (29).

Comments like these—and one finds them scattered throughout the text—demonstrates a problematic naivete regarding his subject matter. One could point out that quantum indeterminacy actually governs the ‘determinism’ he attributes to physical particles. But the bigger problem—the truly ‘unpardonable ignorance’—is that it shows how little he seems to understand the very problem he has set out to solve. His mindset seems to be as antiquated as the sources he cites. He seems to think, for instance, that ‘mechanism’ in the brain sciences refers to something nonstochastic, ‘clockwork,’ that the spectre of Laplace is what drives the unwarranted claims of reductive/eliminative materialists. ‘Decades of research revealing indeterminacy, and still they speak of mechanisms?’

As hard as it is to believe, Johnston pretty clearly thinks the primary problem materialism poses for subjectivity is the problem of determinism. But the problem, simply put, is nothing other than the Second Law of Thermodynamics, the exceptionless irreflexivity of the natural. Ontological freedom is every bit as incompatible with the probabilistic as it is the determined. The freedom of noise is no freedom at all.

This, without a doubt, is his single biggest argumentative oversight, the one that probably explains his wholesale dismissal of any would-be detractor such as myself. His foe here is entropy, not some anachronistic conception of clockwork determinism. Only an appreciation of this allows an appreciation of the difficulty the task Johnston has set himself. Forget the thousands of years of tradition, the lifetime of familiarity, the system of concepts anchored, forget that Johnston is arguing for the most beloved thing—your exceptionality—set aside all this, and what remains, make no mistake, is a perpetual motion machine, something belonging to reality but obeying laws of its own.

So how does one theoretically rationalize a perpetual motion machine?

The metaphor is preposterous, of course, even though it remains analogous in the most important respect. Johnston literally believes it’s possible to “be a partisan of a really and indissolubly free subject while simultaneously and without incoherence or self-contradiction remaining entirely faithful to the uncompromising atheism and immanentism of the combative materialist tradition” (176). He thinks that certain real, physical systems (you and me, as luck would have it) do not obey physical law, at least not the way every single system effectively explained through the history of natural science obeys physical law.

What makes the metaphor preposterous, however, is the apparent immediacy of subjectivity, the way it strikes us as a source of some kind upon reflection, hemmed not by astronomical neural complexities, but by rules, goals, rationality. In a basic sense, what could be more obvious? This is what we experience!

Or… is it just what we remember?

And here’s the rub. The problem that Johnston has set himself to solve is a dastardly one indeed, far, far more difficult than he seems to imagine. Even with the dazzling assurance of experience, a perpetual motion machine is pretty damn hard thing to explain. The fact that most everyone is dazzled by subjectivity in its myriad guises doesn’t change the fact that they are, quite explicitly, betting on a perpetual motion machine. There’s a reason, after all, why everyone but everyone who’s attempted what Johnston has set out to achieve has failed. “Empty-handed adversaries,” as Johnston claims in the same interview, “do not deserve to be feared.” But if they’re empty-handed, then they must know kung-fu, or something lethal, because so far they’ve managed to kill every single theory such as his!

But when you start interrogating that ‘dazzling assurance,’ when you consider just how much we remember, things become even more difficult for Johnston. Because the fact is, we really don’t remember all that much. Certain things escape memory simply because they escape experience altogether. Our brains, for instance, have no more access to the causal complexities of their own function than they do to those of others, so they rely on powerful, yet imperfect systems, ‘fast and frugal heuristics,’ to solve (explain, predict, and manipulate) themselves and others. When abnormalities occur in these systems, such as those belonging, say, to autism spectrum disorder, our capacity to solve is impaired.

As the history of philosophy attests, we seem to experience next to nothing regarding the actual function of these systems, or at least nothing we can remember in the course of pondering our various forms of intentional problem solving. All we seem to intuit are a series of problem-solving modes that we simply cannot square with the problem-solving modes we use to engineer and understand mechanical systems. And, most importantly, we seem to experience (or remember) nothing of just how little we experience (or remember). And so the armchair perpetually remains a live option.

I say ‘most importantly’ because this means remembering doesn’t simply overlook its incapacities, it neglects them. When it comes to experience, we remember everything there is to be remembered, always. We rarely have any inkling of what’s bent, bleached, or lost. What is lost to the system, does not exist for the system, even as something lost.

Add neglect and suddenly a good number of intentional peculiarities begin to make frightening sense. Why, for instance, should we be surprised that problem solving modes adapted to solve complex causal systems absent causal information cannot themselves make sense of causal information? We are mechanically embedded in our environments in such a way that we cannot cognize ourselves as so embedded, and so are forced to cognize ourselves otherwise, acausally, relying on heuristics that theoretical reflection transforms into rules, goals, and reasons, hazy obscurities at the limits of discrimination.

We are astronomically complicated causal systems that cannot remember themselves as such, amnesiac machines that take themselves for perpetual motion machines for the profundity of their forgetting. At any given moment, what we remember is all there is; there is nothing else to blame, no neuromechanistic background we might use to place our thoughts and experiences in their actual functional context, namely, the machinery that bullets and spirochetes and beta-amyloid plaques can destroy. We do not simply lack the access and the resources to intuit ourselves for what we are (something), we lack the resources intuit this lack of resources. Thus the myth of perpetual motion, our conviction in what Johnston calls the “self-determining spontaneity of transcendental subjects.”

The limits of remembering, in other words, provide an elegant, entirely naturalistic, explanation for our metacognitive intuitions of spontaneity, the almost inescapable sense that thought has to represent some kind of fundamental discontinuity in being. Since we cannot cognize the actual activity of cognition, that activity—the function of flesh and blood neural circuits that would seize were you to suffer a midcerebral arterial stroke this instant—does not exist for metacognition. All the informational dimensions of this medial functionality, the dimensions of the material, vanish into oblivion, stranding us with a now that always seems to be the same now, despite its manifest difference, a life that is always in the mysterious process of just beginning.

But Johnston doesn’t buy this story. For him, we actually do remember everything we need to remember to theoretically fathom experience. For him, the fact of subjectivity is nothing less than an “axiomatic intuition” (204), as dazzling as dazzling can be. He never explains how this magic might be possible, how any brain could possibly possess the access and resources to fathom its structure and dynamics in anything but radically privative ways, but then he’s not even aware this is a problem (or more likely, he assumes Freud and Lacan have already solved this problem for him). For him, self-determining spontaneity—perpetual motion—is simply a positive fact of what we are. Everything is remembered that needs to be remembered.

The problem, he’s convinced, doesn’t lie with us. So in order to pass his own test, to craft a materialism absent cryptotheological elements that nevertheless explains (as opposed to explains away) all the perplexing phenomena of intentionality, he needs some different account of nature.

He’s not alone in this regard. The vast majority of theorists who tackle the many angles of this problem are intentional realists of some description. But for many, if not most of them, the tactic is to posit empirical ignorance: though we presently cannot puzzle through the conundrums of intentional phenomena, proponents of so-called ‘spooky emergence’ contend, advances in cognitive neuroscience (and/or physics) will somehow vindicate our remembering. Consciousness and intentionality, they believe, are emergent phenomena, novel physical properties pertaining to as yet unknown natural mechanisms.

Johnston also appropriates the term ‘emergentism’ to describe his project, but it’s hard to see it as much more than a ‘cool by association’ ploy. Emergentism provides a way for physicalists (materialists) to redeem something ‘perpetual enough’ short of committing to ontological pluralism. Emergentists, in other words, are naturalists, convinced that “philosophy can and should limit itself to a deontologized epistemology with nothing more than, at best, a complex conception of the cognizing mental apparatus” (204).

This ‘article of faith,’ however, is one that Johnston explicitly rejects, claiming that “thought cannot indefinitely defer fulfilling its duty to build a realist and materialist ontology” (204). So be warned, no matter how much he helps himself to the term, Johnston is no ‘emergentist’ in the standard sense. He’s an avowed ontologist, as he has to be, given the Zizekian frame he uses to mount his theoretical chassis. “[A] theory of the autonomous negativity of self-relating subjectivity always is accompanied, at a minimum implicitly, by the shadow of a picture of being (as the ground of such subjectivity) that must be made explicit sooner or later” (204). Elsewhere, he writes, “I am tempted to characterize my transcendental materialism as an emergent dual-aspect monism, albeit with the significant qualification that these ‘aspects’ and their eradicable divisions (such as mind and matter, the asubjective and subjectivity, and the natural and the more-than-natural) enjoy the heft of actual existence” (180), that is, he’s a kind of dual-aspect monist so long as the dualities are not aspectual!

Insofar as perpetual motion machines (like autonomous subjects) pretty clearly violate nature as science presently conceives it, one might say that Johnston’s ontological emergentism is honest in a manner that naturalistic emergentism is not. As an eliminative naturalist who finds the notion of systems that violate the laws of physics arising as a consequence of those laws ‘spooky,’ I’m inclined to think so. But in avoiding one credibility conundrum he has simply inherited another, namely, our manifest inability to arbitrate ontological claim-making.

Johnston himself recognizes this problem of ontological credibility, insofar as he makes it the basis of his critiques of Badiou and Meillassoux, who suffer, he argues, “from a Heideggerean hangover, specifically, an acceptance unacceptable for (dialectical) materialism of the veracity of ontological difference, or a clear-cut distinction between the ontological and the ontic” (170). ‘Genuine materialism,’ as he continues, “does not grant anyone the low-effort luxury of fleeing into the uncluttered, fact-free ether of ‘fundamental ontology’ serenely separated from the historically shifting stakes of ontic disciplines” (171). And how could it, now that the machinery of human cognition itself lies on the examination table? He continues, “Although a materialist philosophy cannot be literally falsifiable as are Popperian sciences, it should be contestable as receptive, responsive, and responsible vis-a-vis the sciences” (171).

This, for me, is the penultimate line of the book, the thread from which the credibility of Johnston’s whole project hangs. As Johnston poses the dilemma:

“… the quarrels among the prior rationalist philosophers about being an sich are no more worth taking philosophically seriously than silly squabbles between sci-fi writers about whose concocted fantasy-world is truer or somehow more ‘superior’ than the others; such quarrels are nothing more than fruitless comparisons between equally hallucinatory apples and oranges, again resembling the sad spectacle of a bunch of pulp fiction novelists bickering over the correctness-without-criteria of each others’ fabricated imaginings and illusions.” 170

And yet nowhere could I find any explanation of how his own ontology manages to avoid this ‘fantasy world trap,’ to be ‘receptive’ or ‘responsive’ or ‘responsible’ to any of the sciences—to be anything other than another fundamental ontology, albeit one that rhetorically approves of the natural scientific project. The painful, perhaps even hilarious fact of the matter is that Johnston’s picture of intentionally rising from the cracks and gaps of an intrinsically contradictory reality happens to be the very ontological trope I use to structure the fantasy world of The Second Apocalypse!

There can be little doubt that he believes his picture somehow is receptive, responsive, and responsible, thinking, as he does, that his account

“… will not amount merely to compelling philosophy and psychoanalysis, in a lopsided, one-way movement, to adapt and conform to the current state of the empirical, experimental sciences, with the latter and their images of nature left unchanged in the bargain. Merging philosophy and psychoanalysis with the sciences promises to force profound changes, in a two-way movement, within the latter at least as much as within the former.” 179

Given the way science has ideologically and materially overrun every single domain it has managed to colonize historically, this amounts to a promise to force a conditional surrender with words—unless, that is, he has some gobsmacking way to empirically motivate (as opposed to verify) his peculiar brand of ontological emergentism.

But the closest he comes to genuinely explaining the difference between his ‘good’ ontologism and the ‘bad’ ontologism of those he critiques comes near the end of the text, where he espouses what might be called a qualified Darwinianism, one where “the chasm dividing unnatural humanity from natural animality is … not a top-down imposition inexplicably descending from the enigmatic heights of an always-already there ‘Holy Spirit’ … but, instead a ‘gap’ signalling a transcendence-in-immanence” (178). To advert to Dennettian terms, one might suggest that Johnston sees the bad ontologism of Badiou and Meillasoux as offering ‘skyhooks,’ unexplained explainers set entirely outside the blind irreflexivity of nature. His own good ontologism, on the other hand, he conceives phylogenetically, which is to say more in terms of what Dennett would call ‘cranes,’ a complicating continuity of natural processes and mechanisms culminating in ‘virtual machines’ that we then mistake for skyhooks.

Or perhaps we should label them ‘crane-hooks,’ insofar as Johnston envisions a ‘gap’ or ‘contradiction’ written into the very fundamental structure of existence, a wedge that bootstraps subjectivity as remembered

A perpetual motion machine.

The charitable assumption to make at this point is that he’s saving this bombshell for the ensuing text. But given the egregious way he mischaracterizes the difficulties of his project at the beginning of the text, it’s hard to believe he has much in the way combustible material. As we saw, he flat out equivocates the concrete mechanistic threat—the way the complexities of technology are transforming the complexities of life into more technology—with the abstract philosophical problem of determinism. Creeping depersonalization–be it the medicalization of individuals in numerous institutional (especially educational) contexts, or the ‘nudge’ tactics ubiquitously employed throughout commercial society, or institutional reorganization based on data mining techniques–is nothing if not an obvious social phenomenon. When does it stop? Is there really some essential ‘gap’ between you and all the buzzing, rumbling systems about you, the negentropic machinery of life, the endless lotteries that comprise evolution, the countless matter conversion engines that are stars? Does mechanism, engineered or described, eventually bump into the edge of mere nature, bounce from some redemptive contradiction in the fabric of being? One that just happens to be us?

Are we the perpetual motion machine we’ve sought in vain for millennia?

The fact is, one doesn’t have to look far to conclude that Johnston’s ontologism is just more bad ontology, the same old empty cans strung in a different configuration. After all, he takes the dialectical nature of his materialism quite seriously. As he writes:

“… naturalizing human being (i.e., not allowing humans to stand above-and-beyond the natural world in some immaterial, metaphysical zone) correlatively entails envisioning nature as, at least in certain instances, being divided against itself. An unreserved naturalization of humanity must result in a defamiliarization and reworking of those most foundational and rudimentary proto-philosophical images contributing to any picture of material nature. The new, fully secularized materialism (inspired in part by Freudian-Lacanian psychoanalysis) to be developed and defended in Prolegomena to Any Future Materialism is directly linked to this notion of nature as the self-shattering, internally conflicted existence of a detotalized material immanence.” 19-20

What all this means is that nature, for Johnston, is intrinsically contradictory. Now contradictions are at least three things: first, they logically entail everything; second, they’re analytically difficult to think; and third, they’re conceptually semantic, which is to say, intentional through and through. Setting aside the way the first two considerations raise the spectres of obscurantism and sophistry (where better hide something stolen?), the third should set the klaxons wailing for even those possessing paraconsistent sympathies. Why? Simply because saying that reality is fundamentally contradictory amounts to saying that reality is fundamentally intentional. And this means that what we have here, in effect, is pretty clearly a kind of anthropomorphism, the primary difference being, jargon aside, that it’s a different kind of anthropos that is being externalized, namely, the fragmented, decentred, and oh-so-dreary ‘postmodern subject.’

I don’t care how inured to a discourse’s foibles you become, this has to be a tremendous problem. Johnston writes, “a materialist theory of the subject, in order to adhere to one of the principal tenets of any truly materialist materialism (i.e., the ontological axiom according to which matter is the sole ground), must be able to explain how subjectivity emerges out of materiality—and, correlative to this, how materiality must be configured in and of itself so that such an emergence is a real possibility” (27). Now empirically speaking, we have no clue ‘how materiality must be configured’ because we do not, as yet, understand the mechanisms underwriting consciousness and intentionality. Johnston, of course, rhetorically dismisses this ongoing, ever advancing empirical project, as an obvious nonstarter. He has determined, rather, that the only way subjectivity can be naturally understood is if we come to see that nature itself is profoundly subjective…

I can almost hear Spinoza groaning from his grave on the Spui.

If the contradiction of the human can only be ‘explained’ by recourse to some contradiction intrinsic the entire universe, then why not simply admit that the contradiction of the human cannot be explained? Just declare yourself a mysterian of some kind–I dunno. Johnston devotes considerable space critiquing Meillasoux for using ‘hyperchaos’ as an empty metaphysical gimmick, a post hoc way to rationalize the nonmechanistic efficacy of intentional phenomena. And yet it’s hard to see how Johnston gives his reader even this much, insofar as he’s simply taken the enigma of intentionality and painted it across the cosmos—literally so!

Johnston references the ‘sad spectacle of a bunch of pulp fiction novelists’ arguing their worlds’ (170), but as someone who’s actually participated in that (actually quite hilarious) spectacle, I can assure everyone that we, unlike the sad spectacle of Continental materialists arguing their worlds, know we’re arguing fictions. What makes such spectacles sad is the presumption to a cognitive authority that simply does not exist. Arguing the intrinsically dialectical nature of materiality is of a par with arguing intelligent design, save that the intuitions motivating intelligent design are more immediate (they require nowhere near as much specialized training to appreciate), and that its proponents have done a tremendous amount of work to make their position appear receptive, responsive, and responsible to the sciences they would, in the spirit of share-and-share alike, ‘complement with a deeper understanding.’

A contradictory materiality is an anthropomorphic materiality. It provides redemption, not understanding of some decentred-me-friendly world that science has been unable to find. In his attempt to materially square the circle of subjectivity, Johnston invents a stripped down, intellectualized fantasy world, and then embarks on a series of ‘fruitless comparisons between equally hallucinatory apples and oranges’ (170). And how could it be any other way when all of these pulp philosophy thinkers are trapped arguing memories?

Vivid ones to be sure, but memories all the same.

The vividness, in fact, is a large part of the whole bloody problem. It means that no matter how empty our metacognitive intuitions regarding experience are, they generally strike us as sufficient: What, for instance, could be more obvious than our normative understanding of rules? But there’s powerful evidence suggesting our feeling of willing is only contingently connected to our actions (a matter of interpretation). There’s irrefutable evidence that our episodic memory is not veridical. Likewise, there is powerful evidence suggesting our explanations of our behaviour are only contingently related to our actions (a matter of interpretation). Even if you dispute the findings (with laboratory results, one would hope), or think that psychoanalysis is somehow vindicated by these findings (rather than rendered empirically irrelevant), the fact remains that none of the old assumptions can be trusted.

Do you have any metacognitive sense of the symphony of subpersonal heuristic systems operating inside your skull this very instant, the kinds of problems they’ve adapted to solve versus the kinds of problems that can only generate impasse and confusion? Of course not. The titanic investment in time and resources required to isolate what little we have isolated wouldn’t have been required otherwise. We are almost entirely blind to what we are and what we do. But because we are blind to that blindness, we confuse what little we do see with everything to be seen. We therefore become the ‘object’ that cannot be an ‘object,’ the thing that cannot be intuitively cognized in time and space, that strikes us with the immediacy of this very moment, that appears to somehow stand outside a nature that is all-encompassing otherwise.

The system outside the picture, somehow belonging and not belonging

Or as I once called it, the ‘occluded frame.’

And this just follows from our mechanical nature. For a myriad of reasons, any system originally adapted to systematically engage environmental systems will be structurally incapable of systematically engaging itself in the same manner. So when it develops the capacity to ask, as we have developed the capacity to ask, ‘What am I?’ it will have grounds to answer, ‘Of this world, and not of this world.’

To say, precisely because it is a mechanism, ‘I am contradiction.’

As with the crude thumbnail given above, the Blind Brain Theory attempts to naturalistically explain away the peculiarities of intentionality and phenomenality in terms of neglect. Since we cannot intuit our profound continuity with our environments, we intuit ourselves otherwise, as profoundly discontinuous with our environments. This discontinuity, of course, is the cornerstone of the problem of understanding what we are. Before, when the brain remained a black box, we could take it for granted, we could leverage our ignorance in ways that catered to our conceits, especially our perennial desire to be the great exception to the natural. So long as the box remained sealed, we could speak of beetles without fear of contradiction.

Now that the box has been cracked open with nary a beetle to be found, all those speculative discourses reliant upon our historical ignorance find themselves scrambling. They know the pattern, even if they are loath to speak of it or, like Johnston, prone to denial. Nevertheless, science is nothing if not imperial and industrial. It displaces aboriginal discourses, delegitimizes them in the course of revolutionizing any given domain. Humans, meanwhile, are hardwired to rationalize their interests. When their claims to status and authority are threatened, the moral and intellectual deficiencies of their adversary simply seems obvious. So it should come as no surprise that specialists in those discourses are finally rousing themselves from their ingroup slumber to defend what they must consider manifest authority and hard-earned privileges.

But they face a profound dilemma when it comes to prosecuting their case against science—a dilemma not one of these Continentalists has yet to acknowledge. Before, in the good old black box days, they could rely on simple pejoratives like ‘positivism’ and ‘scientism’ to do all the heavy lifting, simply because science reliably fell silent when it came to issues within their domain. The bind they find themselves in now, however, could scarce be more devious. The most obvious problem lies in the revolutionary revision of their subject matter—the thinking human. But the subject matter of the human is also the subject of the matter, the activity that makes the understanding of any subject matter possible. Continentalists, of course, know this, because it provides the basis for their ontological priority claims. They are describing, so they think, what makes science possible. This is what grants them diplomatic transcendental immunity when they take up residence in scientific domains. But Johnston isolates the dilemma—his dilemma—himself when he points out the empty nature of the Ontological Difference.

Foucault actually provides the most striking image of this that I know of with his analysis of the ‘emprico-transcendental doublet called man’ in The Order of Things. What is transpiring today can be seen as a battle for the soul of the darkness that comes before thought. Is it ontological as so much of philosophy insists? Or is it ontic as science seems to be in the process of discovering? So long as our ontic conditions remained informatically impoverished, so long as the brain remained a black box, then the dazzling vividness of our remembering could easily overcome our abstract, mechanistic qualms. We could rely on the apparent semantic density of ‘lived life’ or ‘conditions of possibility’ or ‘language games’ or ‘epistemes’ or so on (and so on) to silence the rumble of an omnivorous science. We could dwell in the false peace of trench warfare, a stalemate between two general, apparently antithetical claims to one truth. As Foucault writes:

“… either this true discourse finds its foundation and model in the empirical truth whose genesis in nature and in history it retraces, so that one has an analysis of the positivist type (the truth of the object determines the truth of the discourse that describes its foundation); or the true discourse anticipates the truth whose nature and history it defines; it sketches it out in advance and foments it from a distance, so that one has a discourse of the eschatological type (the truth of the philosophical discourse constitutes the truth in formation).” 320

Foucault, of course, has stacked the deck in this characterization of epistemological modes—simply posing the (historically contingent) problem of the human in terms of an ‘empirico-transcendental doublet’ is to concede authority to the transcendental—but he was nevertheless astute–or at least evocative–in his assessment of the form of the problem (as seen from within the subject/object heuristic). Again, as he writes:

“The true contestation of positivism and eschatology does not lie, therefore, in a return to actual experience (which rather, in fact, provides them with confirmation by giving them roots); but if such a contestation could be made, it would be from the starting-point of a question which may well seem aberrant, so opposed is it to what has rendered the whole of our thought historically possible. This question would be: Does man really exist?” 322

A question that was both prescient in his day and premature, given that the empirical remained, for most purposes, locked out of the black box of the human. For all his historicism, Foucault failed to look at this dilemma historically, to realize (as Adorno arguably did) that short of some form reason capable of contesting scientific claims on the human, the domain of the human was doomed to be overrun by scientific reason, and that discourses such as his would eventually be reduced to the status of alchemy or astrology or religion.

And herein lies the rub for Johnston. He thinks the key to a viable Continental materialism turns on getting the ontological nature of the what right, when the problem resides in the how. He says as much himself: anybody can cook up and argue a fantasy world. In my own lectures on fantasy, the most fictional of fictions, I always stress how the anthropomorphic ‘secondary worlds’ depicted could only be counted as ‘fantastic’ given the cognitive dominion of science. This, I think, is the real anxiety lurking beneath his work (despite all his embarrassing claims about ‘empty handed foes’). The only thing preventing the obvious identification of his secondary worlds as fantastic was the scientific inscrutability of the human. Now that the human is becoming empirically scrutable across myriad dimensions, now that the informatic floodgates have been cranked open—now that his claims have a baseline of comparison—the inexorable processes that rendered the anthropomorphic fantastic across external nature are beginning to render internal meaning fantastic as well.

Why do pharmaceuticals impact us? Man is a machine. Why do cochlear implants function? Man is a machine. Why do head injuries so profoundly reorganize experience? Man is a machine. The Problem of Mechanism is material first and only secondarily philosophical. Given what I know about the human capacity for self-deception (having followed the science for years now), I have no doubt that the vast majority of people will find refuge in ‘mere words,’ philosophical or theological rationalization of this or that redeeming ‘axiomatic posit.’ This is what makes the Singularity so bloody crucial to these kinds of debates (and what puts thinkers like David Roden so tragically far ahead of his peers). When we become indistinguishable from our machinery, or when our machines make kindergarten scribbles of our greatest works of genius, will we persist insisting on our ontological exceptionality then?

Or will the ‘human’ merely refer to some eyeless, larval stage? Will noocentrism be seen as last of the three great Centripetal Conceits?

Short of discovering some Messianic form of reason—a form of cognition capable of overpowering a scientific cognition that can cure blindness and vaporize cities—attempts to argue Messianic realities a la Continental materialism are doomed to fail before they even begin. Both the how and the what of the traditional humanities are under siege. As it stands, the profundity of this attack can still be partially hidden, so long as one’s audience wants to be reassured and has no real grasp of the process. A good number of high profile researchers are themselves apologists for the humanistic status quo, so one can, as defenders of various religious beliefs are accustomed, pluck many heartening quotes from the enemy’s own mouth. But since it is the rising tide of black-box information that has generated this legitimacy crisis, it seems more than a little plausible to presume that it will deepen and deepen, until finally it yawns abyssal, no matter how many well-heeled words are mustered to do battle against it.

Not matter how many Johnston’s pawn their cryptotheological perpetual motion machines.

Our only way to cognize our experiencing is via our remembering. The thinner this remembering turns out to be—and it seems to be very thin—the more we should expect to be dismayed and confounded by the sciences of the brain. At the same time we should expect a burgeoning market for apologia, for rationalizations that allow for the dismissal and domestication of the threats posed. Careers will be made, celebrated ones, for those able to concoct the most appealing and slippery brands of theoretical snake-oil. And meanwhile the science will trundle on, the incompatible findings will accumulate, and those of us too suspicious to believe in happy endings will be reduced to arguing against our hopes, and for the honest appraisal of the horror that confronts us all.

Because the bandage of our traditional self-conception will be torn away quicker than you think.


* POSTSCRIPT (17/01/2014): Levi Bryant, it should be noted, is an exception in several respects, and it was remiss of me to include him without qualification. A concise overview of his position can be found here.

Ancient and Modern Enlightenment: from Noosphere to Technosphere (by Ben Cain)

by rsbakker

Enlightenment is elite cognition, the seeing past collective error and illusion to a hidden reality. But the ancient idea of enlightenment differs greatly from the modern one and there may be a further shift in the postmodern era. I’ll try to shed some light on enlightenment, by pursuing these comparisons.


Ancient Enlightenment: Monism and Personification

Enlightenment in the ancient world was made possible by a falling away from our mythopoeic, nomadic prehistory. In that Paleolithic period, symbolized by the wild Enkidu in the Epic of Gilgamesh and by the biblical Adam in Eden, there was no enlightenment since everything was thoroughly personified and so nothing could have been perceived as unfamiliar or alien to the masses. The world was experienced as a noosphere, filled with mentality. Only after the rise of sedentary civilization in the Neolithic Era, when farming replaced nomadic hunting in 10,000 BCE, which allowed for much larger populations, was there a loss of that enchanted mode of experience which actually depended on a sort of blissful collective ignorance. As a population increases, the so-called Law of Oligarchy takes hold, which means that social power must be concentrated to avoid civilizational collapse. Dominance hierarchies are established and those in the lower classes become envious of the stronger and more privileged members who are sure to display their greater wealth and access to women with symbols of their higher status. By doing so, each social class learns its boundaries so that the social structure won’t be overridden, which would invite anarchy.

As Rousseau argued, civilization was the precondition of what we might call the sin of egoism. Contrary to Rousseau, prehistoric life wasn’t utopian; at least, objectively, human life in the Paleolithic Era was likely quite savage. But the ancients seemed to have an easier time perceiving the world in magical terms, judging from the evidence of their religions and extrapolating from what we know of children’s experience, given their similar dearth of content to occupy their collective memory. Thus, even as they killed each other over trifles, the prehistoric people would have interpreted such horror as profoundly meaningful. In any case, I think Rousseau is right that civilization made possible a falling away from a kind of intrinsic innocence. Specifically, the increased social specialization led to an epistemic inequality. As food was stored and more and more people lived together, there was greater need for practical knowledge in such areas as architecture, medicine, sanitation, and warfare. The elites became decadent and alienated from nature, since they found themselves free to indulge their appetites with artificial diversions, as specialists took care of the necessities of survival such as the harvesting of food or the defense of the borders. These elites codified the myths that expressed the population’s mores, but while the uneducated majority clung to their naïve, anthropocentric traditions, the cynical and self-absorbed elites more likely regarded the folk tales as superstitions.

Here, then, was the origin of enlightenment as the opposite of wholesale ignorance—and this was a normative dichotomy. Enlightenment was good and its opposite, mental darkness, was bad. Whereas prior to civilization everyone was enlightened, in a sense, or at least everyone deferred to the shaman’s interpretation of how the spiritual and material worlds are intermixed, civilized people came to believe there’s a secret perspective which alone imparts the ultimate truth, leaving the majority in relative ignorance. As for the content of the enlightened worldview in the ancient world, this was informed by both the egoism and the cynicism that distinguished the hierarchical civilization from the prehistoric past. The content thus had two elements: monism and personification. On the one hand, reality was thought to be a unity, whereas the world appeared to be a multiplicity. Enlightenment was the ability to see past the illusion of change, to the underlying timeless interconnection between all events. Again, in the mythopoeic world, there was no distinction between reality and appearance, because mental projections were given equal weight with the material unfolding of events. The world was a magical place. But the enlightened person had to recover a distorted memory of that childlike, mythopoeic vision, as it were, by theorizing a unity beyond the disenchanted multiplicity that confronted the civilized ancients.

On the other hand, ultimate reality was generally personified. So the absolute unity was called God, equated with the self, and often compared to the particular human who actually ruled the land. That is, the civilizational structure was projected onto the spirit world and the gods were used as symbols to reassure the ancients that their social order was just. There was such personification even in Buddhism, specifically in the Mahayana variety, according to which Bodhisattvas are worshipped and Buddha nature is thought to take not just the inconceivable and thus impersonal form, but ghostly or celestial as well as physical ones.

Ancient enlightenment thus had to reconcile the urge to personify, which was a remnant of the mythopoeic experience that was exacerbated by the advent of egoism even among the masses, and which the elites came to use for political purposes, with the world’s alien, indifferent oneness. That theoretical oneness expressed especially the elites’ growing alienation from nature and their nostalgia for the presumed innocence of the earlier, nomadic period. Monism made egoism out to be preconditioned by ignorance, since if the world were really an ultimate unity, the apparent self’s independence would be an illusion. But because egoism had numerous social and economic causes, the enlightened worldview retained some anthropomorphic projections onto the unity, to rationalize the nature of the civilized individual. There were degrees of enlightenment, so that one or the other factor, impersonal metaphysical unity or personification, predominated. For example, in the Eastern religions, the anthropomorphisms were stripped away as the enlightened person was thought to experience a transcendent unity, in a purified state of consciousness. Alternatively, the monotheistic Western traditions generally took a personal deity to be the highest principle.


Modern Enlightenment: Objectivity and Artificialization

The next epochal change was the birth of modern civilization in the European Renaissance and Scientific Revolution, followed by the Enlightenment and the Industrial Revolution. This transition was marked by profound advances in investigative techniques, which presented the educated upper classes with an altogether impersonal world. Instead of being horrified by this new knowledge, modernists relished the opportunity to conquer a material world that has no prior rights or else they sought refuge in the halfway house of deism. In any case, modernists were forced to reconceptualize the idea of enlightenment. Whereas the ancient kind posited a metaphysical unity that was somehow both transcendent and personal, modernists eventually eliminated personhood altogether, not just in metaphysics but in psychology. And so modern enlightenment is an appreciation of the implications of thoroughgoing metaphysical naturalism. The real world is still a hidden unity and scientists seek to uncover the causal pattern that establishes that unity. Thus, the dichotomy between the reality of the hidden spirit world and the illusion of mundane plurality in the spatiotemporal field of opposites became the split between a rational understanding of nature’s impersonality, as confirmed by the impartiality of cause and effect, and the naïve personification of anything, including ultimate reality or the human self. Enlightened modernists are materialists who think that mind is an illusion and that fundamental reality is bound to be alien to our sensibilities.

However, the conception of enlightenment as a matter of rationality, set off against the darkness of superstition, can’t hold, because rationality is a personal matter which takes for granted the illusion of the personal self. The modern myth of enlightenment as merely the courage to follow the logic and the evidence where they lead can’t be the whole story of the great transition to the modern period. Something else must have happened, not just a rise of rational neutrality, if rationality itself is merely peripheral. Instead of seeing modern enlightenment in terms of the symbol of the Light of Reason, and thus as a mental phenomenon, we should see it as technological: modernists exited the Dark Age through their technological advances which literally made the world brighter in the case of the commercial use of electricity. More broadly, modern enlightenment is the expansion of the “Light” of Artificiality, which makes for a wealth of historical data points. After all, what makes a dark age dark is the lack of lasting evidence of the culture’s identity, due to massive illiteracy and the absence of durable technologies that tell the tale. All of that changed with the printing press and the computer, for example. A Bright Age, then, is bright with cultural information and the light rays should be thought of as being transmitted especially to future historians.

Commercial light bulbs were patented in the late 19th C, although scientists studied electricity as early as 1600 CE. The Age of Enlightenment is primarily an 18th C. period, so the world didn’t literally become much brighter during the modern Enlightenment. However, the paradigmatic rationality of Enlightenment intellectuals, especially that of Isaac Newton, led directly to the Industrial Revolution in the late 18th and early 19th centuries, which included the invention of the light bulb. So we should look at modern enlightenment as beginning with the myth of rationality and giving way to wonder at the undeniable reality of recent technological advance. First came the light of Reason, then scientists realized that personhood and thus reason are illusory. But all along, the modern process was set in motion which replaced the darkness of nature with the light of artificiality (with technological incarnations of culture which endure and testify to our historical identity). Thus, modern enlightenment is only inchoately the dichotomy between neutral (non-personifying) reason and ignorance; the real distinction is between natural, pristine reality, which is dark and monstrous precisely because of its impersonality, and the light we bring to the world by impressing our stamp into it—not subjectively through mere theological interpretation or magical supposition, as in the mythopoeic period, but through the inexorable, objective spread of modern technology.

What’s monumental about modernity isn’t that some white male Europeans learned to think more rigorously, thanks to the scientific methods they invented. Of course, there are such methods, but modern enlightenment shouldn’t be personalized. When you characterize the new kind of enlightenment in that way, you’re left with incoherence since naturalism won’t support naïve personification. Instead, modern enlightenment must be thought of as a great widening of perspective, so that instead of projecting our ego onto indifferent nature, we eliminate our ego through existential encounters with nature’s monstrosity which humiliate us, doing away with our pretensions. Left thusly vacated, the real world is free to flow through us, as it were. In this case, the glory goes not to the great scientists, regardless of how exoteric modern history is told; the scientific methods, for example, must be part of nature’s self-overcoming on our planet, due to a shift from biological processes to artificial ones.

Scientific methods of thought are algorithms which presage the functions of high technology, as in the computer. In other words, before mass technology there was massive regimentation of intellectual life, whereas prior to the Scientific Revolution, social regimentation was confined to the army, to government, farming, and the like, while the business of discovering the nature of reality was still a free-wheeling affair. Ancient philosophy was mostly an artistic kind of speculation, although there are protoscientific aspects of ancient Greek and Indian philosophies. The Presocratics, for example, followed the logic of their hypotheses, however counterintuitive those hypotheses may have been. But what made the Scientific Revolution so special, objectively speaking, was a social transformation. Instead of being ruled mainly by biological norms, such as by the instinct of preserving the genes through sexual reproduction, which were thinly rationalized by the art of myth-making, a new dynamic was introduced: what Jacques Ellul called the necessity of efficiency as a matter of technique.

All species employ techniques, because they’re adapted to their environment, but the Scientific Revolution was the birth of an impersonal, regimented subculture of cognitive elites, one that’s modeled more and more on the machines made possible by that cognitive labour. In place of personification, mystification, or artistic speculation, there’s surrender to rational technique, to algorithms, and to the other scientific methods (public and repeatable testing of hypotheses, mathematical precision, and so on). It’s as though in depersonalizing ourselves, thanks to skepticism, the disempowerment of the Catholic Church, and so forth, we allowed nature’s impersonality to flow more easily through our social structures. Whereas hitherto, our bodies were governed by evolutionary norms and our minds were consumed by myths and illusions of personhood, which we projected onto nature so that we became doubly deluded, modernists abandoned personification, which freed the mind to mimic what the rest of the universe is doing, namely to flow in what I call an undead (impersonal but not inert) fashion.

We still personify techniques when we think of them teleologically, as having a mentally represented goal. However, even if there’s no divine mind desiring nature to end in some way, natural processes do have ends, which is just to say that there are natural processes, as such, or changes that have initial conditions, transitional periods, and probable points of termination. The more we understand nature, the wider our field of vision until we think of everything as a cosmic whole having a beginning (the Big Bang), a middle (evolution and complexification in space and time), and an end state, such as the universe’s heat death. What we call the scientific methods, then, or the more efficient modern techniques of rational thought, are really—according to the enlightened modernist—an inflowing of some underlying natural process besides biological evolution, one which begins with ultra-rational cognition and continues with the elimination of the noosphere and with the transformation of the biosphere into the technosphere.


Counter-Enlightenment and the Return of Mythopoeic Reverie

As long as we’re depersonalizing enlightenment, we should note the Counter-Enlightenment period which leads from the Romantics and other early critics of modern hyper-rationality to postmodern relativism and general jadedness. I won’t attempt to adjudicate this debate here, but I want to close by reflecting on whether the Counter-Enlightenment should be interpreted as an omen indicating that modern enlightenment will itself be transformed. Again, if we ignore the psychological and social levels of inquiry, since an enlightened modernist must regard them as misleading, we can look at historical developments as stages of some larger process. Natural selection explains the design of living bodies, but not the cultural shifts between elite forms of cognition. From mythopoeic animism, to the middle ground of ancient mystical theism, to modern naturalism, there’s a clear elimination of personhood from grand theories. Moreover, there’s exponential progress in technical innovation, as modernists have come to divorce rationality from artistic interpretation. Rather than seeing herself as similar to a shaman, in being a wise person, healer, or hero for venturing into the unknown, an enlightened modernist is more likely to think of herself as a glorified calculator. Modern cognition is hyper-rational in that logic for us is demythologized, and the sciences are separate from the arts and from the humanities, which means that scientific cognition is inhuman (objective and neutral). Science is thus the indwelling of natural mechanisms, due to a breakdown in resistance from religious delusions, resulting in the perfection of the artificial world. Modern geniuses are distorted mirrors held up to undead nature, the reflected image being a technological bastardization of the monstrous original.

And yet we may be witnessing here a cycle rather than a linear progression. Technology may allow us to recover the mythopoeic union of object and subject, so that modern objectivity overcomes itself through its technological progeny. After all, the artificial world caters to our whims and so exacerbates egoism and the urge to personify. Whereas modern enlightenment began with a vision of a lifeless, mechanical universe, the postmodern kind is much less arid and austere. This is because postmodernists are immersed in an artificial world which turns fantasies into realities on a minute-by-minute basis, thus perhaps fulfilling the promise of mythopoeic speculation. For example, if you’re hungry, you may ask your smartphone where the nearest restaurant is and that phone will speak to you; next, you’ll follow the signs in your car which adjusts to your preferences in a hundred ways, and you’ll arrive at the restaurant and be served without having to hunt or cook the animal yourself. The prehistoric fantasy was that nature is alive. Modernists discovered that everything is at best undead and certainly devoid of purpose or of mental, as opposed to biological, life. But perhaps postmodernists are realizing that the world was undead whereas it’s now being imbued with purpose and brought to nonbiological life by us through technology. Instead of mythologizing the world, we postmodernists artificialize it, and whereas natural mechanisms train us to be animals following evolutionary rhythms, artificial mechanisms may train us to be something else entirely, such as infantilized consumers that recapture the prehistoric sense of being at the world’s all-important center, thanks to our history of taming the hostile wilderness.