Three Pound Brain

No bells, just whistling in the dark…

Tag: Heuristic Neglect Theory

Visions of the Semantic Apocalypse: James Andow and Dispositional Metasemantics

by rsbakker

The big problem faced by dispositionalist accounts of meaning lies in their inability to explain the apparent normativity of meaning. Claims that the meaning of X turns on the disposition to utter ‘X’ requires some way to explain the pragmatic dimensions of meaning, the fact that ‘X’ can be both shared and misapplied. Every attempt to pin meaning to natural facts, even ones so low-grained as dispositions, runs aground on the external relationality of the natural, the fact that things in the world just do not stand in relations of rightness or wrongness relative one another. No matter how many natural parameters you pile onto your dispositions, you will still have no way of determining the correctness of any given application of X.

This problem falls into the wheelhouse of heuristic neglect. If we understand that human cognition is fractionate, then the inability of dispositions to solve for correctness pretty clearly indicates a conflict between cognitive subsystems. But if we let metacognitive neglect, our matter of fact blindness to our own cognitive constitution, dupe us into thinking we possess one big happy cognition, this conflict is bound to seem deeply mysterious, a clash of black cows in the night. And as history shows us, mysterious problems beget mysterious answers.

So for normativists, this means that only intentional cognition, those systems adapted to solve problems via articulations of ‘right or wrong’ talk, can hope to solve the theoretical nature of meaning. For dispositionalists, however, this amounts to ceding whole domains of nature hostage to perpetual philosophical disputation. The only alternative, they think, is to collect and shuffle the cards yet again, in the hope that some articulation of natural facts will somehow lay correctness bare. The history of science, after all, is a history of uncovering hidden factors—a priori intuitions be damned. Even still, it remains very hard to understand how to stack external relations into normative relations. Ignorant of the structure of intentional cognition, and the differences between it and natural (mechanical) cognition, the dispositionalist assumes that meaning is real, and that since all real things are ultimately natural, meaning must have a natural locus and function. Both approaches find themselves stalled in different vestibules of the same crash space.

For me, the only way to naturalize meaning is to understand it not as something ‘real out there’ but as a component of intentional cognition, biologically understood. The trick lies in stacking external relations into the mirage of normative relations: laying out the heuristic misapplications generating traditional philosophical crash spaces. The actual functions of linguistic communication turn on the vast differential systems implementing it. We focus on the only things we apparently see. Given the intuition of sufficiency arising out of neglect, we assume these form autonomous systems. And so tools that allow conscious cognition to blindly mediate the function of vast differential systems—histories, both personal and evolutionary—become an ontological nightmare.

In “Zebras, Intransigence & Semantic Apocalypse: Problems for Dispositional Metasemantics,” James Andow considers the dispositionalist attempt to solve for normativity via the notion of ‘complete information.’ The title alone had me hooked (for obvious reasons), but the argument Andow lays out is a wry and fascinating one. Where dispositions to apply terms are neither right nor wrong, dispositions to apply terms given all relevant information seems to enable the discrimination of normative discrepancies between performances. The problem arises when one asks what counts as ‘all relevant information.’ Offloading determinacy onto relevant information simply raises the question of determinacy at the level of relevant information. What constrains ‘relevance’? What about future relevance? Andow chases this inability to delimit complete information to the most extreme case:

It seems pretty likely that there is information out there which would radically restructure the nature of human existence, make us abandon technologies, reconsider our values and place in nature, information that would lead us to restructure the political organization of our species, reconsider national boundaries, and the ‘artificial divisions’ which having distinct languages impose on us. The likely effect of complete information is semantic apocalypse. (Just to be clear—my claim here is not that it is likely we will undergo such a shift. Who is to say what volume of information humankind will become aware of before extinction? Rather, the claim is that the probable result of being exposed to all information which would alter one’s dispositions, i.e., complete information, would involve a radical overhaul in semantic dispositions).

This paragraph is brilliant, especially given the grand way it declares the semantic apocalypse only to parenthetically take it all back! For my money, though, Andow’s throwaway question, “Who is to say what volume of information humankind will become aware of before extinction?” is far and away the most pressing one. But then I see these issues in light of a far different theory of meaning.

What is the information threshold of semantic apocalypse?

Dispositionalism entails the possibility of semantic apocalypse to the degree the tendencies of biological systems are ecologically dependent, and so susceptible to gradual or catastrophic change. This draws out the importance of the semantic apocalypse as distinct from other forms of global catastrophe. A zombie apocalypse, for instance, might also count as a semantic apocalypse, but only if our dispositions to apply terms were radically transformed. It’s possible, in other words, to suffer a zombie apocalypse without suffering a semantic apocalypse. The physical systems underwriting meaning are not the same as the physical systems underwriting modern civilization. So long as some few of us linger, meaning lingers.

Meaning, in other words, can survive radical ecological destruction. (This is one of the reasons we remain, despite all our sophistication, largely blind to the issue of cognitive ecology: so far it’s been with us through thick and thin). The advantage of dispositionalist approaches, Andow thinks, lies in the way it anchors meaning in our nature. One may dispute how ‘meanings’ find themselves articulated in intentional cognition more generally, while agreeing that intentional cognition is biological; a suite of sensitivities attuned to very specific sets of cues, leveraging reliable predictions. One can be agnostic on the ontological status of ‘meaning’ in other words, and still agree that meaning talk turns on intentional cognition, which turns on heuristic capacities whose development we can track through childhood. So long as a catastrophe leaves those cues and their predictive power intact, it will not precipitate a semantic apocalypse.

So the question of the threshold of the semantic apocalypse becomes the question of the stability of a certain biological system of specialized sensitivities and correlations. Whatever collapses this system engenders the semantic apocalypse (which for Andow means the global indeterminacy of meanings, and for me the global unreliability of intentional cognition more generally). The thing to note here, however, is the ease with which such systems do collapse once the correlations between sensitivities and outcomes cease to become reliable. Meaning talk, in other words, is ecological, which is to say it requires its environments be a certain way to discharge ancestral functions.

Suddenly the summary dismissal of the genuine possibility of a semantic apocalypse becomes ill-advised. Ecologies can collapse in a wide variety of ways. The form any such collapse takes turns on the ‘pollutants’ and the systems involved. We have no assurance that human cognitive ecology is robust in all respects. Meaning may be able to survive a zombie apocalypse, but as an ecological artifact, it is bound to be vulnerable somehow.

That vulnerability, on my account, is cognitive technology. We see animals in charcoal across cave walls so easily because our visual systems leap to conclusions on the basis of so little information. The problem is that ‘so little information’ also means so easily reproduced. The world is presently engaged in a mammoth industrial research program bent on hacking every cue-based cognitive reflex we possess. More and more, the systems we evolved to solve our fellow human travellers will be contending with artificial intelligences dedicated to commercial exploitation. ‘Deep information,’ meanwhile, is already swamping the legal system, even further problematizing the folk conceptual (shallow information) staples that ground the system’s self-understanding. Creeping medicalization continues unabated, slowly scaling back warrant for things like character judgment in countless different professional contexts. The list goes on.

The semantic apocalypse isn’t simply possible: it’s happening.

No results found for “cognitive psychology of philosophy”.

by rsbakker

That is, until today.

The one thing I try to continuously remind people is that philosophy is itself a data point, a telling demonstration of what has to be one of the most remarkable facts of our species. We don’t know ourselves for shit. We have been stumped since the beginning. We’ve unlocked the mechanism for aging for Christ’s sake: there’s a chance we might become immortal without having the faintest clue as to what ‘we’ amounts to.

There has to be some natural explanation for that, some story explaining why it belongs to our nature to be theoretically mystified by our nature, to find ourselves unable to even agree on formulations of the explananda. So what is it? Why all the apparent paradoxes?

Why, for instance, the fascination with koans?

Take the famous, “What is the sound of one hand clapping?” Apparently, the point of pondering this lies in realizing the koan is at once the questioning and the questioned, and coming to see oneself as the sound. For many, the pedagogical function of koans lies in revealing one’s Buddha nature, breaking down the folk reasoning habits barring the apprehension of the identity of subject and object.

Strangely enough, the statement I gave you in the previous post could be called a koan, of sorts:

It is true there is no such thing as truth.

But the idea wasn’t so much to break folk reasoning habits as to alert readers to an imperceptible complication belonging to discursive cognition: a complication that breaks the reliability of our folk-reasoning habits. The way deliberative cognition unconsciously toggles between applications and ontologizations of truth talk can generate compelling cognitive illusions—illusions so compelling, in fact, as to hold the whole of humanity in their grip for millennia.

Wittgenstein, and the pragmatists glimpsed the fractionate specialization of cognition, how it operated relative various practical contexts. They understood the problem in terms of concrete application, which for them was pragmatic application, a domain generally navigated via normative cognition. Impressed by the inability of mechanical cognition to double as normative cognition, they decided that only normative cognition could explain cognition, and so tripped into a different version of the ancient trap: that of using intentional cognition to theoretically solve intentional cognition.

Understanding cognition in terms of heuristic neglect lets us frame the problem subpersonally, to look at what’s going on in statements like the above in terms of possible neurobiological systems recruited. The fact that human cognition is heuristic, fractionate, and combinatory means that we should expect koans, puzzles, paradoxes, apories, and the like. We should expect that different systems possessing overlapping domains will come into conflict. We should expect them in the same way and for the same reason we should expect to encounter visual, auditory, and other kinds of systematic illusions. Because the brain picks out only the correlations it needs to predict its environments, cues predicting the systems requiring solution the way they need to be predicted to be solved. Given this, we should begin looking at traditional philosophy as a rich, discursive reservoir of pathologies, breakdowns providing information regarding the systems and misapplications involved. Like all corpses, meaning will provide a feast for worms.

In a sense, then, a koan demonstrates what a great many seem to think it’s meant to demonstrate: a genuine limit to some cognitive modality, a point where our automatic applications fail us, alerting us both to their automaticity and their specialized nature. And this, the idea would be, draws more of the automaticity (and default universal application) of the subject/object (aboutness) heuristic into deliberative purview, leading to… Enlightenment?

Does Heuristic Neglect Theory suggest a path to the Absolute?

I suppose… so long as we keep in mind that ‘Absolute’ means ‘abject stupidity.’ I think we’re better served looking at these kinds of things as boundaries rather than destinations.

The Truth Behind the Myth of Correlationism

by rsbakker

A wrong turn lies hidden in the human cultural code, an error that has scuttled our every attempt to understand consciousness and cognition. So much philosophical activity reeks of dead ends: we try and we try, and yet we find ourselves mired in the same ancient patterns of disputation. The majority of thinkers believe the problem is local, that they need only tinker with the tools they’ve inherited. They soldier on, arguing that this or that innovative modification will overcome our confusion. Some, however, believe the problem lies deeper. I’m one of those thinkers, as is Meillassoux. I think the solution lies in speculation bound to the hip of modern science, in something I call ‘heuristic neglect.’ For me, the wrong turn lies in the application of intentional cognition to solve the theoretical problem of intentional cognition. Meillassoux thinks it lies in what he calls ‘correlationism.’

Since I’ve been accused of ‘correlationism’ on a couple of occasions now, I thought it worthwhile tackling the issue in more detail. This will not be an institutional critique a la Golumbia’s, who manages to identify endless problems with Meillassoux’s presentation, while somehow entirely missing his skeptical point: once cognition becomes artifactual, it becomes very… very difficult to understand. Cognitive science is itself fractured about Meillassoux’s issue.

What follows will be a constructive critique, an attempt to explain the actual problem underwriting what Meillassoux calls ‘correlationism,’ and why his attempt to escape that problem simply collapses into more interminable philosophy. The problem that artifactuality poses to the understanding of cognition is very real, and it also happens to fall into the wheelhouse of Heuristic Neglect Theory (HNT). For those souls growing disenchanted with Speculative Realism, but unwilling to fall back into the traditional bosom, I hope to show that HNT not only offers the radical break with tradition that Meillassoux promises, it remains inextricably bound to the details of this, the most remarkable age.

What is correlationism? The experts explain:

Correlation affirms the indissoluble primacy of the relation between thought and its correlate over the metaphysical hypostatization or representational reification of either term of the relation. Correlationism is subtle: it never denies that our thoughts or utterances aim at or intend mind-independent or language-independent realities; it merely stipulates that this apparently independent dimension remains internally related to thought and language. Thus contemporary correlationism dismisses the problematic of scepticism, and or epistemology more generally, as an antiquated Cartesian hang-up: there is supposedly no problem about how we are able to adequately represent reality; since we are ‘always already’ outside ourselves and immersed in or engaging with the world (and indeed, this particular platitude is constantly touted as the great Heideggerean-Wittgensteinian insight). Note that correlationism need not privilege “thinking” or “consciousness” as the key relation—it can just as easily replace it with “being-in-the-world,” “perception,” “sensibility,” “intuition,” “affect,” or even “flesh.” Ray Brassier, Nihil Unbound, 51

By ‘correlation’ we mean the idea according to which we only ever have access to the correlation between thinking and being, and never to either term considered apart from the other. We will henceforth call correlationism any current of thought which maintains the unsurpassable character of the correlation so defined. Consequently, it becomes possible to say that every philosophy which disavows naive realism has become a variant of correlationism. Quentin Meillassoux, After Finitude, 5

Correlationism rests on an argument as simple as it is powerful, and which can be formulated in the following way: No X without givenness of X, and no theory about X without a positing of X. If you speak about something, you speak about something that is given to you, and posited by you. Consequently, the sentence: ‘X is’, means: ‘X is the correlate of thinking’ in a Cartesian sense. That is: X is the correlate of an affection, or a perception, or a conception, or of any subjective act. To be is to be a correlate, a term of a correlation . . . That is why it is impossible to conceive an absolute X, i.e., an X which would be essentially separate from a subject. We can’t know what the reality of the object in itself is because we can’t distinguish between properties which are supposed to belong to the object and properties belonging to the subjective access to the object. Quentin Meillassoux,”Time without Becoming

The claim of correlationism is the corollary of the slogan that ‘nothing is given’ to understanding: everything is mediated. Once knowing becomes an activity, then the objects insofar as they are known become artifacts in some manner: reception cannot be definitively sorted from projection and as a result no knowledge can be said to be absolute. We find ourselves trapped in the ‘correlationist circle,’ trapped in artifactual galleries, never able to explain the human-independent reality we damn well know exists. Since all cognition is mediated, all cognition is conditional somehow, even our attempts (or perhaps, especially our attempts) to account for those conditions. Any theory unable to decisively explain objectivity is a theory that cannot explain cognition. Ergo, correlationism names a failed (cognitivist) philosophical endeavour.

It’s a testament to the power of labels in philosophy, I think, because as Meillassoux himself acknowledges there’s nothing really novel about the above sketch. Explaining the ‘cognitive difference’ was my dissertation project back in the 90’s, after all, and as smitten as I was with my bullshit solution back then, I didn’t think the problem itself was anything but ancient. Given this whole website is dedicated to exploring and explaining consciousness and cognition, you could say it remains my project to this very day! One of the things I find so frustrating about the ‘critique of correlationism’ is that the real problem—the ongoing crisis—is the problem of meaning. If correlationism fails because correlationism cannot explain cognition, then the problem of correlationism is an expression of a larger problem, the problem of cognition—or in other words, the problem of intentionality.

Why is the problem of meaning an ongoing crisis? In the past six fiscal years, from 2012 to 2017, the National Institute of Health will have spent more than 113 billion dollars funding research bent on solving some corner of the human soul. [1] And this is just one public institution in one nation involving health related research. If you include the cognitive sciences more generally—research into everything from consumer behaviour to AI—you could say that solving the human soul commands more resources than any other domain in history. The reason all this money is being poured into the sciences rather than philosophy departments is that the former possesses real world consequences: diseases cured, soap sold, politicians elected. As someone who tries to keep up with developments in Continental philosophy, I already find the disconnect stupendous, how whole populations of thinkers continue discoursing as if nothing significant has changed, bitching about traditional cutlery in the shadow of the cognitive scientific tsunami.

Part of the popularity of the critique of correlationism derives from anxieties regarding the growing overlap of the sciences of the human and the humanities. All thinkers self-consciously engaged in the critique of correlationism reference scientific knowledge as a means of discrediting correlationist thought, but as far as I can tell, the project has done very little to bring the science, what we’re actually learning about consciousness and cognition, to the fore of philosophical debates. Even worse, the notion of mental and/or neural mediation is actually central to cognitive science. What some neuroscientists term ‘internal models,’ which monolopolize our access to ourselves and the world, is nothing if not a theoretical correlation of environments and cognition, trapping us in models of models. The very science that Meillassoux thinks argues against correlationism in one context, explicitly turns on it in another. The mediation of knowledge is the domain of cognitive science—full stop. A naturalistic understanding of cognition is a biological understanding is an artifactual understanding: this is why the upshot of cognitive science is so often skeptical, prone to further diminish our traditional (if not instinctive) hankering for unconditioned knowledge—to reveal it as an ancestral conceit

A kind of arche-fossil.

If an artifactual approach to cognition is doomed to misconstrue cognition, then cognitive science is a doomed enterprise. Despite the vast sums of knowledge accrued, the wondrous and fearsome social instrumentalities gained, knowledge itself will remain inexplicable. What we find lurking in the bones of Meillassoux’s critique, in other words, is precisely the same commitment to intentional exceptionality we find in all traditional philosophy, the belief that the subject matter of traditional philosophical disputation lies beyond the pale of scientific explanation… that despite the cognitive scientific tsunami, traditional intentional speculation lies secure in its ontological bunkers.

Only more philosophy, Meillassoux thinks, can overcome the ‘scandal of philosophy.’ But how is mere opinion supposed to provide bona fide knowledge of knowledge? Speculation on mathematics does nothing to ameliorate this absurdity: even though paradigmatic of objectivity, mathematics remains as inscrutable as knowledge itself. Perhaps there is some sense to be found in the notion of interrogating/theorizing objects in a bid to understand objectivity (cognition), but given what we now know regarding our cognitive shortcomings in low-information domains, we can be assured that ‘object-oriented’ approaches will bog down in disputation.

I just don’t know how to make the ‘critique of correlationism’ workable, short ignoring the very science it takes as its motivation, or just as bad, subordinating empirical discoveries to some school of ‘fundamental ontological’ speculation. If you’re willing to take such a leap of theoretical faith, you can be assured that no one in the vicinity of cognitive science will take it with you—and that you will make no difference in the mad revolution presently crashing upon us.

We know that knowledge is somehow an artifact of neural function—full stop. Meillassoux is quite right to say this renders the objectivity of knowledge very difficult to understand. But why think the problem lies in presuming the artifactual nature of cognition?—especially now that science has begun reverse-engineering that nature in earnest! What if our presumption of artifactuality weren’t so much the problem, as the characterization? What if the problem isn’t that cognitive science is artifactual so much as how it is?

After all, we’ve learned a tremendous amount about this how in the past decades: the idea of dismissing all this detail on the basis of a priori guesswork seems more than a little suspect. The track record would suggest extreme caution. As the boggling scale of the cognitive scientific project should make clear, everything turns on the biological details of cognition. We now know, for instance, that the brain employs legions of special purpose devices to navigate its environments. We know that cognition is thoroughly heuristic, that it turns on cues, bits of available information statistically correlated to systems requiring solution.

Most all systems in our environment shed information enabling the prediction of subsequent behaviours absent the mechanical particulars of that information. The human brain is exquisitely tuned to identify and exploit the correlation of information available and subsequent behaviours. The artifactuality of biology is an evolutionary one, and as such geared to the thrifty solution of high impact problems. To say that cognition (animal or human) is heuristic is to say it’s organized according to the kinds of problems our ancestors needed to solve, and not according to those belonging to academics. Human cognition consists of artifactualities, subsystems dedicated to certain kinds of problem ecologies. Moreover, it consists of artifactualities selected to answer questions quite different from those posed by philosophers.

These two facts drastically alter the landscape of the apparent problem posed by ‘correlationism.’ We have ample theoretical and empirical reasons to believe that mechanistic cognition and intentional cognition comprise two quite different cognitive regimes, the one dedicated to explanation via high-dimensional (physical) sourcing, the other dedicated to explanation absent that sourcing. As an intentional phenomena, objectivity clearly belongs to the latter. Mechanistic cognition, meanwhile, is artifactual. What if it’s the case that ‘objectivity’ is the turn of a screw in a cognitive system selected to solve in the absence of artifactual information? Since intentional cognition turns on specific cues to leverage solutions, and since those cues appear sufficient (to be the only game in town where that behaviour is concerned), the high-dimensional sourcing of that same behavior generates a philosophical crash space—and a storied one at that! What seems sourceless and self-evident becomes patently impossible.

Short magic, cognitive systems possess the environmental relationships they do thanks to super-complicated histories of natural and neural selection—evolution and learning. Let’s call this their orientation, understood as the nonintentional (‘zombie’) correlate of ‘perspective.’ The human brain is possibly the most complex thing we know of in the universe (a fact which should render any theory of the human neglecting that complexity suspect). Our cognitive systems, in other words, possess physically intractable orientations. How intractable? Enough that billions of dollars in research has merely scratched the surface.

Any capacity to cognize this relationship will perforce be radically heuristic, which is to say, provide a means to solve some critical range of problems—a problem ecology—absent natural historical information. The orientation heuristically cognized, of course, is the full-dimensional relationship we actually possess, only hacked in ways that generate solutions (repetitions of behaviour) while neglecting the physical details of that relationship.

Most significantly, orientation neglects the dimension of mediation: thought and perception (whatever they amount to) are thoroughly blind to their immediate sources. This cognitive blindness to the activity of cognition, or medial neglect, amounts to a gross insensitivity to our physical continuity with our environments, the fact that we break no thermodynamic laws. Our orientation, in other words, is characterized by a profound, structural insensitivity to its own constitution—its biological artifactuality, among other things. This auto-insensitivity, not surprisingly, includes insensitivity to the fact of this insensitivity, and thus the default presumption of sufficiency. Specialized sensitivities are required to flag insufficiencies, after all, and like all biological devices, they do not come for free. Not only are we blind to our position within the superordinate systems comprising nature, we’re blind to our blindness, and so, unable to distinguish table-scraps from a banquet, we are duped into affirming inexplicable spontanieties.

‘Truth’ belongs to our machinery for communicating (among other things) the sufficiency of iterable orientations within superordinate systems given medial neglect. You could say it’s a way to advertise clockwork positioning (functional sufficiency) absent any inkling of the clock. ‘Objectivity,’ the term denoting the supposed general property of being true apart from individual perspectives, is a deliberative contrivance derived from practical applications of ‘truth’—the product of ‘philosophical reflection.’ The problem with objectivity as a phenomenon (as opposed to ‘objectivity’ as a component of some larger cognitive articulation) is that the sufficiency of iterable orientations within superordinate systems is always a contingent affair. Whether ‘truth’ occasions sufficiency is always an open question, since the system provides, at best, a rough and ready way to communicate and/or troubleshoot orientation. Unpredictable events regularly make liars of us all. The notion of facts ‘being true’ absent the mediation of human cognition, ‘objectivity,’ also provides a rough and ready way to communicate and/or troubleshoot orientation in certain circumstances. We regularly predict felicitous orientations without the least sensitivity to their artifactual nature, absent any inkling how their pins lie in intractable high-dimensional coincidences between buzzing brains. This insensitivity generates the illusion of absolute orientation, a position outside natural regularities—a ‘view from nowhere.’ We are a worm in the gut of nature convinced we possess disembodied eyes. And so long as the consequences of our orientations remain felicitous, our conceit need not be tested. Our orientations might as well ‘stand nowhere’ absent cognition of their limits.

Thus can ‘truth’ and ‘objectivity’ be naturalized and their peculiarities explained.

The primary cognitive moral here is that lacking information has positive cognitive consequences, especially when it comes to deliberative metacognition, our attempts to understand our nature via philosophical reflection alone. Correlationism evidences this in a number of ways.

As soon as the problem of cognition is characterized as the problem of thought and being, it becomes insoluble. Intentional cognition is heuristic: it neglects the nature of the systems involved, exploiting cues correlated to the systems requiring solution instead. The application of intentional cognition to theoretical explanation, therefore, amounts to the attempt to solve natures using a system adapted to neglect natures. A great deal of traditional philosophy is dedicated to the theoretical understanding of cognition via intentional idioms—via applications of intentional cognition. Thus the morass of disputation. We presume that specialized problem-solving systems possess general application. Lacking the capacity to cognize our inability to cognize the theoretical nature of cognition, we presume sufficiency. Orientation, the relation between neural systems and their proximal and distal environments—between two systems of objects—becomes perspective, the relation between subjects (or systems of subjects) and systems of objects (environments). If one conflates the manifest artifactual nature of orientation for the artifactual nature of perspective (subjectivity), then objectivity itself becomes a subjective artifact, and therefore nothing objective at all. Since orientation characterizes our every attempt to solve for cognition, conflating it with perspective renders perspective inescapable, and objectivity all but inexplicable. Thus the crash space of traditional epistemology.

Now I know from hard experience that the typical response to the picture sketched above is to simply insist on the conflation of orientation and perspective, to assert that my position, despite its explanatory power, simply amounts to more of the same, another perspectival Klein Bottle distinctive only for its egregious ‘scientism.’ Only my intrinsically intentional perspective, I am told, allows me to claim that such perspectives are metacognitive artifacts, a consequence of medial neglect. But asserting perspective before orientation on the basis of metacognitive intuitions alone not only begs the question, it also beggars explanation, delivering the project of cognizing cognition to never-ending disputation—an inability to even formulate explananda, let alone explain anything. This is why I like asking intentionalists how many centuries of theoretical standstill we should expect before that oft advertised and never delivered breakthrough finally arrives. The sin Meillassoux attributes to correlationism, the inability to explain cognition, is really just the sin belonging to intentional philosophy as a whole. Thanks to medial neglect, metcognition,  blind to both its sources and its source blindness, insists we stand outside nature. Tackling this intuition with intentional idioms leaves our every attempt to rationalize our connection underdetermined, a matter of interminable controversy. The Scandal dwells on eternal.

I think orientation precedes perspective—and obviously so, having watched loved ones dismantled by brain disease. I think understanding the role of neglect in orientation explains the peculiarities of perspective, provides a parsimonious way to understand the apparent first-person in terms of the neglect structure belonging to the third. There’s no problem with escaping the dream tank and touching the world simply because there’s no ontological distinction between ourselves and the cosmos. We constitute a small region of a far greater territory, the proximal attuned to the distal. Understanding the heuristic nature of ‘truth’ and ‘objectivity,’ I restrict their application to adaptive problem-ecologies, and simply ask those who would turn them into something ontologically exceptional why they would trust low-dimensional intuitions over empirical data, especially when those intuitions pretty much guarantee perpetual theoretical underdetermination. Far better trust to our childhood presumptions of truth and reality, in the practical applications of these idioms, than in any one of the numberless theoretical misapplications ‘discovering’ this trust fundamentally (as opposed to situationally) ‘naïve.’

The cognitive difference, what separates the consequences of our claims, has never been about ‘subjectivity’ versus ‘objectivity,’ but rather intersystematicity, the integration of ever-more sensitive orientations possessing ever more effectiveness into the superordinate systems encompassing us all. Physically speaking, we’ve long known that this has to be the case. Short actual difference making differences, be they photons striking our retinas or compression waves striking our eardrums or so on, no difference is made. Even Meillassoux acknowledges the necessity of physical contact. What we’ve lacked is a way of seeing how our apparently immediate intentional intuitions, be they phenomenological, ontological, or normative, fit into this high-dimensional—physical—picture.

Heuristic Neglect Theory not only provides this way, it also explains why it has proven so elusive over the centuries. HNT explains the wrong turn mentioned above. The question of orientation immediately cues the systems our ancestors developed to circumvent medial neglect. Solving for our behaviourally salient environmental relationships, in other words, automatically formats the problem in intentional terms. The automaticity of the application of intentional cognition renders it apparently ‘self-evident.’

The reason the critique of correlationism and speculative realism suffer all the problems of underdetermination their proponents attribute to correlationism is that they take this very same wrong turn. How is Meillassoux’s ‘hyper-chaos,’ yet another adventure in a priori speculation, anything more than another pebble tossed upon the heap of traditional disputation? Novelty alone recommends them. Otherwise they leave us every bit as mystified, every bit as unable to accommodate the torrent of relevant scientific findings, and therefore every bit as irrelevant to the breathtaking revolutions even now sweeping us and our traditions out to sea. Like the traditions they claim to supersede, they peddle cognitive abjection, discursive immobility, in the guise of fundamental insight.

Theoretical speculation is cheap, which is why it’s so frightfully easy to make any philosophical account look bad. All you need do is start worrying definitions, then let the conceptual games begin. This is why the warrant of any account is always a global affair, why the power of Evolutionary Theory, for example, doesn’t so much lie in the immunity of its formulations to philosophical critique, but in how much it explains on nature’s dime alone. The warrant of Heuristic Neglect Theory likewise turns on the combination of parsimony and explanatory power.

Anyone arguing that HNT necessarily presupposes some X, be it ontological or normative, is simply begging the question. Doesn’t HNT presuppose the reality of intentional objectivity? Not at all. HNT certainly presupposes applications of intentional cognition, which, given medial neglect, philosophers pose as functional or ontological realities. On HNT, a theory can be true even though, high-dimensionally speaking, there is no such thing as truth. Truth talk possesses efficacy in certain practical problem-ecologies, but because it participates in solving something otherwise neglected, namely the superordinate systematicity of orientations, it remains beyond the pale of intentional resolution.

Even though sophisticated critics of eliminativism acknowledge the incoherence of the tu quoque, I realize this remains a hard twist for many (if not most) to absorb, let alone accept. But this is exactly as it should be, both insofar as something has to explain why isolating the wrong turn has proven so stupendously difficult, and because this is precisely the kind of trap we should expect, given the heuristic and fractionate nature of human cognition. ‘Knowledge’ provides a handle on the intersection of vast, high-dimensional histories, a way to manage orientations without understanding the least thing about them. To know knowledge, we will come to realize, is to know there is no such thing, simply because ‘knowing’ is a resolutely practical affair, almost certainly inscrutable to intentional cognition. When you’re in the intentional mode, this statement simply sounds preposterous—I know it once struck me as such! It’s only when you appreciate how far your intuitions have strayed from those of your childhood, back when your only applications of intentional cognition were practical, that you can see the possibility of a more continuous, intersystematic way to orient ourselves to the cosmos. There was a time before you wandered into the ancient funhouse of heuristic misapplication, when you could not distinguish between your perspective and your orientation. HNT provides a theoretical way to recover that time and take a radically different path.

As a bona fide theory of cognition, HNT provides a way to understand our spectacular inability to understand ourselves. HNT can explain ‘aporia.’ The metacognitive resources recruited for the purposes of philosophical reflection possess alarm bells—sensitivities to their own limits—relevant only to their ancestral applications. The kinds of cognitive apories (crash spaces) characterizing traditional philosophy are precisely those we might expect, given the sudden ability to exercise specialized metacognitive resources out of school, to apply, among other things, the problem-solving power of intentional cognition to the question of intentional cognition.

As a bona fide theory of cognition, HNT bears as much on artificial cognition as on biological cognition, and as such, can be used to understand and navigate the already radical and accelerating transformation of our cognitive ecologies. HNT scales, from the subpersonal to the social, and this means that HNT is relevant to the technological madness of the now.

As a bona fide empirical theory, HNT, unlike any traditional theory of intentionality, will be sorted. Either science will find that metacognition actually neglects information in the ways I propose, or it won’t. Either science will find this neglect possesses the consequences I theorize, or it won’t. Nothing exceptional and contentious is required. With our growing understanding of the brain and consciousness comes a growing understanding of information access and processing capacity—and the neglect structures that fall out of them. The human brain abounds in bottlenecks, none of which are more dramatic than consciousness itself.

Cognition is biomechanical. The ‘correlation of thought and being,’ on my account, is the correlation of being and being. The ontology of HNT is resolutely flat. Once we understand that we only glimpse as much of our orientations as our ancestors required for reproduction, and nothing more, we can see that ‘thought,’ whatever it amounts to, is material through and through.

The evidence of this lies strewn throughout the cognitive wreckage of speculation, the alien crash site of philosophy.



[1] This includes, in addition to the neurosciences proper, research into Basic Behavioral and Social Science (8.597 billion), Behavioral and Social Science (22.515 billion), Brain Disorders (23.702 billion), Mental Health (13.699 billion), and Neurodegenerative (10.183 billion). 21/01/2017


Framing “On Alien Philosophy”…

by rsbakker


Peter Hankins of Conscious Entities fame has a piece considering “On Alien Philosophy.” The debate is just getting started, but I thought it worthwhile explaining why I think this particular paper of mine amounts to more than yet just another interpretation to heap onto the intractable problem of ourselves.

Consider the four following claims:

1) We have biologically constrained (in terms of information access and processing resources) metacognitive capacities ancestrally tuned to the solution of various practical problem ecologies, and capable of exaptation to various other problems.

2) ‘Philosophical reflection’ constitutes such an exaptation.

3) All heuristic exaptations inherit, to some extent, the problem-solving limitations of the heuristic exapted.

4) ‘Philosophical reflection’ inherits the problem-solving limitations of deliberative metacognition.

Now I don’t think there’s much anything controversial about any of these claims (though, to be certain, there’s a great many devils lurking in the details adduced). So note what happens when we add the following:

5) We should expect human philosophical practice will express, in a variety of ways, the problem-solving limitations of deliberative metacognition.

Which seems equally safe. But note how the terrain of the philosophical debate regarding the nature of the soul has changed. Any claim purporting the exceptional nature of this or that intentional phenomena now needs to run the gauntlet of (5). Why assume we cognize something ontologically exceptional when we know we are bound to be duped somehow? All things being equal, mediocre explanations will always trump exceptional ones, after all.

The challenge of (5) has been around for quite some time, but if you read (precritical) eliminativists like Churchland, Stich, or Rosenberg, this is where the battle grinds to a standstill. Why? Because they have no general account of how the inevitable problem-solving limitations of deliberative metacognition would be expressed in human philosophical practice, let alone how they would generate the appearance of intentional phenomena. Since all they have are promissory notes and suggestive gestures, ontologically exceptional accounts remain the only game in town. So, despite the power of (5), the only way to speak of intentional phenomena remains the traditional, philosophical one. Science is blind without theory, so absent any eliminativist account of intentional phenomena, it has no clear way to proceed with their investigation. So it hews to exceptional posits, trusting in their local efficacy, and assuming they will be demystified by discoveries to come.

Thus the challenge posed by Alien Philosophy. By giving real, abductive teeth to (5), my account overturns the argumentative terrain between eliminativism and intentionalism by transforming the explanatory stakes. It shows us how stupidity, understood ecologically, provides everything we need to understand our otherwise baffling intuitions regarding intentional phenomena. “On Alien Philosophy” challenges the Intentionalist to explain more with less (the very thing, of course, he or she cannot do).

Now I think I’ve solved the problem, that I have a way to genuinely naturalize meaning and cognition. The science will sort my pretensions in due course, but in the meantime, the heuristic neglect account of intentionality, given its combination of mediocrity and explanatory power, has to be regarded as a serious contender.

It Is What It Is (Until Notified Otherwise)

by rsbakker



The thing to always remember when one finds oneself in the middle of some historically intractable philosophical debate is that path-dependency is somehow to blame. This is simply to say that the problem is historical in that squabbles regarding theoretical natures always arises from some background of relatively problem-free practical application. At some point, some turn is taken and things that seem trivially obvious suddenly seem stupendously mysterious. St. Augustine, in addition to giving us one of the most famous quotes in philosophy, gives us a wonderful example of this in The Confessions when he writes:

“What, then, is time? If no one asks of me, I know; if I wish to explain to him who asks, I know not.” XI, XIV, 17

But the rather sobering fact is that this is the case with a great number of the second order questions we can pose. What is mathematics? What’s a rule? What’s meaning? What’s cause? And of course, what is phenomenal consciousness?

So what is it with second order interrogations? Why is ‘time talk’ so easy and effortlessly used even though we find ourselves gobsmacked each and every time someone asks what time qua time is? It seems pretty clear that either we lack the information required or the capacity required or some nefarious combination of both. If framing the problem like this sounds like a no-brainer, that’s because it is a no-brainer. The remarkable thing lies in the way it recasts the issue at stake, because as it turns out, the question of the information and capacity we have available is a biological one, and this provides a cognitive ecological means of tackling the problem. Since practical solving for time (‘timing’) is obviously central to survival, it makes sense that we would possess the information access and cognitive capacity required to solve a wide variety of timing issues. Given that theoretical solving for time (qua-time) isn’t central to survival (no species does it and only our species attempts it), it makes sense that we wouldn’t possess the information access and cognitive capacity required, that we would suffer time-qua-time blindness.

From a cognitive ecological perspective, in other words, St. Augustine’s perplexity should come as no surprise at all. Of course solving time-qua-time is mystifying: we evolved the access and capacity required for solving the practical problems of timing, and not the theoretical problem of time. Now I admit if the cognitive ecological approach ground to a halt here it wouldn’t be terribly illuminating, but there’s quite a bit more to be said: it turns out cognitive ecology is highly suggestive of the different ways we might expect our attempts to solve things like time-qua-time to break down.

What would it be like to reach the problem-solving limits of some practically oriented problem-solving mode? Well, we should expect our assumptions/intuitions to stop delivering answers. My daughter is presently going through a ‘cootie-catcher’ phase and is continually instructing me to ask questions, then upbraiding me when my queries don’t fit the matrix of possible ‘answers’ provided by the cootie-catcher (yes, no, and versions of maybe). Sometimes she catches these ill-posed questions immediately, and sometimes she doesn’t catch them until the cootie-catcher generates a nonsensical response.


Now imagine your child never revealed their cootie-catcher to you: you asked questions, then picked colours or numbers or animals, and it turned out some were intelligibly answered, and some were not. Very quickly you would suss out the kinds of questions that could be asked, and the kinds that could not. Now imagine unbeknownst to you that your child replaced their cootie-catcher with a computer running two separately tasked, distributed AlphaGo type programs, the first trained to provide well-formed (if not necessarily true) answers to basic questions regarding causality and nothing else, the second trained to provide well-formed (if not necessarily true) answers to basic questions regarding goals and intent. What kind of conclusions would you draw, or more importantly, assume? Over time you would come to suss out the questions generating ill-formed answers versus questions generating well-formed ones. But you would have no way of knowing that two functionally distinct systems were responsible for the well-formed answers: causal and purposive modes would seem the product of one cognitive system. In the absence of distinctions you would presume unity.

Think of the difference between Plato likening memory to an aviary in the Theaetetus and the fractionate, generative memory we now know to be the case. The fact that Plato assumed as much, unity and retrieval, shouts something incredibly important once placed in a cognitive ecological context. What it suggests is that purely deliberative attempts to solve second-order problems, to ask questions like what is memory-qua-memory, will almost certainly run afoul the problem of default identity, the identification that comes about for the want of distinctions. To return to our cootie-catcher example, it’s not simply that we would report unity regarding our child’s two AlphaGo type programs the way Plato did with memory, it’s that information involving its dual structure would play no role in our cognitive economy whatsoever. Unity, you could say, is the assumption built into the system. (And this applies as much to AI as it does to human beings. The first ‘driverless fatality’ died because his Tesla Model S failed to distinguish a truck trailer from the sky.)

Default identity, I think, can play havoc with even the most careful philosophical interrogations—such as the one Eric Schwitzgebel gives in the course of rebutting Keith Frankish, both on his blog and in his response in The Journal of Consciousness Studies, “Phenomenal Consciousness, Defined and Defended as Innocently as I Can Manage.”

According to Eric, “Illusionism as a Theory of Consciousness” presents the phenomenal realist with a dilemma: either they commit to puzzling ontological features such as simple, ineffable, intrinsic, or so on, or they commit to explaining those features away, which is to say, some variety of Illusionism. Since Eric both believes that phenomenal consciousness is real, and that the extraordinary properties attributed to it are likely not real, he proposes a third way, a formulation of phenomenal experience that neither inflates it into something untenable, nor deflates into something that is plainly not phenomenal experience. “The best way to meet Frankish’s challenge,” he writes, “is to provide something that the field of consciousness studies in any case needs: a clear definition of phenomenal consciousness, a definition that targets a phenomenon that is both substantively interesting in the way that phenomenal consciousness is widely thought to be interesting but also innocent of problematic metaphysical and epistemological assumptions” (2).

It’s worth noting the upshot of what Eric is saying here: the scientific study of phenomenal consciousness cannot, as yet, even formulate their primary explanandum. The trick, as he sees it, is to find some conceptual way to avoid the baggage, while holding onto some semblance of a wardrobe. And his solution, you might say, is to wear as many outfits as he possibly can. He proposes that definition by example is uniquely suited to anchor an ontologically and epistemologically innocent concept of phenomenal consciousness.

He has but one caveat: any adequate formulation of phenomenal consciousness has to account or allow for what Eric terms its ‘wonderfulness’:

If the reduction of phenomenal consciousness to something physical or functional or “easy” is possible, it should take some work. It should not be obviously so, just on the surface of the definition. We should be able to wonder how consciousness could possibly arise from functional mechanisms and matter in motion. Call this the wonderfulness condition. 3

He concedes the traditional properties ascribed to phenomenal experience outrun naturalistic credulity, but the feature of begging belief remains to be explained. This is the part of Eric’s position to keep an eye on because it means his key defense against eliminativism is abductive. Whatever phenomenal consciousness is, it seems safe to say it is not something easily solved. Any account purporting to solve phenomenal consciousness that leaves the wonderfulness condition unsatisfied is likely missing phenomenal consciousness altogether.

And so Eric provides a list of positive examples including sensory and somatic experiences, conscious imagery, emotional experience, thinking and desiring, dreams, and even other people, insofar as we continually attribute these very same kinds of experiences to them. By way of negative examples, he mentions a variety of intimate, yet obviously not phenomenally conscious processes, such as fingernail growth, intestinal lipid absorption, and so on.

He writes:

Phenomenal consciousness is the most folk psychologically obvious thing or feature that the positive examples possess and that the negative examples lack. I do think that there is one very obvious feature that ties together sensory experiences, imagery experiences, emotional experiences, dream experiences, and conscious thoughts and desires. They’re all conscious experiences. None of the other stuff is experienced (lipid absorption, the tactile smoothness of your desk, etc.). I hope it feels to you like I have belabored an obvious point. Indeed, my argumentative strategy relies upon this obviousness. 8

Intuition, the apparent obviousness of his examples, is what he stresses here. The beauty of definition by example is that offering instances of the phenomenon at issue allows you to remain agnostic regarding the properties possessed by that phenomenon. It actually seems to deliver the very metaphysical and epistemological innocence Eric needs to stave off the charge of inflation. It really does allow him to ditch the baggage and travel wearing all his clothes, or so it seems.

Meanwhile the wonderfulness condition, though determining the phenomenon, does so indirectly, via the obvious impact it has on human attempts to cognize experience-qua-experience. Whatever phenomenal consciousness is, contemplating it provokes wonder.

And so the argument is laid out, as spare and elegant as all of Eric’s arguments. It’s pretty clear these are examples of whatever it is we call phenomenal consciousness. Of course, there’s something about them that we find downright stupefying. Surely, he asks, we can be phenomenal realists in this austere respect?

For all its intuitive appeal, the problem with this approach is that it almost certainly presumes a simplicity that human cognition does not possess. Conceptually, we can bring this out with a single question: Is phenomenal consciousness the most folk psychologically obvious thing or feature the examples share, or is it obvious in some other respect? Eric’s claim amounts to saying the recognition of phenomenal consciousness as such belongs to everyday cognition. But is this the case? Typically, recognition of experience-qua-experience is thought to be an intellectual achievement of some kind, a first step toward the ‘philosophical’ or ‘reflective’ or ‘contemplative’ attitude. Shouldn’t we say, rather, that phenomenal consciousness is the most obvious thing or feature these examples share upon reflection, which is to say, philosophically?

This alternative need only be raised to drag Eric’s formulation back into the mire of conceptual definition, I think. But on a cognitive ecological picture, we can actually reframe this conceptual problematization in path-dependent terms, and so more forcefully insist on a distinction of modes and therefore a distinction in problem-solving ecologies. Recall Augustine, how we understand time without difficulty until we ask the question of time qua time. Our cognitive systems have no serious difficulty with timing, but then abruptly break down when we ask the question of time as such. Even though we had the information and capacity required to solve any number of practical issues involving time, as soon as we pose the question of time-qua-time that fluency evaporates and we find ourselves out-and-out mystified.

Eric’s definition by example, as an explicitly conceptual exercise, clearly involves something more than everyday applications of experience talk. The answer intuitively feels as natural as can be—there must be some property X these instances share or exclude, certainly!—but the question strikes most everyone as exceptional, at least until they grow accustomed to it. Raising the question, as Augustine shows us, is precisely where the problem begins, and as my daughter would be quick to remind Eric, cootie-catchers only work if we ask the right question. Human cognition is fractionate and heuristic, after all.


All organisms are immersed in potential information, difference making differences that could spell the difference between life and death. Given the difficulties involved in the isolation of causes, they often settle for correlations, cues reliably linked to the systems requiring solution. In fact, correlations are the only source of information organisms have, evolved and learned sensitivities to effects systematically correlated to those environmental systems relevant to reproduction. Human beings, like all other living organisms, are shallow information consumers adapted to deep information environments, sensory cherry pickers, bent on deriving as much behaviour from as little information as possible.

We only have access to so much, and we only have so much capacity to derive behaviour from that access (behaviour which in turn leverages capacity). Since the kinds of problems we face outrun access, and since those problems and the resources required to solve them are wildly disparate, not all access is equal.

Information access, I think, divides cognition into two distinct forms, two different families of ‘AlphaGo type’ programs. On the one hand we have what might be called source sensitive cognition, where physical (high-dimensional) constraints can be identified, and on the other we have source insensitive cognition, where they cannot.

Since every cause is an effect, and every effect is a cause, explaining natural phenomena as effects always raises the question of further causes. Source sensitive cognition turns on access to the causal world, and to this extent, remains perpetually open to that world, and thus, to the prospect of more information. This is why it possesses such wide environmental applicability: there are always more sources to be investigated. These may not be immediately obvious to us—think of visible versus invisible light—but they exist nonetheless, which is why once the application of source sensitivity became scientifically institutionalized, hunting sources became a matter of overcoming our ancestral sensory bottlenecks.

Since every natural phenomena has natural constraints, explaining natural phenomena in terms of something other than natural constraints entails neglect of natural constraints. Source insensitive cognition is always a form of heuristic cognition, a system adapted to the solution of systems absent access to what actually makes them tick. Source insensitive cognition exploits cues, accessible information invisibly yet sufficiently correlated to the systems requiring solution to reliably solve those systems. As the distillation of specific, high-impact ancestral problems, source insensitive cognition is domain-specific, a way to cope with systems that cannot be effectively cognized any other way.

(AI approaches turning on recurrent neural networks provide an excellent ex situ example of the necessity, the efficacy, and the limitations of source insensitive (cue correlative) cognition. Andrei Cimpian’s lab and the work of Klaus Fiedler (as well as that of the Adaptive Behaviour and Cognition Research Group more generally) are providing, I think, an evolving empirical picture of source insensitive cognition in humans, albeit, absent the global theoretical framework provided here.)

So what are we to make of Eric’s attempt to innocently (folk psychologically) pose the question of experience-qua-experience in light of this rudimentary distinction?

If one takes the brain’s ability to cognize its own cognitive functions as a condition of ‘experience talk,’ it becomes very clear very quickly that experience talk belongs to a source insensitive cognitive regime, a system adapted to exploit correlations between the information consumed (cues) and the vastly complicated systems (oneself and others) requiring solution. This suggests that Eric’s definition by example is anything but theoretically innocent, assuming, as it does, that our source insensitive, experience-talk systems pick out something in the domain of source sensitive cognition… something ‘real.’ Defining by example cues our experience-talk system, which produces indubitable instances of recognition. Phenomenal consciousness becomes, apparently, an indubitable something. Given our inability to distinguish between our own cognitive systems (given ‘cognition-qua-cognition blindness’), default identity prevails; suddenly it seems obvious that phenomenal experience somehow, minimally, belongs to the order of the real. And once again, we find ourselves attempting to square ‘posits’ belonging to sourceless modes of cognition with a world where everything has a source.

We can now see how the wonderfulness condition, which Eric sees working in concert with his definition by example, actually cuts against it. Experience-qua-experience provokes wonder precisely because it delivers us to crash space, the point where heuristic misapplication leads our intuitions astray. Simply by asking this question, we have taken a component from a source insensitive cognitive system relying (qua heuristic) on strategic correlations to the systems requiring solution to solve, and asked a completely different, source sensitive system to make sense of it. Philosophical reflection is a ‘cultural achievement’ precisely because it involves using our brains in new ways, applying ancient tools to novel questions. Doing so, however, inevitably leaves us stumbling around in a darkness we cannot see, running afoul confounds we have no way of intuiting, simply because they impacted our ancestors not at all. Small wonder ‘phenomenal consciousness’ provokes wonder. How could the most obvious thing possess so few degrees of cognitive freedom? How could light itself deliver us to darkness?

I appreciate the counterintuitive nature of the view I’m presenting here, the way it requires seeing conceptual moves in terms of physical path-dependencies, as belonging to a heuristic gearbox where our numbness to the grinding perpetually convinces us that this time, at long last, we have slipped from neutral into drive. But recall the case of memory, the way blindness to its neurocognitive intricacies led Plato to assume it simple. Only now can we run our (exceedingly dim) metacognitive impressions of memory through the gamut of what we know, see it as a garden of forking paths. The suggestion here is that posing the question of experience-qua-experience poses a crucial fork in the consciousness studies road, the point where a component of source-insensitive cognition, ‘experience,’ finds itself dragged into the court of source sensitivity, and productive inquiry grinds to a general halt.

When I employ experience talk in a practical, first-order way, I have a great deal of confidence in that talk. But when I employ experience talk in a theoretical, second-order way, I have next to no confidence in that talk. Why would I? Why would anyone, given the near-certainty of chronic underdetermination? Even more, I can see of no way (short magic) for our brain to have anything other than radically opportunistic and heuristic contact with its own functions. Either specialized, simple heuristics comprise deliberative metacognition or deliberative metacognition does not exist. In other words, I see no way of avoiding experience-qua-experience blindness.

This flat out means that on a high dimensional view (one open to as much relevant physical information as possible), there is just no such thing as ‘phenomenal consciousness.’ I am forced to rely on experience related talk in theoretical contexts all the time, as do scientists in countless lines of research. There is no doubt whatsoever that experience-talk draws water from far more than just ‘folk psychological’ wells. But this just means that various forms of heuristic cognition can be adapted to various experimentally regimented cognitive ecologies—experience-talk can be operationalized. It would be strange if this weren’t the case, and it does nothing to alleviate the fact that solving for experience-qua-experience delivers us, time and again, to crash space.

One does not have to believe in the reality of phenomenal consciousness to believe in the reality of the systems employing experience-talk. As we are beginning to discover, the puzzle has never been one of figuring out what phenomenal experiences could possibly be, but rather figuring out the biological systems that employ them. The greater our understanding of this, the greater our understanding of the confounds characterizing that perennial crash space we call philosophy.

Breakneck: Review and Critical Commentary of Whiplash: How to Survive our Faster Future by Joi Ito and Jeff Howe

by rsbakker


The thesis I would like to explore here is that Whiplash by Joi Ito and Jeff Howe is at once a local survival guide and a global suicide manual. Their goal “is no less ambitious than to provide a user’s manual to the twenty-first century” (246), a “system of mythologies” (108) embodying the accumulated wisdom of the storied MIT Media Lab. Since this runs parallel to my own project, I applaud their attempt. Like them, I think understanding the consequences of the ongoing technological revolution demands “an entirely new mode of thinking—a cognitive evolution on the scale of a quadruped learning to stand on its hind feet” (247). I just think we need to recall the number of extinctions that particular evolutionary feat required.

Whiplash was a genuine delight for me to read, and not simply because I’m a sucker for technoscientific anecdotes. At so many points I identified with the collection of misfits and outsiders that populate their tales. So, as an individual who fairly embodies the values promulgated in this book, I offer my own amendments to Ito and Howe’s heuristic source code, what I think is a more elegant and scientifically consilient way to understand not only our present dilemma, but the kinds of heuristics we will need to survive them…

Insofar as that is possible.


Emergence over Authority

General Idea: Pace of change assures normative obsolescence, which in turn requires openness to ‘emergence.’

“Emergent systems presume that every individual within that system possesses unique intelligence that would benefit the group.” 47

“Unlike authoritarian systems, which enable only incremental change, emergent systems foster the kind of nonlinear innovation that can react quickly to the kind of change of rapid changes that characterize the network age.” 48

Problems: Insensitive to the complexities of the accelerating social and technical landscape. The moral here should be, Does this heuristic still apply?

The quote above also points to the larger problem, which becomes clear by simply rephrasing it to read, ‘emergent systems foster the kind of nonlinear transformation that can react quickly to the kind of nonlinear transformations that characterize the network age.’ The problem, in other words, is also the solution. Call this the Putting Out Fire with Gasoline Problem. I wish Ito and Howe would have spent some more time considering it since it really is the heart of their strategy: How do we cope with accelerating innovation? We become as quick and innovative as we can.


Pull over Push

General Idea: Command and control over warehoused resources lacks the sensitivity to solve many modern problems, which are far better resolved by allowing the problems themselves to attract the solvers.

“In the upside-down, bizarre universe created by the Internet, the very assets on your balance sheet—from printing presses to lines of code—are now liabilities from the perspective of agility. Instead, we should try to use resources that can be utilized just in time, for just that time necessary, then relinquished.” 69

“As the cost of innovation continues to fall, entire communities that have been sidelined by those in power will be able to organize themselves and become active participants in society and government. The culture emergent innovation will allow everyone to feel a sense of both ownership and responsibility to each other and to the rest of the world, which will empower them to create more lasting change that the authorities who write policy and law.” 71

Problems: In one sense, I think this chapter speaks to the narrow focus of the book, the degree it views the world through IT glasses. Trump examples the power of Pull. ISIS examples the power of Pull. ‘Empowerment’ is usually charged with positive connotations, until one applies it to criminals, authoritarian governments and so on. It’s important to realize that ‘pull’ runs any which way, rather than directly toward better.


Compasses over Maps

General Idea: Sensitivity to ongoing ‘facts on the ground’ generally trumps reliance on high-altitude appraisals of yesterday’s landscape.

“Of all the nine principles in the book, compasses over maps has the greatest potential for misunderstanding. It’s actually very straightforward: a map implies a detailed knowledge of the terrain, and the existence of an optimum route; the compass is a far more flexible tool and requires the user to employ creativity and autonomy in discovering his or her own path.” 89

Problems: I actually agree that this principle is the most apt to be misunderstood because I’m inclined to think Ito and Howe themselves might be misunderstanding it! Once again, we need to see the issue in terms of cognitive ecology: Our ancestors, you could say, suffered a shallow present and enjoyed a deep future. Because the mechanics of their world eluded them, they had no way of re-engineering them, and so they could trust the machinery to trundle along the way it always had. We find ourselves in the opposite predicament: As we master more and more of the mechanics of our world, we discover an ever-expanding array of ways to re-engineering them, meaning we can no longer rely on the established machinery the way our ancestors—and here’s the important bit—evolved to. We are shallow present, deep future creatures living in a deep present, shallow future world.

This, I think, is what Ito and Howe are driving at: just as the old rules (authorities) no longer apply, the old representations (maps) no longer apply either, forcing us to gerrymander (orienteer) our path.


Risk over Safety

General Idea: The cost of experimentation has plummeted to such an extent that being wrong no longer has the catastrophic market consequences it once had.

“The new rule, then, is to embrace risk. There may be nowhere else in this book that exemplifies how far our collective brains have fallen behind our technology.” 116

“Seventy million years ago it was great to be a dinosaur. You were a complete package; big, thick-skinned, sharp-toothed, cold-blooded, long-lived. And it was great for a long, long time. Then, suddenly… it wasn’t so great. Because of your size, you needed an awful lot of calories. And you needed an awful lot of room. So you died. You know who outlived you? The frog.” 120

Problems: Essentially the argument is that risky ventures in the old economy are now safe, and that safe ventures are now risky, which means the argument is actually a ‘safety over risk’ one. I find this particular maxim so interesting because I think it really throws their lack of any theory of the problem they take themselves to be solving/ameliorating into relief. Really the moral here is experimentation pays.

This means the cognitive ecology Ito and Howe are both describing and advocating is in some sense antithetical—and therefore alienating—to our ancestral ways of making sense of ourselves.


Disobedience over Compliance

General Idea: Traditional forms of development stifle the very creativity institutions require to adapt to the accelerating pace of technological change.

“Since the 1970’s, social scientists have recognized the positive impact of “positive deviants,” people whose unorthodox behavior improves their lives and has the potential to improve their communities if it’s adopted more widely.” 141

“The people who will be the most successful in this environment will be the ones who ask questions, trust their instincts, and refuse to follow the rules when the rules get in their way.” 141

Problems: Disobedience is not critique, and Ito and Howe are careful to point this out, but they fail to mention what role, if any, criticality plays in their list of principles. Another problem has to do with the obvious exception bias at work in their account. Sure, being positive deviants has served Ito and Howe and the generally successful people they count as their ingroup, but what about the rest of us? This is why I cringe every time I hear Oscar acceptance speeches urging young wannabe thespians to ‘never give up on their dream,’ because winners—who are winners by virtue of being the exception—see themselves as proof positive that it can be done if you just try-try-try… This stuff is what powers the great dream smashing factory called Hollywood—as well as Silicon Valley. All things being equal, I think being a ‘positive deviant’ is bound to generate far more grief than success.

And this, I think, underscores the fundamental problem with the book, which is the question of application. I like to think of myself as a ‘positive deviant,’ but I’m aware that I am often identified as a ‘contrarian flake’ in the various academic silos I piss in now and again. By opening research ingroups to the wider world, the web immediately requires members to vet communications in a manner they never had to before. The world, as it turns out, is filled with contrarian flakes, so the problem becomes one of sorting positive deviants (like myself (maybe)), extra-institutional individuals with positive contributions to make, from all those contrarian flakes (like myself (maybe)).

Likewise, given that every communal enterprise possesses wilful, impassioned, but unimaginative employees, how does a manager sort the ‘positive deviant’ out?

When does disobedience over compliance apply? This is where the rubber hits the road, I think. The whole point of the (generally fascinating) anecdotes is to address this very issue, but aside from some gut estimation of analogical sufficiency between cases, we really have nothing to go on.


Practice over Theory

General Idea: Traditional forms of education and production emphasize planning before and learning outside the relevant context of applications, when humans are simply not wired for this, and when those contexts are transforming so quickly.

“Putting practice over theory means recognizing that in a faster future, in which change has become a new constant, there is often a higher cost to waiting and planning that there is to doing and improvising.” 159

“The Media Lab is focussed on interest-driven, passion-driven learning through doing. It is also trying to understand and deploy this form of creative learning into a society that will increasingly need more creative learners and fewer human beings who can solve problems better tackled by robots and computers.” 170

Problems: Humans are the gerrymandering species par excellence, leveraging technical skills into more and more forms of environmental mastery. In this respect it’s hard to argue against Ito and Howe’s point, given the caveats they are careful to provide.

The problem lies in the supercomplex environmental consequences of that environmental mastery: Whiplash is advertised as a how-to environmentally master the consequences of environmental mastery manual, so obviously, environmental mastery, technical innovation, ‘progress’—whatever you want to call it—has become a life and death matter, something to be ‘survived.’

The thing people really need to realize in these kinds of discussions is just how far we have sailed into uncharted waters, and just how fast the wind is about to grow.


Diversity over Ability

General Idea: Crowdsourcing, basically, the term Jeff Howe coined referring to the way large numbers of people from a wide variety of backgrounds can generate solutions eluding experts.

“We’re inclined to believe the smartest, best trained people in a given discipline—the experts—are the best qualified to a solve a problem in their specialty. And indeed, they often are. When they fail, as they will from time to time, our unquestioning faith in the principle of ‘ability’ leads us to imagine that we need to find a better solver: other experts with similarly high levels of training. But it is in the nature of high ability to reproduce itself—the new team of experts, it turns out, trained at the same amazing schools, institutes, and companies as the previous experts. Similarly brilliant, out two sets of experts can be relied on to apply the same methods to the problem, and share as well the same biases, blind spots, and unconscious tendencies.” 183

Problems: Again I find myself troubled not so much by the moral as by the articulation. If you switch the register from ‘ability’ to competence and consider the way ingroup adjudications of competence systematically perceive outgroup contributions to be incompetent, then you have a better model to work with here, I think. Each of us carry a supercomputer in our heads and all cognition exhibits path-dependency and is therefore vulnerable to blind alleys, so the power of distributed problem solving should come as no surprise. The problem, here, rather, is one of seeing though our ingroup blinders, and coming to understand how we instinctively identify competence forecloses on distributed cognitive resources (which can take innumerable forms).

Institutionalizing diversity seems like a good first step. But what about overcoming ingroup biases more generally? And what about the blind-alley problem (which could be called the ‘double-blind alley problem,’ given the way reviewing the steps taken tends to confirm the necessity of the path taken)? Is there a way to suss out the more pernicious consequences of cognitive path-dependency?


Resilience over Strength

General Idea: The reed versus the tree.

Problems: It’s hard to bitch about a chapter beginning with a supercool Thulsa Doom quote.

Strike that—impossible.


Systems over Objects

General Idea: Unravelling contemporary problems means unravelling complex problems necessitating adoption of the systems view.

“These new problems, whether we’re talking about curing Alzheimer’s or learning to predict volatile weather systems, seem to be fundamentally different, in that they seem to require the discovery of all the building blocks in a complex system.” 220

“Systems over objects recognizes that responsible innovation requires more than speed and efficiency. It also requires a constant focus on the overall impact of new technologies, and an understanding of the connections between people, their communities, and their environments.” 224

Problems: Since so much of Three Pound Brain is dedicated to understanding human experience and cognition in naturally continuous terms, I tend to think that ‘Systems over Subjects’ offers a more penetrating approach. The idea that things and events cannot be understood or appreciated in isolation is already firmly rooted in our institutional DNA, I think. The challenge, here, lies in squaring this way of thinking with everyday cognition, with our default ways of making sense of each other and ourselves. We are hardwired to see simple essences and sourceless causes everywhere we look. This means the cognitive ecology Ito and Howe are both describing and advocating is in some sense antithetical—and therefore alienating—to our ancestral ways of making sense of ourselves.

Algorithms are set to flood this space, to begin cuing social cognition to solve biological brains in the absence of any biological brains.



When I decided to post a review on this book, I opened an MSWord doc the way I usually do and began jotting down jumbled thoughts and impressions, including the reminder to “Bring up the problem of theorizing politics absent any account of human nature.” I had just finished reading the introduction by that point, so I read the bulk of Whiplash with this niggling thought in the back of my mind. Ito and Howe take care to avoid explicit political references, but as I’m sure they will admit, their project is political through and through. Politics has always involved science fiction; after all, how do you improve a future you can’t predict? Knowing human nature, their need to eat, to secure prestige, to mate, to procreate, and so on, is the only thing that allows us to predict human futures at all. Dystopias beg Utopias beg knowing what makes us tick.

In a time of radical, exponential social and environmental transformation, the primary question regarding human nature has to involve adaptability, our ability to cope with social and environmental transformation. The more we learn about human cognition, however, the more we discover that the human capacity to solve new problems is modular as opposed to monolithic, complex as opposed to simple. This in turn means that transforming different elements in our environments (the way technology does) can have surprising results.

So for example, given the ancestral stability of group sizes, it makes sense to suppose we would assess the risk of victimization against a fixed baseline whenever we encountered information regarding violence. Our ability to intuitively assess threats, in other words, depends upon a specific cognitive ecology, one where the information available is commensurate with the small communities of farmers and/or hunter-gatherers. This suggests the provision of ‘deep’ (ancestrally unavailable) threat information, such as that provided by the web or the evening news, would play havoc with our threat intuitions—as indeed seems to be the case.

Human cognition is heuristic, through and through, which is to say dependent on environmental invariances, the ancestral stability of different relevant backgrounds. The relation between group size and threat information is but one of countless default assumptions informing our daily lives. The more technology transforms our cognitive ecologies, the more we should expect our intuitions to misfire, to prompt ineffective problem-solving behaviour like voting for ‘tough-on-crime’ political candidates. The fact is technology makes things easy that were never ‘meant’ to be easy. Consider how humans depended on all the people they knew before the industrial concentration of production, and so were forced to compromise, to see themselves as requiring friends and neighbours. You could source your clothes, your food, even your stories and religion to some familiar face. You grew up in an atmosphere of ambient, ingroup gratitude that continually counterbalanced your selfish impulses. After the industrial concentration of production, the material dependencies enforcing cooperation evaporated, allowing humans to indulge egocentric intuitions, the sweet-tooth of themselves, and ‘individualism’ was born, and with it all the varieties of social isolation comprising the ‘modern malaise.’

This cognitive ecological lens is the reason why I’ve been warning that the web was likely to aggravate processes of group identification and counter-identification, why I’ve argued that the tactics of 20th century progressivism had actually become more pernicious than efficacious, and suggested that forms of political atavism, even the rise of demagoguery, would become bigger and bigger problems. Where most of the world saw the Arab Spring as a forceful example of the web’s capacity to emancipate, I saw it as an example of ‘flash civil unrest,’ the ability of populations to spontaneously organize and overthrow existing institutional orders period, and only incidentally ‘for the better.’

If you entertained extremist impulses before the internet, you had no choice but to air your views with your friends and neighbours, where, all things being equal, the preponderance of views would be more moderate. The network constraints imposed by geography, I surmised, had the effect of ameliorating extremist tendencies. Absent the difficulty of organizing about our darker instincts, rationalizing and advertising them, I think we have good reason to fear. Humans are tribal through and through, as prone to acts of outgroup violence as ingroup self-sacrifice. On the cognitive ecological picture, it just so happens that technological progress and moral/political progress have marched hand in hand thus far. The bulk of our prosocial, democratic institutions were developed—at horrendous cost, no less—to maximize the ‘better angels’ of our natures and to minimize the worst, to engineer the kind of cognitive ecologies we required to flourish in the new social and technical environments—such as the industrial concentration of material dependency—falling out of the Renaissance and Enlightenment.

I readily acknowledge that better accounts can be found for the social phenomena considered above: what I contend is that all of those accounts will involve some nuanced understanding of the heuristic nature of human cognition and the kinds of ecological invariance they take for granted. My further contention is that any adequate understanding of that heuristic nature raises the likelihood, perhaps even the inevitability, that human social cognition will effectively breakdown altogether. The problem lies in the radically heuristic nature of the cognitive modes we use to understand each other and ourselves. Since the complexity of our biocomputational nature renders it intractable, we had to develop ways of predicting/explaining/manipulating behaviour that have nothing to do with the brains behind that behaviour, and everything to do with its impact on our reproductive fortunes. Social problem-solving, in other words, depends on the stability of a very specific cognitive ecology, one entirely innocent to the possibility of AI.

For me, the most significant revelation from the Ashley Madison scandal was the ease with which men were fooled into thinking they were attracting female interest. And this just wasn’t an artifact of the venue: Ito’s MIT colleague Sherry Turkle, in addition to systematically describing the impact of technology on interpersonal relationships, often warns of the ease with which “Darwinian buttons” can be pushed. What makes simple heuristics so powerful is precisely what renders them so vulnerable (and it’s no accident that AI is struggling to overcome this issue now): they turn on cues physically correlated to the systems they track. Break those correlations, and those cues are connected to nothing at all, and we enter Crash Space, the kind of catastrophic cognitive ecological failure that warns away everyone but philosophers.

Virtual and Augmented Reality, or even Vegas magic acts, provide excellent visual analogues. Whether one looks at stereoscopic 3-D systems like Occulus Rift, or the much-ballyhooed ‘biomimetics’ of Magic Leap, or the illusions of David Copperfield, the idea is to cue visual environments that do not exist as effectively and as economically as possible. Goerztal and Levesque and others can keep pounding at the gates of general cognition (which may exist, who knows), but research like that of the late Clifford Nass is laying bare the landscape of cues comprising human social cognition, and given the relative resources required, it seems all but inevitable that the ‘taking to be’ approach, designing AIs focused not so much on being a genuine agent (whatever that is) as cuing the cognition of one, will sweep the field. Why build Disney World when you can project it? Developers will focus on the illusion, which they will refine and refine until the show becomes (Turing?) indistinguishable from the real thing—from the standpoint of consumers.

The differences being, 1) that the illusion will be perspectivally robust (we will have no easy way of seeing through it); and 2) the illusion will be a sociocognitive one. As AI colonizes more and more facets of our lives, our sociocognitive intuitions will become increasingly unreliable. This prediction, I think, is every bit as reliable as the prediction that the world’s ecosystems will be increasingly disrupted as human activity colonizes more and more of the world. Human social cognition turns access to cues into behaviour solving otherwise intractable biological brains—this is a fact. Algorithms are set to flood this space, to begin cuing social cognition to solve biological brains in the absence of any biological brains. Neil Lawrence likens the consequences to the creation of ‘System Zero,’ an artificial substratum for the System 1 (automatic, unconscious) and System 2 (deliberate, conscious) organization of human cognition. He writes:

“System Zero will come to understand us so fully because we expose to it our inner most thoughts and whims. System Zero will exploit massive interconnection. System Zero will be data rich. And just like an elephant, System Zero will never forget.”

Even as we continue attempting to solve it with systems we evolved to solve one another—a task which is going to remain as difficult as it always has, and will likely grow less attractive as fantasy surrogates become increasingly available. Talk about Systems over Subjects! The ecology of human meaning, the shared background allowing us to resolve conflict and to trust, will be progressively exploited and degraded—like every other ancestral ecology on this planet. When I wax grandiloquent (I am a crazy fantasy writer after all), I call this the semantic apocalypse.

I see no way out. Everyone thinks otherwise, but only because the way that human cognition neglects cognitive ecology generates the illusion of unlimited, unconstrained cognitive capacity. And this, I think, is precisely the illusion informing Ito and Howe’s theory of human nature…

Speaking of which, as I said, I found myself wondering what this theory might be as I read the book. I understood I wasn’t the target audience of the book, so I didn’t see its absence as a failing so much as unfortunate for readers like me, always angling for the hard questions. And so it niggled and niggled, until finally, I reached the last paragraph of the last page and encountered this:

“Human beings are fundamentally adaptable. We created a society that was more focussed on our productivity than our adaptability. These principles will help you prepare to be flexible and able to learn the new roles and to discard them when they don’t work anymore. If society can survive the initial whiplash when we trade our running shoes for a supersonic jet, we may yet find that the view from the jet is just what we’ve been looking for.” 250

This first claim, uplifting as it sounds, is simply not true. Human beings, considered individually or collectively, are not capable of adapting to any circumstance. Intuitions systematically misfire all the time. I appreciate how believing as much balms the conscience of those in the innovation game, but it is simply not true. And how could it be, when it entails that humans somehow transcend ecology, which is a far different claim than saying humans, relative to other organisms, are capable of spanning a wide-variety of ecologies. So long as human cognition is heuristic it depends on environmental invariances, like everything else biological. Humans are not capable of transcending system, which is precisely why we need to think the human in systematic terms, and to look at the impact of AI ecologically.

What makes Whiplash such a valuable book (aside from the entertainment factor) is that it is ecologically savvy. Ito and Howe’s dominant metaphor is that of adaptation and ecology. The old business habitat, they argue, has collapsed, leaving old business animals in the ecological lurch. The solution they offer is heuristic, a set of maxims meant to transform (at a sub-ideological level no less!) old business animals into newer, more adaptable ones. The way to solve the problem of innovation uncertainty is to contribute to that problem in the right way—be more innovative. But they fail to consider the ecological dimensions of this imperative, to see how feeding acceleration amounts to the inevitable destruction of cognitive ecologies, how the old meaning habitat is already collapsing, leaving old meaning animals in the ecological lurch, grasping for lies because those, at least, they can recognize.

They fail to see how their local survival guide likely doubles as a global suicide manual.

The meta-heuristics they offer, the new guiding mythologies, are meant to encapsulate the practical bases of evolvability itself… They’re teaching ferns how to grow flowers.


PS: The Big Picture

“In the past twenty-five years,” Ito and Howe write, “we have moved from a world dominated by simple systems to a world beset and baffled by complex systems” (246). This claim caught my attention because it is both true and untrue, depending how you look at it. We are pretty much the most complicated thing we know of in the universe, so it’s certainly not the case that we’ve ever dwelt in a world dominated by simple systems. What Ito and Howe are referring to, of course, is our tools. We are moving from a world dominated by simple tools to a world beset and baffled by complex ones. Since these tools facilitate tool-making, we find the great ratchet that lifted us out of the hominid fog clicking faster and faster and faster.

One of these ‘simple tools’ is what we call a ‘company’ or ‘business,’ an institution itself turning on the systematic application of simple tools, ones that intrinsically value authority over emergence, push over pull, maps over compasses, safety over risk, compliance over disobedience, theory over practice, ability over diversity, strength over resilience, and objects over systems. In the same way the simplicity of our physical implements limited the damage they could do to our physical ecologies, the simplicity of our cognitive tools limited the damage they could do to our cognitive ecology. It’s important to understand that the simplicity of these tools is what underwrites the stability of the underlying cognitive ecology. As the growing complexity and power of our physical tools intensified the damage done to our physical ecologies, the growing complexity and power of our cognitive tools is intensifying the damage done to our cognitive ecologies.

Now, two things. First, this analogy suggests that not all is hopeless, that the same way we can use the complexity and power of our physical tools to manage and prevent the destruction of our physical environment, we should be able to use the complexity and power of our cognitive tools to do the same. I concede the possibility, but I think the illusion of noocentrism (the cognitive version of geocentrism) is simply too profound. I think people will endlessly insist on the freedom to concede their autonomy. System Zero will succeed because it will pander ever so much better than a cranky old philosopher could ever hope to.

Second, notice how this analogy transforms the nature of the problem confronting that old animal, business, in the light of radical ecological change. Ancestral human cognitive ecology possessed a shallow present and a deep future. For all his ignorance, a yeoman chewing his calluses in the field five hundred years ago could predict that his son would possess a life very much resembling his own. All the obsolete items that Ito and Howe consider are artifacts of a shallow present. When the world is a black box, when you have no institutions like science bent on the systematic exploration of solution space, the solutions happened upon are generally lucky ones. You hold onto the tools you trust, because it’s all guesswork otherwise and the consequences are terminal. Authority, Push, Compliance, and so on are all heuristics in their own right, all ways of dealing with supercomplicated systems (bunches of humans), but selected for cognitive ecologies where solutions were both precious and abiding.

Oh, how things have changed. Ambient information sensitivity, the ability to draw on everything from internet search engines, to Big Data, to scientific knowledge more generally, means that businesses have what I referred to earlier as a deep present, a vast amount of information and capacity to utilize in problem solving. This allows them to solve systems as systems (the way science does) and abandon the limitations of not only object thinking, but (and this is the creepy part) subject thinking as well. It allows them to correct for faulty path-dependencies by distributing problem-solving among a diverse array of individuals. It allows them to rationalize other resources as well, to pull what they need when they need it rather than pushing warehoused resources.

Growing ambient information sensitivity means growing problem-solving economy—the problem is that this economy means accelerating cognitive ecological transformation. The cheaper optimization becomes, the more transient it becomes, simply because each and every new optimization transforms, in ways large or small but generally unpredictable, the ecology (the network of correlations) prior heuristic optimizations require to be effective. Call this the Optimization Spiral.

This is the process Ito and Howe are urging the business world to climb aboard, to become what might be called meta-ecological institutions, entities designed in the first instance, not to build cars or to mediate social relations or to find information on the web, but to evolve. As an institutionalized bundle of heuristics, a business’s ability to climb the Optimization Spiral, to survive accelerating ecological change, turns on its ability to relinquish the old while continually mimicking, tinkering, and birthing with the new. Thus the value of disobedience and resilience and practical learning: what Ito and Howe are advocating is more akin to the Precambrian Explosion or the rise of Angiosperms than simply surviving extinction. The meta-heuristics they offer, the new guiding mythologies, are meant to encapsulate the practical bases of evolvability itself… They’re teaching ferns how to grow flowers.

And stepping back to take the systems view they advocate, one cannot but feel an admixture of awe and terror, and wonder if they aren’t sketching the blueprint for an entirely unfathomable order of life, something simultaneously corporate and corporeal.

Real Systems

by rsbakker


Now I’ve never had any mentors; my path has been too idiosyncratic, for the better, since I think it’s the lack of institutional constraints that has allowed me to experiment the way I have. But if I were pressed to name any spiritual mentor, Daniel Dennett would be the first name to cross my lips—without the least hesitation. Nevertheless, I see the theoretical jewel of his project, the intentional stance, as the last gasp of what will one day, I think, count as one of humanity’s great confusions… and perhaps the final one to succumb to science.

A great many disagree, of course, and because I’ve been told so many times to go back to “Real Patterns” to discover the error of my ways, I’ve decided I would use it to make my critical case.

Defenders of Dennett (including Dennett himself) are so quick to cite “Real Patterns,” I think, because it represents his most sustained attempt to situate his position relative to his fellow philosophical travelers. At issue is the reality of ‘intentional states,’ and how the traditional insistence on some clear cut binary answer to this question—real/unreal—radically underestimates the ontological complexity charactering both everyday life and the sciences. What he proposes is “an intermediate doctrine” (29), a way of understanding intentional states as real patterns.

I have claimed that beliefs are best considered to be abstract objects rather like centers of gravity. Smith considers centers of gravity to be useful fictions while Dretske considers them to be useful (and hence?) real abstractions, and each takes his view to constitute a criticism of my position. The optimistic assessment of these opposite criticisms is that they cancel each other out; my analogy must have hit the nail on the head. The pessimistic assessment is that more needs to be said to convince philosophers that a mild and intermediate sort of realism is a positively attractive position, and not just the desperate dodge of ontological responsibility it has sometimes been taken to be. I have just such a case to present, a generalization and extension of my earlier attempts, via the concept of a pattern. My aim on this occasion is not so much to prove that my intermediate doctrine about the reality of psychologcal states is right, but just that it is quite possibly right, because a parallel doctrine is demonstrably right about some simpler cases. 29

So what does he mean by ‘real patterns’? Dennett begins by considering a diagram with six rows of five black boxes each characterized by varying degrees of noise, so extreme in some cases as completely obscure the boxes. He then, following the grain of his characteristic genius, provides a battery of different ways these series might find themselves used.

This crass way of putting things-in terms of betting and getting rich-is simply a vivid way of drawing attention to a real, and far from crass, trade-off that is ubiquitous in nature, and hence in folk psychology. Would we prefer an extremely compact pattern description with a high noise ratio or a less compact pattern description with a lower noise ratio? Our decision may depend on how swiftly and reliably we can discern the simple pattern, how dangerous errors are, how much of our resources we can afford to allocate to detection and calculation. These “design decisions” are typically not left to us to make by individual and deliberate choices; they are incorporated into the design of our sense organs by genetic evolution, and into our culture by cultural evolution. The product of this design evolution process is what Wilfrid Sellars calls our manifest image, and it is composed of folk physics, folk psychology, and the other pattern-making perspectives we have on the buzzing blooming confusion that bombards us with data. The ontology generated by the manifest image has thus a deeply pragmatic source. 36

The moral is straightforward: the kinds of patterns that data sets yield are both perspectival and pragmatic. In each case, the pattern recognized is quite real, but bound upon some potentially idiosyncratic perspective possessing some potentially idiosyncratic needs.

He then takes this moral to Conway’s Game of Life, a computer program where cells in a grid are switched on or off in successive turns depending on the number of adjacent cells switched on. The marvelous thing about this program lies in the kinds of dynamic complexities arising from this simple template and single rule, subsystems persisting from turn to turn, encountering other subsystems with predictable results. Despite the determinism of this system, patterns emerge that only the design stance seems to adequately capture, a level possessing “it’s own language, a transparent foreshortening of the tedious descriptions one could give at the physical level” (39).

For Dennett, the fact that one can successfully predict via the design stance clearly demonstrates that it’s picking out real patterns somehow. He asks us to imagine transforming the Game into a supersystem played out on a screen miles wide and using the patterns picked out to design a Turing Machine playing chess against itself. Here, Dennett argues, the determinacy of the microphysical picture is either intractable or impracticable, yet we need only take up a chess stance or a computational stance to make, from a naive perspective, stunning predictions as to what will happen next.

And this is of course as true of life life as it is the Game of Life: “Predicting that someone will duck if you throw a brick at him is easy from the folk-psychological stance; it is and will always be intractable if you have to trace the photons from brick to eyeball, the neurotransmitters from optic nerve to motor nerve, and so forth” (42). His supersized Game of Life, in other words, makes plain the power and the limitations of heuristic cognition.

This brings him to his stated aim of clarifying his position vis a vis his confreres and Fodor. As he points out, everyone agrees there’s some kind of underlying “order which is there,” as Anscombe puts it in Intention. The million dollar question, of course, is what this order amounts to:

Fodor and others have claimed that an interior language of thought is the best explanation of the hard edges visible in “propositional attitude psychology.” Churchland and I have offered an alternative explanation of these edges… The process that produces the data of folk psychology, we claim, is one in which the multidimensional complexities of the underlying processes are projected through linguistic behavior, which creates an appearance of definiteness and precision, thanks to the discreteness of words. 44-45

So for traditional realists, like Fodor, the structure beliefs evince in reflection and discourse expresses the structure beliefs must possess in the head. For Dennett, on the other hand, the structure beliefs evince in reflection and discourse expresses, among other things, the structure of reflection and discourse. How could it be otherwise, he asks, given the ‘stupendous scale of compression’ (42) involved?

As Haugeland points out in “Pattern and Being,” this saddles Dennett’s account of patterns with a pretty significant ambiguity: if the patterns characteristic of intentional states express the structure of reflection and discourse, then the ‘order which is there’ must be here as well. Of course, this much is implicit in Dennett’s preamble: the salience of certain patterns depends on the perspective we possess on them. But even though this implicit ‘here-there holism’ becomes all but explicit when Dennett turns to Radical Translation and the distinction between his and Davidson’s views, his emphasis nevertheless remains on the order out there. As he writes:

Davidson and I both like Churchland’s alternative idea of propositional-attitude statements as indirect “measurements” of a reality diffused in the behavioral dispositions of the brain (and body). We think beliefs are quite real enough to call real just so long as belief talk measures these complex behavior-disposing organs as predictively as it does. 45-46

Rhetorically (even diagrammatically if one takes Dennett’s illustrations into account), the emphasis is on the order there, while here is merely implied as a kind of enabling condition. Call this the ‘epistemic-ontological ambiguity’ (EOA). On the one hand, it seems to make eminent sense to speak of patterns visible only from certain perspectives and to construe them as something there, independent of any perspective we might take on them. But on the other hand, it seems to make jolly good sense to speak of patterns visible only from certain perspectives and to construe them as around here, as something entirely dependent on the perspective we find ourselves taking. Because of this, it seems pretty fair to ask Dennett which kind of pattern he has in mind here. To speak of beliefs as dispositions diffused in the brain seems to pretty clearly imply the first. To speak of beliefs as low dimensional, communicative projections, on the other hand, seems to clearly imply the latter.

Why this ambiguity? Do the patterns underwriting belief obtain in individual believers, dispositionally diffused as he says, or do they obtain in the communicative conjunction of witnesses and believers? Dennett promised to give us ‘parallel examples’ warranting his ‘intermediate realism,’ but by simply asking the whereabouts of the patterns, whether we will find them primarily out there as opposed to around here, we quickly realize his examples merely recapitulate the issue they were supposed to resolve.



Welcome to crash space. If I’m right then you presently find yourself strolling through a cognitive illusion generated by the application of heuristic capacities outside their effective problem ecology.

Think of how curious the EOA is. The familiarity of it should be nothing short of gobsmacking: here, once again we find ourselves stymied by the same old dichotomies: here versus there, inside versus outside, knowing versus known. Here, once again we find ourselves trapped in the orbit of the great blindspot that still, after thousands of years, stumps the wise of the world.

What the hell could be going on?

Think of the challenge facing our ancestors attempting cognize their environmental relationships for the purposes of communication and deliberate problem-solving. The industrial scale of our ongoing attempt to understand as much demonstrates the intractability of that relationship. Apart from our brute causal interactions, our ability to cognize our cognitive relationships is source insensitive through and through. When a brick is thrown at us, “the photons from brick to eyeball, the neurotransmitters from optic nerve to motor nerve, and so forth” (42) all go without saying. In other words, the whole system enabling cognition of the brick throwing is neglected, and only information relevant to ancestral problem-solving—in this case, brick throwing—finds its way to conscious broadcast.

In ancestral cognitive ecologies, our high-dimensional (physical) continuity with nature mattered as much as it matters now, but it quite simply did not exist for them. They belonged to any number of natural circuits across any number of scales, and all they had to go on was the information that mattered (disposed them to repeat and optimize behaviours) given the resources they possessed. Just as Dennett argues, human cognition is heuristic through and through. We have no way of cognizing our position within any number of the superordinate systems science has revealed in nature, so we have to make do with hacks, subsystems allowing us to communicate and troubleshoot our relation to the environment while remaining almost entirely blind to it. About talk belongs to just such a subsystem, a kluge communicating and troubleshooting our relation to our environments absent cognition of our position in larger systems. As I like to say, we’re natural in such a way as to be incapable of cognizing ourselves as natural.

About talk facilitates cognition and communication of our worldly relation absent any access to the physical details of that relation. And as it turns out, we are that occluded relation’s most complicated component—we are the primary thing neglected in applications of about talk. As the thing most neglected, we are the thing most presumed, the invariant background guaranteeing the reliability of about talk (this is why homuncular arguments are so empty). This combination of cognitive insensitivity to and functional dependence upon the machinations of cognition (what I sometimes refer to as medial neglect) suggests that about talk would be ideally suited to communicating and troubleshooting functionally independent systems, processes generally insensitive to our attempts to cognize them. This is because the details of cognition make no difference to the details cognized: the automatic distinction about talk poses between cognizing system and the system cognized poses no impediment to understanding functionally independent systems. As a result, we should expect about talk to be relatively unproblematic when it comes to communicating and troubleshooting things ‘out there.’

Conversely, we should expect about talk to generate problems when it comes to communicating and troubleshooting functionally dependent systems, processes somehow sensitive to our attempts to cognize them. Consider ‘observer effects,’ the problem researchers themselves pose when their presence or their tools/techniques interfere with the process they are attempting to study. Given medial neglect, the researchers themselves always constitute a black box. In the case of systems functionally sensitive to the activity of cognition, as is often the case in psychology and particle physics, understanding the system requires we somehow obviate our impact on the system. As the interactive, behavioural components of cognition show, we are in fact quite good (though far from perfect) at inserting and subtracting our interventions in processes. But since we remain a black box, since our position in the superordinate systems formed by our investigations remains occluded, our inability to extricate ourselves, to gerrymander functional independence, say, undermines cognition.

Even if we necessarily neglect our positions in superordinate systems, we need some way of managing the resulting vulnerabilities, to appreciate that patterns may be artifacts of our position. This suggests one reason, at least, for the affinity of mechanical cognition and ‘reality.’ The more our black box functions impact the system to be cognized, the less cognizable that system becomes in source sensitive terms. We become an inescapable source of noise. Thus our intuitive appreciation of the need for ‘perspective,’ to ‘rise above the fray’: The degree to which a cognitive mode preserves (via gerrymandering if not outright passivity) the functional independence of a system is the degree to which that cognitive mode enables reliable source sensitive cognition is the degree to which about talk can be effectively applied.

The deeper our entanglements, on the other hand, the more we need to rely on source insensitive modes of cognition to cognize target systems. Even if our impact renders the isolation of source signals impossible, our entanglement remains nonetheless systematic, meaning that any number of cues correlated in any number of ways to the target system can be isolated (which is really all ‘radical translation’ amounts to). Given that metacognition is functionally entangled by definition, it becomes easy to see why the theoretical question of cognition causes about talk to crash the spectacular ways it does: our ability to neglect the machinations of cognition (the ‘order which is here’) is a boundary condition for the effective application of ‘orders which are there’—or seeing things as real. Systems adapted to work around the intractability of our cognitive nature find themselves compulsively applied to the problem of our cognitive nature. We end up creating a bestiary of sourceless things, things that, thanks to the misapplication of the aboutness heuristic, have to belong to some ‘order out there,’ and yet cannot be sourced like anything else out there… as if they were unreal.

The question of reality cues the application of about talk, our source insensitive means of communicating and troubleshooting our cognitive relation to the world. For our ancient ancestors, who lacked the means to distinguish between source sensitive and source insensitive modes of cognition, asking, ‘Are beliefs real?’ would have sounded insane. HNT, in fact, provides a straightforward explanation for what might be called our ‘default dogmatism,’ our reflex for naive realism: not only do we lack any sensitivity to the mechanics of cognition, we lack any sensitivity to this insensitivity. This generates the persistent illusion of sufficiency, the assumption (regularly observed in different psychological phenomena) that the information provided is all the information there is.

Cognition of cognitive insufficiency always requires more resources, more information. Sufficiency is the default. This is what makes the novel application of some potentially ‘good trick,’ as Dennett would say, such tricky business. Consider philosophy. At some point, human culture acquired the trick of recruiting existing metacognitive capacities to explain the visible in terms of the invisible in unprecedented (theoretical) ways. Since those metacognitive capacities are radically heuristic, specialized consumers of select information, we can suppose retasking those capacities to solve novel problems—as philosophers do when they, for instance, ‘ponder the nature of knowledge’—would run afoul some pretty profound problems. Even if those specialized metacognitive consumers possessed the capacity to signal cognitive insufficiency, we can be certain the insufficiency flagged would be relative to some adaptive problem-ecology. Blind to the heuristic structure of cognition, the first philosophers took the sufficiency of their applications for granted, much as very many do now, despite the millennia of prior failure.

Philosophy inherited our cognitive innocence and transformed it, I would argue, into a morass of competing cognitive fantasies. But if it failed to grasp the heuristic nature of much cognition, it did allow, as if by delayed exposure, a wide variety of distinctions to blacken the photographic plate of philosophical reflection—that between is and ought, fact and value, among them. The question, ‘Are beliefs real?’ became more a bona fide challenge than a declaration of insanity. Given insensitivity to the source insensitive nature of belief talk, however, the nature of the problem entirely escaped them. Since the question of reality cues the application of about talk, source insensitive modes of cognition struck them as the only game in town. Merely posing the question springs the trap (for as Dennett says, selecting cues is “typically not left to us to make by individual and deliberate choices” (36)). And so they found themselves attempting to solve the hidden nature of cognition via the application of devices adapted to ignore hidden natures.

Dennett runs into the epistemic-ontological ambiguity because the question of the reality of intentional states cues the about heuristic out of school, cedes the debate to systems dedicated to gerrymandering solutions absent high-dimensional information regarding our cognitive predicament—our position within superordinate systems. Either beliefs are out there, real, or they’re in here, merely, an enabling figment of some kind. And as it turns out, IST is entirely amenable to this misapplication, in that ‘taking the intentional stance’ involves cuing the about heuristic, thus neglecting our high-dimensional cognitive predicament. On Dennett’s view, recall, an intentional system is any system that can be predicted/explained/manipulated via the intentional stance. Though the hidden patterns can only be recognized from the proper perspective, they are there nonetheless, enough, Dennett thinks, to concede them reality as intentional systems.

Heuristic Neglect Theory allows us to see how this amounts to mistaking a CPU for a PC. On HNT, the trick is to never let the superordinate systems enabling and necessitating intentional cognition out of view. Recall the example of the gaze heuristic from my prior post, how fielders essentially insert—functionally entangle—themselves into the pop fly system to let the ball itself guide them in. The same applies to beliefs. When your tech repairs your computer, you have no access to her personal history, the way thousands of hours have knapped her trouble-shooting capacities, and even less access to her evolutionary history, the way continual exposure to problematic environments has sculpted her biological problem-solving capacities. You have no access, in other words, to the vast systems of quite natural relata enabling her repair. The source sensitive story is unavailable, so you call her ‘knowledgeable’ instead; you presume she possesses something—a fetish, in effect—possessing the sourceless efficacy explaining her almost miraculous ability to make your PC run: a mass of true beliefs (representations), regarding personal computer repair. You opt for a source insensitive means that correlates with her capacities well enough to neglect the high-dimensional facts—the natural and personal histories—underwriting her ability.

So then where does the ‘real pattern’ gainsaying the reality of belief lie? The realist would say in the tech herself. This is certainly what our (heuristic) intuitions tell us in the first instance. But as we saw above, squaring sourceless entities in a world where most everything has a source is no easy task. The instrumentalist would say in your practices. This certainly lets us explain away some of the peculiarities crashing our realist intuitions, but at the cost of other, equally perplexing problems (this is crash space, after all). As one might expect, substituting the use heuristic for the about heuristic merely passes the hot potato of source insensitivity. ‘Pragmatic functions’ are no less difficult to square with the high-dimensional than beliefs.

But it should be clear by now that the simple act of pairing beliefs with patterns amounts to jumping the same ancient shark. The question, ‘Are beliefs real?’ was a no-brainer for our preliterate ancestors simply because they lived in a seamless shallow information cognitive ecology. Outside their local physics, the sources of things eluded them altogether. ‘Of course beliefs are real!’ The question was a challenge for our philosophical ancestors because they lived in a fractured shallow information ecology. They could see enough between the cracks to appreciate the potential extent and troubling implications of mechanical cognition, it’s penchant to crash our shallow (ancestral) intuitions. ‘It has to be real!’

With Dennett, entire expanses of our shallow information ecology have been laid low and we get, ‘It’s as real as it needs to be.’ He understands the power of the about heuristic, how ‘order out there’ thinking effects any number of communicative solutions—thus his rebuttal of Rorty. He understands, likewise, the power of the use heuristic, how ‘order around here’ thinking effects any number of communicative solutions—thus his rebuttal of Fodor. And most importantly, he understands the error of assuming the universal applicability of either. And so he concludes:

Now, once again, is the view I am defending here a sort of instrumentalism or a sort of realism? I think that the view itself is clearer than either of the labels, so I shall leave that question to anyone who stills find [sic] illumination in them. 51

What he doesn’t understand is how it all fits together—and how could he, when IST strands him with an intentional theorization of intentional cognition, a homuncular or black box understanding of our contemporary cognitive predicament? This is why “Real Patterns” both begins and ends with EOA, why we are no closer to understanding why such ambiguity obtains at all. How are we supposed to understand how his position falls between the ‘ontological dichotomy’ of realism and instrumentalism when we have no account of this dichotomy in the first place? Why the peculiar ‘bi-stable’ structure? Why the incompatibility between them? How can the same subject matter evince both? Why does each seem to inferentially beg the other?



The fact is, Dennett was entirely right to eschew outright realism or outright instrumentalism. This hunch of his, like so many others, was downright prescient. But the intentional stance only allows him to swap between perspectives. As a one-time adherent I know first-hand the theoretical versatility IST provides, but the problem is that explanation is what is required here.

HNT argues that simply interrogating the high-dimensional reality of belief, the degree to which it exists out there, covers over the very real system—the cognitive ecology—explaining the nature of belief talk. Once again, our ancestors needed some way of communicating their cognitive relations absent source-sensitive information regarding those relations. The homunculus is a black box precisely because it cannot source its own functions, merely track their consequences. The peculiar ‘here dim’ versus ‘there bright’ character of naive ontological or dogmatic cognition is a function of medial neglect, our gross insensitivity to the structure and dynamics of our cognitive capacities. Epistemic or instrumental cognition comes with learning from the untoward consequences of naive ontological cognition—the inevitable breakdowns. Emerging from our ancestral, shallow information ecologies, the world was an ‘order there’ world simply because humanity lacked the ability to discriminate the impact of ‘around here.’ The discrimination of cognitive complexity begets intuitions of cognitive activity, undermines our default ‘out there’ intuitions. But since ‘order there’ is the default and ‘around here’ the cognitive achievement, we find ourselves in the peculiar position of apparently presuming ‘order there’ when making ‘around here’ claims. Since ‘order there’ intuitions remain effective when applied in their adaptive problem-ecologies, we find speculation splitting along ‘realist’ versus ‘anti-realist’ lines. Because no one has any inkling of any of this, we find ourselves flipping back and forth between these poles, taking versions of the same obvious steps to trod the same ancient circles. Every application is occluded, and so ‘transparent,’ as well as an activity possessing consequences.

Thus EOA… as well as an endless parade of philosophical chimera.

Isn’t this the real mystery of “Real Patterns,” the question of how and why philosophers find themselves trapped on this rickety old teeter-totter? “It is amusing to note,” Dennett writes, “that my analogizing beliefs to centers of gravity has been attacked from both sides of the ontological dichotomy, by philosophers who think it is simply obvious that centers of gravity are useful fictions, and by philosophers who think it is simply obvious that centers of gravity are perfectly real” (27). Well, perhaps not so amusing: Short of solving this mystery, Dennett has no way of finding the magic middle he seeks in this article—the middle of what? IST merely provides him with the means to recapitulate EOA and gesture to the possibility of some middle, a way to conceive all these issues that doesn’t deliver us to more of the same. His instincts, I think, were on the money, but his theoretical resources could not take him where he wanted to go, which is why, from the standpoint of his critics, he just seems to want to have it both ways.

On HNT we can see, quite clearly, I think, the problem with the question, ‘Are beliefs real?’ absent an adequate account of the relevant cognitive ecology. The bitter pill lies in understanding that the application conditions of ‘real’ have real limits. Dennett provides examples where those application conditions pretty clearly seem to obtain, then suggests more than argues that these examples are ‘parallel’ in all the structurally relevant respects to the situation with belief. But to distinguish his brand from Fodor’s ‘industrial strength’ realism, he has no choice but to ‘go instrumental’ in some respect, thus exposing the ambiguity falling out of IST.

It’s safe to say belief talk is real. It seems safe to say that beliefs are ‘real enough’ for the purposes of practical problem-solving—that is, for shallow (or source insensitive) cognitive ecologies. But it also seems safe to say that beliefs are not real at all when it comes to solving high-dimensional cognitive ecologies. The degree to which scientific inquiry is committed to finding the deepest (as opposed to the most expedient) account, should be the degree to which it views belief talk as components of real systems and views ‘belief’ as a source insensitive posit, a way to communicate and troubleshoot both oneself and one’s fellows.

This is crash space, so I appreciate the kinds of counter-intuitiveness involved in this view I’m advancing. But since tramping intuitive tracks has hitherto only served to entrench our controversies and confusions, we have good reason to choose explanatory power over intuitive appeal. We should expect synthesis in the cognitive sciences will prove every bit as alienating to traditional presumption as it was in biology. There’s more than a little conceit involved in thinking we had any special inside track on our own nature. In fact, it would be a miracle if humanity had not found itself in some version of this very dilemma. Given only source insensitive means to troubleshoot cognition, to understand ourselves and each other, we were all but doomed to be stumped by the flood of source sensitive cognition unleashed by science. (In fact, given some degree of interstellar evolutionary convergence, I think one can wager that extraterrestrial intelligences will have suffered their own source insensitive versus source sensitive cognitive crash spaces. See my, “On Alien Philosophy,” The Journal of Consciousness Studies, (forthcoming))

IST brings us to the deflationary limit of intentional philosophy. HNT offers a way to ratchet ourselves beyond, a form of critical eliminativism that can actually explain, as opposed to simply dispute, the traditional claims of intentionality. Dennett, of course, reserves his final criticism for eliminativism, perhaps because so many critics see it as the upshot of his interpretivism. He acknowledges the possibility that “that neuroscience will eventually-perhaps even soon-discover a pattern that is so clearly superior to the noisy pattern of folk psychology that everyone will readily abandon the former for the latter (50),” but he thinks it unlikely:

For it is not enough for Churchland to suppose that in principle, neuroscientific levels of description will explain more of the variance, predict more of the “noise” that bedevils higher levels. This is, of course, bound to be true in the limit-if we descend all the way to the neurophysiological “bit map.” But as we have seen, the trade-off between ease of use and immunity from error for such a cumbersome system may make it profoundly unattractive. If the “pattern” is scarcely an improvement over the bit map, talk of eliminative materialism will fall on deaf ears-just as it does when radical eliminativists urge us to abandon our ontological commitments to tables and chairs. A truly general-purpose, robust system of pattern description more valuable than the intentional stance is not an impossibility, but anyone who wants to bet on it might care to talk to me about the odds they will take. 51

The elimination of theoretical intentional idiom requires, Dennett correctly points out, some other kind of idiom. Given the operationalization of intentional idioms across a wide variety of research contexts, they are not about to be abandoned anytime soon, and not at all if the eliminativist has nothing to offer in their stead. The challenge faced by the eliminativist, Dennett recognizes, is primarily abductive. If you want to race at psychological tracks, you either enter intentional horses or something that can run as fast or faster. He thinks this unlikely because he thinks no causally consilient (source sensitive) theory can hope to rival the combination of power and generality provided by the intentional stance. Why might this be? Here he alludes to ‘levels,’ suggest that any causally consilient account would remain trapped at the microphysical level, and so remain hopelessly cumbersome. But elsewhere, as in his discussion of ‘creeping depersonalization’ in “Mechanism and Responsibility,” he readily acknowledges our ability to treat with one another as machines.

And again, we see how the limited resources of IST have backed him into a philosophical corner—and a traditional one at that. On HNT, his claim amounts to saying that no source sensitive theory can hope to supplant the bundle of source insensitive modes comprising intentional cognition. On HNT, in other words, we already find ourselves on the ‘level’ of intentional explanation, already find ourselves with a theory possessing the combination of power and generality required to eliminate a particle of intentional theorization: namely, the intentional stance. A way to depersonalize cognitive science.

Because IST primarily provides a versatile way to deploy and manage intentionality in theoretical contexts rather than any understanding of its nature, the disanalogy between ‘center of gravity’ and ‘beliefs’ remains invisible. In each case you seem to have an entity that resists any clear relation to the order which is there, and yet finds itself regularly and usefully employed in legitimate scientific contexts. Our brains are basically short-cut machines, so it should come as no surprise that we find heuristics everywhere, in perception as much as cognition (insofar as they are distinct). It also should come as no surprise that they comprise a bestiary, as with most all things biological. Dennett is comparing heuristic apples and oranges, here. Centers of gravity are easily anchored to the order which is there because they economize otherwise available information. They can be sourced. Such is not the case with beliefs, belonging as they do to a system gerrymandering for the want of information.

So what is the ultimate picture offered here? What could reality amount to outside our heuristic regimes? Hard to think, as it damn well should be. Our species’ history posed no evolutionary challenges requiring the ability to intuitively grasp the facts of our cognitive predicament. It gave us a lot of idiosyncratic tools to solve high impact practical problems, and as a result, Homo sapiens fell through the sieve in such a way as to be dumbfounded when it began experimenting in earnest with its interrogative capacities. We stumbled across a good number of tools along the way, to be certain, but we remain just as profoundly stumped about ourselves. On HNT, the ‘big picture view’ is crash space, in ways perhaps similar to the subatomic, a domain where our biologically parochial capacities actually interfere with our ability to understand. But it offers a way of understanding the structure and dynamics of intentional cognition in source sensitive terms, and in so doing, explains why crashing our ancestral cognitive modes was inevitable. Just consider the way ‘outside heuristic regimes’ suggests something ‘noumenal,’ some uber-reality lost at the instant of transcendental application. The degree to which this answer strikes you as natural or ‘obvious’ is the degree you have been conditioned to apply that very regime out of school. With HNT we can demand those who want to stuff us into this or that intellectual Klein bottles define their application conditions, convince us this isn’t just more crash space mischief.

It’s trivial to say some information isn’t available, so why not leave well enough alone? Perhaps the time has come to abandon the old, granular dichotomies and speak in terms of dimensions of information available and cognitive capacities possessed. Imagine that

Moving on.