- Language use and cognition are different things. Cognition is vastly older than language: it is based on discriminating high-level features and doing things with those discriminations that promote environmental success. Arguably, many types of single-celled organisms are already harboring versions of this. Language, in contrast, evolved in a hominin or proto-hominin societal setting to transmit encodings of the upshots of discrimination and the stages of cognition to other individuals in the community, in order to support what we might fairly call 'chain of thought' evaluation and refinement in a social setting. Perhaps only in a final stage did chain-of-thought evaluation start getting used re-entrantly in individual psychologies to review and hopefully improve on the individual's own thought processes.
- It's hardly surprising that discrimination and purely discrimination-based, alinguistic cognition in different humans should involve similar if not contextually identical feature discriminations. That's because it's implicit in the ontogeny of the individual brain, which in turn reflects the phylogeny of the species, which evolved having to repeatedly solve the same environmentally imposed categories of problems. This baseline similarity is a necessary condition for language being possible at all.
- Point 2 is necessary for language, but not sufficient; it compels the question of why, modulo some close-but-no-cigar communication systems in other species, we don't see grammatical language happening anywhere else in nature. Here, I strongly suspect the key lies in those genetic bottlenecks that we now know happened early in our history, one of which we strongly suspect came close to resulting in the extinction of our species. In short, the explanation is that we're considerably inbred, with the result that our respective neural ontogenies, architectures, and functions aren't merely similar but more nearly identical than obtains for any other vertebrate. This gives the background context on why the mutations that enabled language were actually supportive reproductive success. For any two individuals of our species, the odds were overwhelmingly in favor of there being enough points in common between their fundamental cognitive processes to support probably-approximately-correct exchange of encodings of those processes.
- Like many other signalling systems in biology, human language is extremely flexible, with extreme ‘canalization’ of function: i.e., there are many different ways of conveying essentially the same thing. In some degree, imprecision can be an advantage, when linguistic contextual understanding on the part of the audience can be counted on to elide any mismatch that might otherwise occur; indeed, it seems plausible that our extreme efficiency at internalizing the subsymbolic tropisms of ‘language games’ in the Wittgensteinian sense evolved precisely because of this consideration. The cautionary note is that there might be a scalability limit: one that is reached when science and especially data science become a part of culture. Communicating, and, especially, successfully applying scientific verification and falsification criteria demands rigid and extensionally static categories. However, the more exacting categorical definitions become, the greater the likelihood of misunderstanding if these are not explicitly and carefully articulated - and such precision in explication and the disputation it can give rise to carries its own information-theoretic cost. In short, it seems reasonable to think that it is precisely when we transition from the realm of unstructured to structured data that ‘probably approximately correct’ ceases to be adequate for linguistic communication and the broad semantic bandwidth that is the great strength of natural language in turn becomes a weakness. On a related note, it may be no coincidence that the technicality of the reasoning required appears to be a primary factor imposing significant limit on the performance of large language models.
Friday, June 13, 2025
Sunday, April 14, 2024
Alfred Korzybski famously said the map is not the territory; there are different ways in which the two may be conflated, and these are instructive.
The ontological absolutist mistakenly believes in the possibility of an ideal map such that every feature of the territory (and indeed, of territories beyond the borders of the map) can be predicted from it: if a feature is not in the map, it is not in the territory. Which is to say, they believe the reality of the territory consistently and completely determines the content of this map to the point where it would be an infallible guide, and they believe furthermore that such a map is discoverable. The totalitarian ontological absolutist (which is to say, the religious or ideological fanatic) believes that such a map is already in their possession, and the path to truth consists merely in forcing everyone to use it. The nominalist or moral relativist, conversely, believes that the reality of a map consistently and completely determines the territory: everybody’s map is equally real for them, invent a map and you reify a truth. And the totalitarian nihilist as a corollary believes that if you force everyone to use the map you have invented, you can create reality to suit yourself. From one perspective, the moral absolutist and the moral relativist might seem to be at opposing ends of an ideological spectrum; but both at bottom make the same fundamental mistake of assuming a unidirectional and omnivalent determinacy between the territory and its representation, and it is for this reason that, though in apparent opposition, they are so often indistinguishable in their practice and its outcome.
None of this is how representation actually works. Representation is an internalization of an external selection filter: it makes pre-emptive selections based on past experience of the external filter’s operation in order to game filtration. But nothing constrains the external filter’s operation such that future behavior will perfectly replicate the past - indeed, things are guaranteed not to be so, insofar as the agent deploying the internal filter is by definition a part of the external and thus reciprocally subject to selection based on consequences of their own action in a way which is not learnable from experience of the prior filter before the agent learned it, becoming thereby a novel element in the scenario. The universe, as J.B.S.Haldane said, is not only queerer than we suppose, but queerer than we can suppose, and reality will always, as a matter of necessity, be, not less, but much greater than what we are able to articulate.
Sunday, May 10, 2020
Thoughts on the present times.
Thursday, August 20, 2015
The Morbius Endgame
Saturday, March 01, 2014
The Importance of Selection History
In particular, in automated recognition systems in AI, it is fatal to ignore the semantic aspect even at the lowest levels - which is to say, fatal to conceive the relevant features as being somehow given in the data, as opposed it being a case of assimilating data to feature, where the process of assimilation is fundamentally the product of a genealogy. And while complete reconstruction of that genealogy may not be necessary (let alone recapitulation), it’s still safe to say that there can be no effective engineering of feature extraction in the long term without a fundamental understanding of the historical biological contingencies conditioning extraction.
Saturday, December 28, 2013
Of Transcription Factors, Experiential Inheritance, Second-Order Selection Of Proteins, and Saltatory Evolution
But, beyond facilitation of what I’m sorely tempted to call ‘soft Lamarckianism’, transcription factors appear to allow for an even more important plasticity - namely evolution stemming from selection pressures operating on a population comprised of different sections of the same genome. It appears that a given genome exhibits significant redundancy with respect to a number of different protein types: that is, different sections of the genome code for essentially the same protein. Critically, this is redundancy with variation: while the protein types coded for are similar and almost functionally equivalent, the codon sequence, and hence the amino acid sequence, are not quite the same, and there is every reason to believe that the environmental context could be such as to insure that one version of the functional enzyme does a slightly better job than another. Now consider the role transcription factors might play. Suppose that there is some feedback mechanism, however indirect, that leads from differential performance to upregulation or downregulation - more specifically, from suboptimal performance of the enzyme variant to downregulation of the section of the genome that coded for it, or from optimal performance to upregulation of the corresponding genome section. Now we have all the conditions in place for natural selection to occur, operating over a population of genome sections coding for functionally analogous but variant enzymes. Be it noted this is ‘second-order’ natural selection occurring on top of the first order selection that operates between whole genomes: when its consequences are viewed at the molecular scale, they will appear as a kind of adaptive learning; while at the macroscopic scale they may manifest as saltatory evolution.
It is also worth observing - as is generally true of cases involving adaptation via second-order selection - that selection pressures favor conservation of such a feedback mechanism, if there are any means by which it can be stumbled upon and conserved. That is, a mechanism that allows for optimization of enzyme function through selective adaptation constitutes a first-order advantage for the whole genome that possesses it; thus, genomes might be expected to evolve such a feature if a reliably replicable version can be hit on accidentally. Indeed, it may be the primary reason for the apparent redundancy of protein-encodings in the genome.
Sunday, July 21, 2013
Thinking About Information Value
Therefore we cannot speak of the absolute entropy, or the absolute information value of any state of a physical system independent of an organism or device which assesses that system’s internal and external boundaries. The entropy is a function of the system boundary and the space of possible microstates, and these in turn are conditional on what the detector is configured to detect. Use a different detector, and you draw the boundaries differently get a different entropy. Or take any arbitrary ‘slice’ of the plenum that you please; it is always possible to imagine some evolutionary history that would produce a detector able to extract useful work from, and therefore to detect, just that configuration within its context of internal and external boundaries.