Thursday, February 5, 2026
Saturday, January 24, 2026
Philosophy for an expanding universe
I have come to a set of conceptions that reimagine ‘philosophy’ in light of recent discoveries in cosmology. More particularly, moral philosophy – ethics – arguably the essence of philosophy – occupies a changed landscape in the wake of revolutionary science: relativity, uncertainty, and the gravitational redshift.
(1) Energy
A photon’s energy is equal to E = hv. Yet v depends on an observer’s proper time. Different observers at different gravitational potentials have different clock rates. Therefore, different observers assign different energies to the same photon. Conclusion: energy is not an absolute property of a system – only a relational one.
There is no single global energy; there is no observer-independent frequency of energy; there is no absolute time.
Conservation is global and relational, not intrinsic and local. There is no local energy density.
(2) Conservation
Noether’s theorem implicitly raises the issue whether conservation laws are an artifact of an observer’s perspective. Noether shows that conservation laws exist relative to symmetries – i.e., to symmetries an observer can legitimately identify. Different observers may not agree on what is being conserved. Conservation laws are relational, they are observer-indexed, they are globally defined but not locally intrinsic – they are relational, not absolute.
This is roughly like the case of temperature which is locally well defined but dependent on a frame of reference – tied to an observer’s motion – as we see in the Unruh effect. The upshot is something like contextual realism. Conservation laws are not fundamental truths about reality. They are results of the way in which reality lets itself be sliced up and ‘timed.’
(3) Steady state
Einstein and Noether precede Hubble. After Hubble – after the realization that Lambda is not theoretical – then what? The universe is expanding. The universe is not stationary. There is no global energy conservation in the universe. Energy conservation is not globally definable in an expanding universe.
Einstein anticipated this – introducing Lambda to avoid expansion – after Hubble he abandoned this motivation but not the placeholder – he never reinstated energy conservation for cosmology. He emphasized that the divergence of the stress energy tensor = 0. Gravitational redshift does not violate conservation at this level – the redshift does not transfer energy anywhere. Setting the stress energy tensor to zero is simply the residue of the idea of conservation after it becomes obvious that spacetime itself is dynamic.
No observer can define a globally conserved energy for the universe. Thus conservation laws are not fundamental principles imposed on spacetime – they are emergent consequences of spacetime symmetries that may or may not exist.
(4) Information
Is information more fundamental than energy? Looking at a relationship between the geometry of spacetime and the distribution of matter in the universe suggests something like this – Black holes suggest this – it from bit. Information is causality itself, plus entanglement. Thus if you ask the question, What is conserved in the universe? – if not energy – the answer seems to be: causality – minimally: the consistency of causal correlations – within causal limits – within a horizon.
Conclusion: information is horizon-relative.
Knowledge is no longer a representation of a total state of affairs but the maintenance of consistent correlations within a causal domain.
Knowledge can no longer be thought of converging towards a single unitary absolute description.
This settles the Einstein-Heisenberg debate about whether science is about what we can say about nature versus the idea that science is an attempt to closely track what is really there. Science – and philosophy – confront the same ontological limits.
Truth is indexed, partial, non-aggregable, but not at all arbitrary. Observer accounts must mesh where they overlap. Ontology is relational – entities exist not as substances but as nodes of interaction, properties are actualized relative to conditions, structure outruns substance.
Conclusion: reality does not consist in a single conserved totality but in the fact that no local perspective ever encounters contradiction within its causal reach.
Thus we go from a question like what is the ultimate inventory of reality? to a question like What constraints ensure mutual consistency among the partial descriptions that we have to deal with? The problem is coherence under limitation.
(5) Unitarity
Reasoning can be simplified to the directive not to multiply entities needlessly. The unity principle is primary. This is not simply Platonism or Idealism but simple mathematical economy. What happens to unity and the drive towards fundamental simplicity as we encounter the Hubble expansion? What happens to the Unified Field Theory/UFT? What happens to the whole idea of a Theory of Everything/TOE? What happens to cosmology?
Unitarity itself cannot be an absolute global principle in cosmology. It presupposes a global time parameter, a closed system, and a notion of the whole. None of these are possible. There is no preferred global time – there is no global observer – there is no single causal domain containing everything.
A horizon just on its own hides degrees of freedom, induces entropy, forces tracing over inaccessible states. Unitarity is preserved relative to an observer’s accessible code of existence; it is not preserved relative to the universe itself. Unitarity is extra theoretical. No observer ever witnesses the breakdown of their own physical laws – even though no single observer can survey the whole.
Conclusion: Reality is a patchwork of mutually consistent but incomplete descriptions.
(6) (After -------- ) Philosophy in the expanding universe
We get to a relational ontology – a relational epistemology – a relational cosmology – a new set of assumptions for philosophy. There are some big steps here.
The view from nowhere is no longer coherent. There is and can be no absolute description of reality. Global states of affairs themselves are no longer possible. There is no unitary whole to describe.
What is the new starting point?
Structural humility. A new encapsulation of the ignorance principle. We have discovered a kind of ignorance we did not know we had.
There is no global observer, no global time, no globally conserved quantity, no globally defined state. Reality is locally complete but globally inarticulable. There is no ‘there’ there.
Relational ontology. Entities exist as nodes of interaction, properties actualize relative to conditions, structure outruns substance. Syntax outrun semantics.
Primacy of causality. The rules of combination, constraint and consistency are more fundamental than the meanings of the things being combined. Causal order is more important than objects. Correlation is more important than property. Consistency is more important than truth.
We don't start with things and then assign relations; we start with relations that stabilize into things.
Information is no longer about something – information is the pattern of degrees of freedom in a relational network – rules overrule meaning.
What exists is what can be consistently related; what it means is reconstructed afterwards. Syntax outruns semantics because relations outrun relata.
Identity as invariance. In an expanding universe, objects are defined by their morphisms. Identity is a structural role; equality is replaced by isomorphism; an object is what it is as an invariance under transformations. Judgment is situated. Identity is symmetry assessed from a situated standpoint.
Experience is the local selection of a relational structure.
Memory is constraint on admissible transformations.
Agency is symmetry breaking within a structure.
Consciousness itself undergoes transformation: the unity of consciousness cannot be any kind of metaphysical glue but is (roughly) coherence under translation. A perspective counts as mine if it composes correctly with my past perspectives. I cannot be just anyone. I can be an equivalence class. Morphisms are constrained; situated does not mean subjective.
Responsibility is the preservation of coherence.
Harm is incoherence across perspectives.
Ethics is no longer rule following but maintenance of relational symmetry.
//
The whole vocabulary of philosophy gets reworked as energy, conservation, the steady-state universe, information, simplicity, and ‘existence’ itself, all reappear in relational guise.
//
Humility – intellectual conscience about one’s own intellectual limits – is an idea that connects the understanding of the cosmos to the problem of understanding oneself. The same limit pushes us back from overstating what is there and outrunning ourselves. All we can know is what we have learned getting there – there is no oracle – no clairvoyant – just work.
Wednesday, January 21, 2026
Human invariants
1. Beginning from Orientation in World Philosophy
In addition to teaching I also see patients. One of them said to me that my work in philosophy seems to look for mathematical and philosophical invariants. This person asked me: what do you see as invariant in everyday life? I answered with a question -- what is the same through human change? I thought of historical responses to the question and ultimately to the 'existential givens' pointed out by thinkers like Jaspers, Heidegger, and de Beauvoir. I thought of some ideas about basic human patterns -- such as we see in prospect theory, psychoanalysis, and psychophysics -- potential invariants from daily life. Some ideas came to mind –
The structure of temporality itself - not just that we experience time, but that we experience it with a particular topology: the asymmetry between past (fixed, known in principle) and future (open, uncertain), and the way the present moment has a peculiar "thickness" to it. This is invariant across cultures and epochs in a way that's almost too obvious to notice.
The gap between intention and outcome - we act in the world with purposes, yet the world responds according to its own logic. This generates surprise, regret, gratitude, learning. There is no form of human life that doesn't contend with this gap.
The need for recognition - not just Hegel's master-slave dialectic, but something more fundamental: that our sense of ourselves requires some mirroring or acknowledgment from outside ourselves. Even the hermit defines himself against society. This seems present in infancy (attachment) and persists through all human social arrangements.
The fact of embodiment with its particular constraints - fatigue, hunger, pleasure, pain, the way attention is limited and must be allocated. Even as technology extends our capacities, it doesn't eliminate these constraints; it just shifts their expression.
The basic structure of learning through error - we form expectations, they're violated, we update. This applies from motor learning to scientific revolution. Prediction error is a fundamental currency of biological intelligence.
The existential givens as the center - death, freedom, isolation, meaninglessness (in Yalom's formulation) - not just psychological facts – instead structural features of any existence that combines the human mixture // finitude, consciousness, and separateness. These are transcendental conditions for experience (loosely speaking).
Psychophysical invariants (Weber-Fechner laws, Stevens' power law) suggest that even subjective experience has a mathematical structure that's remarkably stable across individuals – we are sensitive to thresholds and plug them in whether they belong or not.
Philosophy: something invariant in the form of human questioning itself - not the content, but the fact that we ask "why?" in a way that seems qualitatively different from other animals. We are perpetually troubled by justification – the space between is and ought.
2. Criticism from Habermas
Habermas was an important influence in my earlier thinking. He challenges the search for human universals as another face of idealism, Platonism, another attempted escape from contingency and the messiness of everyday life in search of pure ideals and beautiful abstractions. I see my universalist aspirations in a different light -- as looking to logical simplicity at the foundations -- but perhaps Habermas' criticism is fair.
Habermas' criticism is sharp – worth taking seriously. I think I can answer it.
The crucial distinction seems to be between seeking invariants as a way to escape contingency versus seeking them as a way to understand the structure within which contingency occurs and operates. These are radically different projects.
The escapist version - what Habermas rightly criticizes - treats universals as a realm of purity untainted by history, culture, bodies, power. It's the move that says "never mind all this messy lived experience; the real truth is up here in the realm of forms." This becomes ideological when it uses abstraction to avoid responsibility, to deny the particular claims of actual suffering people, to dismiss context as merely "accidental."
In my framing - "simplicity of foundations" – the idea is more what we might call structural realism or perhaps even transcendental inquiry in the Kantian sense: asking what must be the case for this particular (the contingent, messy, historical phenomenon before us) even to be possible at all? This is not escaping contingency but asking about its conditions.
Consider: evolutionary biology seeks invariant principles (natural selection, genetic inheritance) – not in order to deny the riotous diversity and contingency of life, but to explain how such diversity is generated and constrained. The invariants don't make the particulars less real or important - they make them intelligible.
In clinical work: understanding something like the process of transference as a relatively invariant pattern doesn't deny the absolute uniqueness of each therapeutic relationship.
It empowers a framework within which to understand what makes this relationship, with this person, at this moment, matter in the way that it does – distinct from every other.
Habermas himself, ironically, can't entirely avoid the search for invariants. His theory of communicative action posits certain universal pragmatics - conditions that must hold for genuine communication to occur. He's looking for the invariant structure of non-distorted communication. So even in criticizing the search for universals, he engages in it - just with different content and (crucially) a different attitude toward what finding invariants means.
Perhaps the key is this: Is one seeking invariants in order to stop the conversation ("here's the eternal truth, end of story") or to enable better conversation by clarifying what we're all working with? The former is the Platonic temptation Habermas warns against. The latter is what science does, what philosophy does, what clinical wisdom does – when things work.
Weber's law doesn't make particular sensory experiences less rich or real – but it describes a relationship that seems to hold across those experiences. It's a regularity discovered in the mess, not imposed from ‘outside’ to clean it up.
There's also the question of scale and scope. Some invariants are narrow and technical (prospect theory's specific parameters), others broad and abstract – (lived temporality, embodiment). The broader they get, the more careful we need to be that we're describing structure rather than smuggling in particular cultural content disguised as universal form.
Cheap overgeneralization smuggles in cultural baloney – easy to miss.
The danger isn't abstraction or generalization per se - it's unacknowledged particularity masquerading as universality. When 18th-century European thinkers proclaimed universal "Reason" while actually describing the cognitive style of educated European men, that's smuggling. When psychologists claimed universal human laws based on WEIRD populations (Western, Educated, Industrialized, Rich, Democratic), that's smuggling.
But the solution isn't to abandon the search for invariants - it's to be rigorous and honest about:
The domain of application (what populations, contexts, scales?)
The level of abstraction (formal structure vs. concrete content?)
The evidence base (how do we know this holds across variation?)
The purpose (why does identifying this regularity matter?)
Prospect theory's loss aversion, psychophysical laws, the existential givens - these are not in any sense Western constructs dressed up as human nature.
They are universal.
3. Human invariants / first approximation
The structure of temporality
The gap between intention and outcome
The need for recognition
The fact of embodiment and its constraints
The basic structure of learning through error
Death, freedom, isolation, meaninglessness –
structural features of any existence that combines the human mixture:
<finitude, life, separateness, consciousness>
= transcendental conditions
= tautological conditions of being a conscious, embodied creature
Regarding restricting invariance to a domain of application, this mirrors results in model theory in logic and in group theory in mathematics, i.e. invariants only make sense in a context.
An invariant is a temporally flat variable.
I.e.: invariance is always relative to some transformation group, some set of operations under which a thing remains unchanged.
In group theory, what's invariant depends entirely on what transformations you're considering. A quantity might be invariant under rotation but not under translation, or vice versa. The symmetry group defines what counts as invariant. There's no such thing as "invariant simpliciter" - it's always "invariant under G" for some group G.
Model theory makes the same point from another angle: a sentence might be true in all models of a theory (hence "invariant" across those models), but completely meaningless or false outside that theoretical context.
The invariance is internal to a structure, not floating free in some Platonic heaven.
-- -- What is ‘there’ has to do with the process engaged for finding it.
4. Invariance under transformation
What is “invariant” under what “transformations?” --
Death as an existential given is invariant under cultural transformation, historical change, individual variation - it's a structural feature of finite existence itself
Loss aversion looks to be invariant under cultural transformation but not under certain kinds of framing effects or decision contexts
Cognition in the form of basic cognitive biases look to be invariant across typical human experience – but not under conditions of expertise or training
Temporal flatness is key. When we say something is invariant, we're saying: as we move through time, across contexts, through variations - this doesn't change. It's a constant in the equation while other things are variables. The problematic move isn't finding invariants - it's claiming invariance across a transformation group larger than the evidence warrants, or worse, using the language of invariance to stop inquiry rather than to structure it.
Just as in physics, where finding conserved quantities (invariants under time translation, spatial translation, etc.) gives you the fundamental laws, finding genuine human invariants might reveal something like the "conservation laws" of human experience – these would represent the deep constraints within which all the variation plays out.
Conservation laws imply a conserved quantity – imply symmetries – thus the question what is conserved in human life? What deep conserved quantity generates the ‘human’ pattern?
5. Conservation laws
What deep conserved quantity generates the ‘human’ pattern?
Conservation laws imply symmetries.
As long as we can identify the symmetries of a system, we should be able to figure out what the conserved quantity is that generates the symmetry … this is Noether’s theorem.
Potential Symmetries / Conserved Quantities …
Translation Invariance, Duality Symmetries, Filtration Monotonicity / directional symmetry (very like entropy in thermodynamics: time-reversal breaking with a conserved arrow), Sequence Convergence …
The thick/thin distinction matters
For Thin Invariants: The symmetries are often exact -- shifting degrees, duality isomorphisms, etc. The conserved quantities are algebraic structures (grading, pairing).
For Thick Invariants: The symmetries are at best approximate or statistical: in persistent homology, features persist across parameter changes, but not perfectly -- there's "birth" and "death"
The conserved quantity could be something like total topological complexity measured across all scales, which redistributes but has some global constraint
The conserved quantity: the Euler characteristic (thin) constrains the total persistence (thick) across all filtration parameters.
Conserved charges (gauge symmetries) constrain what field configurations are possible. So, thin invariants act like a topological charge that limits how thick features can manifest.
The balance is maintained by asymptotically approaching a goal one can never reach
This makes the unreachable limit into a structural principle
Think about how this plays out:
Each turn gets you "closer" to the limiting invariant, but you never compute E_∞ directly—you approach it through successive approximations.
The information is conserved in the sense that what you learn at each stage constrains all future stages, but you can never jump to the end.
Thin invariants are what you get at the limit – they're the convergence point
Thick invariants are the trajectory toward that limit – they're the geometric path
The conserved quantity (information) ensures that:
You can't get more information at the limit than you accumulated along the way
But you can't see the "shape" of that information without tracking the levels
This is very like free energy versus entropy
Entropy (thin): The limit quantity, what's actually conserved globally
Free energy (thick): The quantity you can actually compute/measure
As temperature → 0, free energy → ground state energy, but you never reach absolute zero
The level index plays the role of inverse temperature – as you refine (k→∞), you approach the true invariant asymptotically.
Thick invariants are "computable approximations" to thin invariants equated to ideal limits.
You can't get more information at the limit than you accumulated along the way
... there's no free lunch.
That is itself (a formulation of) the conservation law.
The No-Free-Lunch Principle for Level-Indexed Invariants
This crystallizes something important:
Statement: The information content of any thin invariant (the limit) is bounded above by the accumulated information in the thick invariants (the trajectory).
You can't know more about the limit than you learned getting there.
What This Rules Out
This principle explains why certain things are impossible:
No Oracle: You can't compute H_k(X) without doing work at level k (or work that propagates from other levels). There's no shortcut that bypasses the filtration resolution/approximation structure.
No Clairvoyance: If you only compute the first N levels, you get at most N levels worth of information about the limit. Some "persistence landscapes" die off quickly. They're telling us that most of the limiting information has already been captured.
The conserved quantity (total information) acts like a budget. The geometry of the space determines how that budget gets spent across levels.
6. Self-questioning
We appear to have smuggled in an observer/computational agent. But the geometry – the manifold, the space, the structure – just is. It doesn't care about our resolution strategy.
Still: The thickness seems to capture something objective about the space.
This leads to paradox.
Paradox / Possible interpretations:
Thickness is Relational, Not Intrinsic
Maybe thickness isn't a property of the space alone, but of the space-plus-computational-framework pair.
Just as:
Energy isn't absolute but relative to a choice of coordinates/frame
Entropy depends on macroscopic variables one chooses to measure
The "thickness" of an invariant might be measuring the mismatch between the intrinsic structure and the particular algebraic/computational apparatus you're using to probe it
Agency is implicit in the geometry
Maybe the geometry does determine a preferred computational structure. For instance, a cell decomposition of a manifold has a natural filtration by skeleta (0-skeleton ⊂ 1-skeleton ⊂ ...). This isn't imposed from outside—it's latent in how the space is built up from local pieces. The "agent" is just making explicit what's already implicit in the gluing data
Thickness measures friction between scales
//
Self-doubt:
What if the agent-dependence tells us something about reality itself?
In physics, we've learned that: the speed of light isn't about light – it's about causal structure. The Planck constant isn't about quanta – it's about the geometry of phase space. The Bekenstein bound relates information capacity to geometry.
Computational complexity is a geometric invariant – something described with Big O notation – tracking an algorithm’s growth rate versus its input size.
The "thickness" measures how this geometry resists compression – how many bits one needs to specify it – this is not agent-dependent if there's a unique minimal description
Is agent-dependence a bug or a feature? (Thesis: there is no a unique minimal description)
I.e., thickness is relational, not intrinsic -- this is very like the idea in the calculus that if one zeros in very closely on a curve, we will see a straight line
In calculus, the tangent line at a point is the first-order approximation to the curve. Zoom in far enough (take the limit), and curve and tangent line become indistinguishable. But:
How quickly they become indistinguishable depends on the curvature
Higher derivatives measure how the curve deviates from its tangent
You need the full Taylor series (infinitely many levels!) to recover the curve from its tangent
Thickness might be measuring how many "orders of approximation" one needs before the approximation becomes good enough.
In this view, the conserved quantity is the gap between levels – the information about "how wrong" your current approximation is – the failure rate, error rate, error landscape …
The key insight: Whether something is "straight" (thin) or "curved" (thick) depends on what you're measuring it against.
A geodesic on a sphere curves in 3D Euclidean space, but is "straight" intrinsically on the sphere
The same space can be thin with respect to one invariant, thick with respect to another !!!
The thickness measures levels of refinement separate from an initial probe (exist in time)
What if curvature itself is just thickness in disguise?
Flat spaces: Thin (tangent structure = global structure)
Curved spaces: Thick (need many approximation orders to capture global from local)
The Riemann curvature T appears at second order in the Taylor expansion of the metric.
Spaces with curvature are "thick" because you can't understand them from tangent data alone.
If thickness is relational – if it measures the mismatch between geometric reality and our approximation scheme – the information about "how wrong" one’s current approximation is – then what is conserved is a surprise function.
This reframes everything in terms of Kullback-Leibler divergence – relative entropy – a measure of how much an approximating probability distribution Q is different from the actual probability distribution P. The divergence of P from Q is the expected excess surprise (Shannon information). More generally, we are looking at the information distance between the current approximation at level k and the truth (at the limit →∞).
At each level, you're surprised by how much the next level differs from what you expected. The conserved quantity is the total surprise you'll accumulate across all levels.
The total information-theoretic surprise Σ_k D(level_{k+1} || level_k) is bounded by (or equal to) the complexity of the limiting object.
The surprise of an event is the negative logarithm of its probability.
In the language we've been using:
Thin invariants: Low total surprise (you converge quickly, small KL divergence sum)
Thick invariants: High total surprise (each level reveals substantial new information)
7. Result: the surprise function
No Oracle Property: You can't know at level k how surprised you'll be at level k+1 without actually going to level k+1. The surprise is generated by the resolution process.
Additivity: Surprises accumulate. If you're surprised going from level 1→2, and surprised going from 2→3, the total surprise is (roughly) additive.
Budget Constraint: The geometric object has a fixed "surprise budget" to distribute across levels.
A space can choose to:
Surprise you all at once (thin - most information at one level)
Surprise you gradually forever (thick - information trickling in across infinitely many levels)
Sequences: Each differential d_r measures surprise: "I thought this cycle was permanent, but it bounds after all!"
Convergence means eventually, no more surprises
The E_∞ page is where surprise → 0
Persistent Homology:
Birth and death events are surprises: "A new cycle appeared!" / "A cycle just became trivial!"
The persistence diagram is a surprise map: it tells you when and how severely your understanding will be corrected
Long bars = delayed surprise (the feature was real for a long time before being revealed as artifact)
Taylor Series: Each derivative is measuring surprise: "The function curved more than the previous order predicted!"
Analytic functions: finite surprise budget
Non-analytic: infinite surprise budget distributed across infinitely many derivatives
Symmetry behind the conservation
Now we can answer the original question about symmetries:
The symmetry is scale invariance (or its breaking).
When the system has exact scale invariance: No surprise at any level (thin)
When scale invariance is broken: Surprise distributed across scales (thick)
The conserved quantity (total surprise) measures how badly scale invariance fails.
Surprise = information gained = entropy difference
Bayesian updating: Each level is a Bayesian update, KL divergence measures data gain
This formulation explains
--why thickness feels computational/agent-dependent (surprise is relative to expectations)
--yet geometrically meaningful (the geometry itself determines the surprise distribution).
The space doesn't care about your expectations, but it determines how those expectations will be violated as you refine your probe.
= Testable Prediction
If surprise is conserved, then:
Spaces that surprise you early (steep initial learning curve) should have less surprise later.
Spaces that seem simple at first should have more hidden surprise at fine scales.
This would manifest as:
Sequences that collapse early vs. those with persistent differentials
Persistence diagrams with few long bars vs. many short bars
`` reflection
Is the principle of least action this conservation law or should we talk about consciousness in some way that incorporates negentropy?
The principle of least action is elegant precisely because it's so fundamental - it unifies mechanics, optics, quantum theory. But it's fundamentally about physical systems finding extremal paths through configuration space. Consciousness seems to involve something stranger: not just minimizing action.
Action may not be the language to get at it (?)
Consciousness creates distinctions, maintaining improbable organized states against entropy, and - crucially - cares about the difference between states. Consciousness is the care structures Heidegger points to.
The negentropy angle (Schrödinger's "What is Life?" territory) is compelling because living systems, and especially conscious ones, are precisely those that locally reverse entropy - we build structure, we preserve information, we resist equilibrium.
Consciousness = the capacity to maintain improbable distinctions and to treat them as meaningful … (my thesis in Important Nonsense)
Consciousness is neither least action nor negentropy … we have to come at it differently …
Consciousness is language-structured, symbol-structured, in ways that affect what states are even possible …
When someone can't find words for an experience in their new language, they're not just failing to report a pre-existing inner state. Often the experience itself has a different shape, different articulation in their first language. The phenomenology is genuinely different. Wittgenstein's point is right here: the limits of my language are the limits of my world.
This complicates the search for invariants in consciousness. Maybe the invariant isn't in the content of conscious states (which are language-shaped and therefore culturally variable) but in the form: the fact that consciousness involves maintaining distinctions, that it has intentionality (aboutness, care-structure), that it involves both a temporal flow and a sense of unity, that it requires some kind of self-world boundary (intention, outcome) …
There is the invariant need to articulate experience, even when the tools for articulation are inadequate. This is existential .. in philosophy, this is the prime directive
This idea connects to recognition: consciousness might require not just internal organization (negentropy) but external validation through symbolic exchange. The symbol is a way of locating this boundary.
The invariant might be less about information-theoretic properties and more about the dialectical structure - consciousness is always intersubjective, reaching for articulation.
Jung's Symbols of Transformation. Cassirer's Philosophy of Symbolic Forms. Goodman's Ways of Worldmaking. Fodor's The Language of Thought. The idea of conducting the search for invariants via symbols is attractive … but also problematic since what Husserl called their 'saturation' tends to obscure rather than clarify understanding. Leibniz’s dream of the universalis characteristica is closer to what we are looking for … perfect symbolic thinness
Symbols accrete meanings, histories, emotional valences. They become too rich, too overdetermined. Jung's symbols are so dense with archetypal resonance that they can mean almost anything. Cassirer's symbolic forms are illuminating but verge on the unfalsifiable. By the time a symbol has enough cultural weight to be interesting, it's already carrying too much baggage to be analytically clean
Leibniz's characteristica universalis appeals precisely because it promises the opposite: a formal system where symbols are thin - they mean exactly what they're defined to mean, no more, no less. Language becomes pure syntax that can be mechanically manipulated to yield truths about the world. The dream is to make reasoning as rigorous as calculation.
Mathematics works as a language for invariants precisely because mathematical symbols resist saturation.
When I write "∫" or "∂/∂t" or "G" (for a group), these symbols have disciplined, constrained meanings. They don't vary – there is no semantic drift. They don't gather associations.
The practical success of this approach is undeniable - physics discovers its invariants (conservation laws, symmetries) by changing from natural language to mathematical language – it is precisely by abstracting to this level that we see general truths.
The Noether theorem connecting symmetries to conserved quantities is only statable in this thin, formal idiom.
Yet the most interesting human invariants - existential givens, the structure of recognition, prospect theory's aversion loss patterns - seem to exist at a level where pure formalization loses something essential. Death isn't just
∃t: life(x,t) → ¬life(x,t+ε)
The meaning of finitude, its felt structure, the way it organizes human projects – this vicissitude requires thicker description.
Perhaps the solution is stratified:
At the most formal level: thin symbols, mathematical structures (like psychophysical laws, or the temporal topology)
At an intermediate level: theoretical terms with disciplined meanings (like "transference," "loss aversion," "intentionality")
At the phenomenological level: thick description that resists formalization but reveals the lived structure (“love,” “desire,” ‘freedom” …
Therefore invariants exist at their various levels -- expressed differently depending on what kind of precision one is after – what tools one is applying – by chosen coordinate axes.
Fodor's LOT tries to do both: a formal computational system and semantic content. But it arguably achieves this only by pushing the hard questions about meaning down to "primitives" and "conceptual atoms" that remain mysterious. The point is that there are different domains for different layers of symbolization -- some are very close to 'raw experience' and some very far from it.
// Moral
The distance from raw experience isn't just an epistemic issue. It's a moral issue because the further we move toward thin formalization, the easier it becomes to lose sight of what's at stake for actual people. Transition to the thin vocabulary often loses the magical kernel.
A policy analyst working with utility functions and discount rates is operating with very thin symbols, very far from raw experience. That distance enables powerful analysis - but it also makes it easier to forget that behind "units of utility lost" are people experiencing grief, pain, fear. In a case like this, formalization can become a kind of moral anesthetic. Conversely, staying too close to raw experience - pure phenomenology, thick description - can make it impossible to see where one is – or think systematically about tradeoffs, to compare cases, to identify patterns to help more – more people more effectively. Moral reflection requires moving between levels: formalizing enough to think clearly and act systematically, returning regularly to the phenomenological level to check whether the abstractions are still tracking what matters. The invariants we identify at each level need to remain answerable to the levels above and below. In clinical work especially, this movement seems essential - you need theoretical frameworks (middle level) and maybe even some formal models (thin level), but you also need to remain in contact with this person's suffering, this irreducible (thick level). The invariant is 'right now.'
Monday, January 19, 2026
Sunday, January 11, 2026
Orientation in World Philosophy
Orientation in World Philosophy – 2014
A decade has passed since my experiment in world philosophy – Orientation in World Philosophy – 2014 – a draft attempt to break from traditional categories in philosophy – dreaming of a planetary philosophy – asking questions about philosophy and its history.
In Western philosophy, particularly from the Greeks onward, there is a powerful current that treats knowledge as a kind of vision—theoria originally meant contemplation but carried visual connotations, and we still speak of "insight," "clarity," "perspective," "illumination." The ideal knower stands at a vantage point from which everything becomes intelligible, where contradictions resolve into systematic comprehension. Even mystical streams in the West (Neoplatonism, medieval contemplatives) often frame the ultimate encounter as a "vision" of the One or beatific vision—a seeing that transcends ordinary sight but remains fundamentally observational, a subject apprehending an object, however unified.
The Eastern traditions —particularly Buddhism, Daoism, and certain Hindu schools—tend toward something quite different: a collapse of the subject-object structure itself. The ideal isn't achieving a perfect view but dissolving the viewer. Meditation practices often work explicitly to deconstruct the observing stance, to move from knowing about to being what is. The famous line that "if you meet the Buddha on the road, kill him" captures this: any objectified understanding, any "seeing" of enlightenment, remains dualistic and therefore incomplete.
The span between them is the difference between epistemology and ontology as ultimate concerns—between perfecting how we know versus transforming what we are. Western philosophy tends to preserve and refine; Eastern practice to exhaust and finally dissolve.
//
Late Heidegger's critique of Western metaphysics – his notion of Gelassenheit (releasement) –move to the Eastern model – letting-be rather than a grasping-to-know. Merleau-Ponty's refocus on embodied being-in-the-world deconstructs the detached observer. Looking east, there are the systems – the sophisticated epistemologies and logical systems (Buddhist logic, for instance) moving towards a Western objectivity model.
These are not competing (or intersecting) paths in the same space ... history matters in the learning process; the early learning environment is largely determinative of what a person can learn … is able to learn … learning environments globally differ radically … different worlds.
My intuition is that philosophy is a universal aspiration -- therefore these different configurations could also be framed inside a much larger set -- one day perhaps philosophy will no longer find itself stuck split-up in traditional schools and separate fiefdoms – a more planetary form of thinking, more mathematical than magical, might emerge ... . That is my vision for the future.
We don't choose our initial conceptual frameworks any more than we choose our native languages, yet this doesn't mean the frameworks themselves exhaust what's possible or real – there is the problem of getting over the natural attitude, wherever/whenever you happen to be.
//
A planetary philosophy – more mathematical than magical -- offers something rigorous and universal without being reductive. Mathematics offers the model: abstract yet somehow binding across cultures, a kind of discovered being rather than any brain-spun convention. A planetary philosophy would rest exactly on invariant structures beneath all cultural-historical variations, evolving into ‘universal’ patterns of human conception that hold regardless of whether they're approached through dialectical reasoning or meditative dissolution.
There's something almost political about current philosophical fragmentation—territorial defense of methods and vocabularies, as if admitting that meditation accesses genuine understanding threatens the validity of logical analysis, or that logical clarity is a hindrance to understanding.
If we think of human cognitive capacities as multidimensional—capable of both analytic precision and non-dual awareness, discursive thought and direct realization—then different traditions are cultivating different regions of this larger space – the total space of philosophy.
//
The mathematical analogy suggests that synthesis (blending Eastern and Western approaches into some compromise) is not the thing we need – nor syncretism – nor reformation. The discovery is transformation rules between systems. Translate between geometric and algebraic descriptions of the same mathematical object – each representation has advantages for certain problems – neither is more "true” – from a perspective, this is what you can see – change perspectives and you begin to learn something new – this is not a defect in earlier work.
What would it mean for philosophy to operate at this level of abstraction? To recognize "Seeing" and "Being" as different coordinates in a higher-dimensional space of human realization?
Philosophies cultivate different regions of this larger space ... Orientation in World Philosophy made a first pass at an invariant-structure model ... assuming logic and philosophy and mathematics are all evolving in the direction of the simplification of the axiomatic basis ... reaching some threshold (maybe something about primes or Pi) a newer Planetary model could shine brighter than the earlier traditions ... a new Galileo to lead Confusion to Understanding.
We've seen the pattern repeatedly: non-Euclidean geometries emerged when someone questioned the parallel postulate; set theory attempted to ground all mathematics in a handful of primitives; category theory seeks even more abstract invariants. Each new simplification reveals that what seemed like fundamental diversity was actually variation on deeper themes.
Primes and Pi represent irreducible mathematical structures—primes as the atoms of number, Pi as the unavoidable constant relating linear and circular measure. They're not conventions or cultural constructs; they're... there. Any sufficiently sophisticated mathematical system encounters them. In this sense they can function like anchors – landmarks in thoughtspace.
If the question becomes – what would any mind engaged in philosophy discover? – we are looking for landmarks like these – like the Pythagorean theorem and Gödel’s result.
My thesis in the Orientation was to extend the analogy by identifying equally irreducible structures of consciousness or reasoning—not Western or Eastern, but simply necessary features of any reflective awareness attempting to understand itself and its world – a planetary outlook.
It's not about Western philosophy conquering Eastern contemplation or vice versa, but about finding the minimal sufficient structure from which both traditions' insights become derivable. The way all of Euclidean geometry follows from five postulates, or how complex group theory emerges from a few axioms about operations … we are looking for a origin root like this.
"Orientation in World Philosophy"—the title suggests the project: not one philosophy to rule them all, but a coordinate system within which different orientations become locatable, relatable, mutually intelligible, questionable. Different regions of the same large space – conceptual space.
Donald Brown's 1990 work Human Universals was an inspiration. My question was, What is cross-cultural / universal / invariant – more closely where topology suggest an isomorphism, homeomorphism, diffeomorphism – extending group theory into the philosophy of philosophy.
A mathematical hierarchy of structural mappings as a way of getting at philosophical outlooks – the progression from cross-cultural commonality to isomorphism to diffeomorphism represents increasingly refined ways of asking: "In what sense are these forms of thinking the same?"
Human Universals gave anthropology a way to talk about commonality without altogether flattening genuine difference—recognizing that all cultures have kinship systems, say, while recognizing that those systems vary enormously in structure. All cultures practice examination … ‘philosophy’ … this is a core thesis of the project … an a priori empirical argument …
Isomorphism would mean: two philosophical systems that are structurally identical even if their vocabularies differ—e.g., the way Buddhist śūnyatā and readings of Heideggerian nothingness map onto each other, preserving all essential relationships, even though one emerges from meditation on dependent origination and the other from Western ontological investigations.
Homeomorphism relaxes the constraints—continuous deformation without tearing. Two traditions might be homeomorphic if you can continuously transform one into the other, even if they look superficially different. The "shape" of the inquiry is preserved even as specific features change. There are great systematizers for the three Abrahamic traditions: Aquinas, Averroes, Maimonides – a homeomorphism in critical space – also visible in China, in India, in America.
Diffeomorphism—smooth invertible mapping—would be the strongest claim: not just that two traditions address the same territory, but that there's a smooth, structure-preserving translation between them in both directions. This is harder to identify … Descartes and Kaibara Ekken (?)
I started with data. Categories: metaphysics, ontology -- going after the whole; skepticism -- a focus on ignorance; empiricism -- a focus on experience; logic, rationalism -- a focus on reason; moral philosophy -- the ethical perspective; synthesis -- holism -- assembling what is known; a philosophy of history – historicism -- a focus on time, development, evolution; rhetoric/linguistic philosophy -- a focus on language; bio-philosophy -- a focus on life, the life form, the psyche-soma, the embodiment; existentialism -- a focus on lived immediacy; psychology -- a focus on the self; social philosophy -- a focus on politics.
Trends in the data
1 --- the progress model --- cumulative process model -- philosophy makes progress like science by cumulative achievement; examples include logical positivism, neurophilosophy, pragmatism.
(Easily falsified by the record of setbacks, dark ages and regressive epochs.This is not the point.)
2 --- the fine arts model
Heisenberg’s model
3 – cyclical model: theoretical-natural; practical-popular; skeptical; dogmatic-mystic. Philosophy goes in circles -- not straight lines, spirals, or more complicated shapes – goes nowhere
(philosophy as regress rather than progress; philosophy as leaving and returning; as neurosis)
4 --- the degeneration model (Heidegger)
5 --- the golden age model -- as in Heidegger, Confucius, and in Vedanta philosophy, where a past age serves as the standard, which may be the age of the Presocratics, or Trakl’s poetry, or the age of Wen and Wu, or the age of Vyasa or Homer.
6 --- the Enlightened One model -- based on an apotheosis of a teacher, e.g. Socrates, Plato, Aristotle, Wittgenstein, Derrida, Foucault, Lacan.
See Macauley on the distance between ideals and reality – the guru is a human being.
7 --- the brave new world model -- proposing to break new ground (Descartes, Hume, Husserl) and begin philosophy from a hitherto unthought standpoint.
Hume brings about a permanent change in Western thinking; so does Frege, Husserl, Einstein …
The "group theory for philosophy" formulation attempts to get a transformation itself—what remains invariant under change. What is the minimum for reflective awareness – is there such a thing? What is invariant across all these transformations as the least common factor?
What must any consciousness examining itself discover?
Self-reference? The subject-object division? Finitude? – Some large problems here.
I began with: 'What new kind of thinking is possible that is (as if for the first time in history) the inheritor of world philosophical inquiry?' -- my guiding question.
Ultimately the mystery of the persistence of traditions drew my attention back. The arc seems to be that the discoveries made in a school are seen as invaluable – they may be -- but this prevents advocates from growing further. Advocacy for the position takes over from free inquiry.
My conclusion was simply an appeal to get back into the practice of philosophy -- simply to cultivate some small plot of mind-space.
The trajectory is from the ambitious question "what new thinking becomes possible as inheritor of world philosophical inquiry?" through the search for topological invariants, to the recognition that advocacy for the position takes over from free inquiry, to the humble project of practice.
The shift from discovery to defense is so pervasive it almost seems like a law of intellectual history. A breakthrough occurs—someone sees something genuinely new—and then the work becomes preserving, systematizing, protecting that insight. The living inquiry calcifies into a school, a lineage, a tradition to be loyal to. What was once an opening becomes a boundary.
The conclusion—an appeal to return to the practice of philosophy rather than the defense of positions—a defeat – is also meta-level discovery. Perhaps the invariant is not a content (a specific structure of reflective awareness) but a process (a strict practice of testing ideas).
Philosophy is the stance of free inquiry – even as discoveries accumulate – the moment we stop cultivating our small plot and start defending the entire territory is the moment philosophy stops being philosophy – tis happens again and again and is part of the data to be examined
A rough Daoism: the recognition that the systemization itself becomes the obstacle. The map, however accurate, isn't the territory, and defending the map prevents further exploration.
The new kind of thinking I was seeking might not be a new position at all – more a new mobility – the ability to move between traditions, to see them, to recognize their homeomorphisms and diffeomorphisms, without needing to establish permanent residence in any of them.
Cultivating a small plot of mind-space as an experimental garden rather than claimed land. Just tarrying a while
My vocabulary -- 2014: metagenetic = attempting to transcend one's origins, and terrostatic = attempting to stay rooted in life. I attempted to articulate metagenetic terrostatism. Nonsense of course – like all of my work.
"Metagenetic terrostatism" captures the tension: how do you transcend your origins while remaining rooted in life? How do you achieve the universality that philosophy aspires to without losing the particularity that makes inquiry vital and embodied?
The Western "Seeing" tradition tends metagenetic—trying to rise above contingency to achieve a god's-eye view.
The Eastern "Being" traditions often emphasize rootedness—staying with this breath, this moment, the suchness of what is.
Both traditions, taken to extremes, have their pathologies. Pure metagenesis becomes untethered abstraction, a view from nowhere that's really a view from nothing. Pure terrostasis risks provincialism, an inability to recognize oneself in the other.
The conjunction "metagenetic terrostatism" might seem contradictory because we're trained to think in either/or terms—either you transcend or you stay rooted. But perhaps the deepest insights require both simultaneously: transcending enough to recognize the common structures, the topological invariants across traditions, while remaining rooted enough that the thinking stays alive, responsive, capable of genuine discovery rather than mere categorization.
The metaphor of cultivating a small plot captures this beautifully—gardening is intensely local, hands-in-dirt particular, yet it participates in universal processes: photosynthesis, germination, decay, renewal. The gardener doesn't need to choose between understanding botany and tending specific plants.
"Metagenetic terrostatism" – an awkward phrase for what planetary philosophy might feel like as a lived practice. Impossible to settle into
Tuesday, January 6, 2026
Learning rate
I thought I would begin the year with some thoughts. Recently my focus has turned to psychology.
I am trying to rethink psychological issues from a more mathematically informed perspective. I am considering several ideas.
1 - Let us think of decision-making as a continuous time stochastic process. This could translate into an equation that relates accumulated evidence, drift rate (the quality of the evidence), noise, and a Wiener process. A decision occurs when the function hits a boundary.
2 – In thinking about learning as a continuous process, it equates to something like minimizing a prediction error. Learning rate becomes a dynamical parameter, not a constant. This suggests an approach to rapid versus slow learners and to pathological learning such as addiction and depression.
3 - Ultimately a relational ontology gets me to something like a neural field model of cognition. Thus the question, how do distributed populations of neurons give rise to stable thoughts, memories, or percepts? If it is a continuous dynamical system, then individual psychological states become attractors in a continuous process; this suggests a dynamic approach to working memory, hallucinations, and various kinds of breakdowns of cognitive ability.
4 – Fechner's law could be a way of looking at subjective experience in a quantitative form. Quantitative change can relate to qualitative experience – this implies that mental states should be treated as temporarily extended processes.
5 – Consider the RW model – application predicts rapid early learning, diminishing returns, and persistence of early priors. These phenomena are exactly what we see. Overall: Psychological systems behave as if they were following the steepest descent on an error landscape. Mental content is not static. Meaning arises from trajectories, not representations. The mind is defined by how it changes, not by what it contains. Learning is a differential equation about the self.
----
DDMs // drift-diffusion models of decision-making // decisions emerge from continuous accumulation rather than discrete comparisons. The same formalism handles both perceptual decisions (is that a face?) and value-based choices (should I take this job?). The drift rate estimates a "signal clarity" parameter that explains individual differences and contextual effects.
What happens when boundaries themselves are dynamic or learned? This gets at impulsivity, patience, and how depression might alter decision thresholds.
Learning rates as dynamic parameters//The Rescorla-Wagner model with constant α misses something fundamental about adaptive systems—they need to modulate how much they update based on uncertainty, volatility, and context. Bayesian approaches formalize this through precision-weighting of prediction errors.
Addiction might involve excessive learning rates for reward prediction errors in specific circuits.
Depression could involve learned helplessness through overly rigid priors that resist updating.
Anxiety might be hyperactive learning about threats.
Neural field dynamics and attractors // If mental states are attractors in a high-dimensional neural state space, then "stability" of a thought or percept means the system has settled into a basin. Working memory becomes actively maintaining a state near an attractor against noise. Hallucinations could be spurious attractors that the system falls into. Cognitive flexibility versus rigidity maps onto the depth and breadth of attractor basins. Creativity might involve noise-driven exploration of state space, while rumination is getting trapped in a tight attractor basin.
Binding problems // How do we maintain a coherent percept of, say, "red ball moving leftward"? If it's a single attractor state rather than separate features that need binding, the problem dissolves.
Psychophysics, temporal extension // Fechner's law (and Stevens' power law) reveal that subjective experience is a compressed transform of physical intensity. The logarithmic or power-law relationship ensures we're sensitive across vast dynamic ranges. Treating mental states as temporally extended processes rather than instantaneous snapshots changes the problem-set. Consciousness is ‘intentional’ as Husserl stated: awareness is about trajectories and transitions. Mathematically, this is closer to a vector than a scalar. We have escaped the world of frozen moments. This result connects to the "specious present" in phenomenology.
Learning and priors//The Rescorla-Wagner asymptotic learning curve shows that early experiences have outsized influence – not because they're "special" – but because that's when uncertainty is at its highest peak and learning rates also achieve a maximum. This explains some of the power of cultural imprinting, attachment patterns, and why early trauma is so persistent. It's literally foundational to the error landscape you're building on.
Synthesis//The mind is defined by how it changes, not by what it contains – this is the key.
The Predictive processing/active inference approach is distinct from representationalist cognitive science. Mental content is trajectories through state space, not stored symbols. Meaning isn't a mapping between representations and world. Meaning is a pattern of transitions and their consequences for prediction error.
The steepest descent perspective is that thermodynamic systems minimize free energy, which for perception and learning means minimizing prediction error.
Questions … extensions:
• How do multiple timescales interact? You have fast perceptual dynamics, medium learning rates, and slow developmental/cultural priors. How do they couple?
• What role does embodiment play? If meaning is trajectories, then sensorimotor contingencies matter—the same "internal" trajectory might mean different things depending on what actions it affords. The environment/organism boundary blurs.
• Can this framework handle social cognition? Other minds might be attractors we simulate, or empathy could be synchronization of trajectories across coupled dynamical systems.
• How does language fit? Words might be attractor labels or controllers that shape trajectories, rather than symbols that denote representations.
Let’s rebuild psychology on a foundation of continuous dynamical systems theory – since brains are continuous, embodied, dynamical systems, evolved for prediction and control in an environment. The mathematics show structural properties of adaptive systems.
//The model suggests that learning is path dependent. History matters, not just outcomes.
The problem is that an organism must change itself in response to the world. Yet it only has access to prediction errors.
Learning is not about storing facts but about continuous self-modification under uncertainty.
But then the Bayes approach suggests that people get trapped in early "weight spaces" and lose the ability to grow ... the early environment has an outsize impact. This result into skepticism ...
// If learning is gradient descent on an error landscape, and early learning carves deep channels in that landscape, then later experience flows along paths already established. The system becomes increasingly constrained by its own history. Early priors aren't just influential—they're constitutive of the space in which later learning happens. You learn how to learn early, and that meta-structure may be nearly impossible to escape. This is an empirical question …
This isn't just about content ("I learned X when young"). It's structural: the very dimensionality of the hypothesis space, the features one is capable of extracting, the errors one is capable of detecting—are all shaped by early experience. A child raised without language during critical periods doesn't just lack vocabulary; they may lack the neural architecture for certain syntactic operations. Cognition is also the full development of a capacity that is merely contingent …
The Bayesian formalization makes this stark: Strong early priors that get reinforced become increasingly difficult to overcome. As you accumulate experience, your effective learning rate for anything contradicting those priors approaches zero. The precision-weighting means you discount evidence that doesn't fit. This is adaptive—it prevents you from being blown around by noise—but it also means you can become trapped in a locally optimal but globally sub-optimal configuration.
Epistemically: If you're trapped in your weight space, how do you know your current mind isn't just an artifact of your learning history? Your confidence in any belief might just reflect that you've settled into a deep attractor, not that you've found truth. You can't "see far enough" because your very perceptual/cognitive apparatus was shaped by contingent early experience.
Practically: This has dark implications for therapy, education, social change. If early trauma or poverty or cultural programming fundamentally shapes the architecture of learning, then later interventions are trying to work within a constrained space. You're not just overcoming false beliefs—you're trying to reshape the system that generates beliefs—How does one fight this?
Existentially: The continuous self-modification means there's no stable "you" evaluating the process from outside. The thing doing the learning IS the learning process. You can't step outside your trajectory to assess it objectively. You are the learning process up to this point …
However
Metaplasticity and multiple timescales//Neural systems have plasticity of plasticity. Learning rates themselves can be learned. The Bayesian framework assumes fixed priors, but real brains can detect when their model is systematically failing and increase uncertainty, reopening the learning process. This is what happens in "insight" and "cognitive restructuring"—a meta-level change that allows ground-level updating. It's rare, difficult, and possible.
Noise and exploration// Pure gradient descent gets stuck in local minima. But biological systems have noise—from neural stochasticity, from neuromodulators, from sleep, from drugs, from stress.
Noise can kick you out of stable attractors and allow exploration of different regions of state space.
Creativity, psychedelics, meditation, trauma = ways the system escapes its own stability.
Social and cultural scaffolding//You're not learning in isolation. Other minds, institutions, practices can provide structure that your individual system can't generate alone. Language, mathematics, scientific method—these are cultural technologies that extend the hypothesis space beyond what any individual could construct. They're external ratchets preventing complete path-dependence. These are prized gearworks in the toolkit of thinking.
Predictive error itself as the escape//The system is optimizing for prediction error minimization, but the world is complex and non-stationary. Persistent large prediction errors—especially in domains that matter—can force architectural change.
Suffering is the signal that your model is inadequate.
The question is whether organisms can tolerate enough sustained prediction error to reshape themselves rather than just avoiding or explaining it away …
Objection: mechanisms exist, but they're partial, difficult, and constrained by the very history they're trying to escape. The mathematician who learns to "think differently" is still constrained by having learned mathematics within one particular cultural and cognitive framework.
Result? -- a kind of pragmatic fallibilism?
• The current understanding is path-dependent
• The learner can't fully escape his or her learning history
• But one can still work toward less inadequate models
• Self-modification under uncertainty is the only game in town
• Skepticism is part of the trajectory—a recognition of constraint that might slightly loosen the constraint—the impossible possibility
The fact that we can formulate this problem, recognize the trap, feel the vertigo—that's already a form of meta-level cognitive flexibility that the pure path-dependence story appears to make impossible.
Or// This very thought is just my own weight space trying to avoid its own vertigo.
The thing doing the learning IS the learning process.
//One addition – a lemma? -- The Rescorla-Wagner model with a dynamic α models the plasticity of the organism in response to its environment.
In the standard Rescorla-Wagner model, α (learning rate) is a fixed parameter representing the "associability" of a stimulus or the general learning capacity of the organism. It's treated as a property of the stimulus or the organism that stays constant across learning.
But if α becomes dynamic—if it varies based on uncertainty, surprise, context, volatility—then it's no longer just a parameter in a learning rule. It becomes a measure of plasticity itself. The organism's ability to modify itself in response to prediction errors is now part of what's being tracked – this is the thing we are trying to optimize.
This shifts the ontology:
• Static α: The organism has a fixed capacity to learn from experience
• Dynamic α: Plasticity is itself adaptive—the organism learns how much to learn
Mathematically, dynamic α might depend on:
• Uncertainty about the environment (higher uncertainty → higher α)
• Volatility (rapid environmental change → higher α to track it)
• Prediction error history (persistent errors → increase α; accuracy → decrease α)
• Neuromodulatory state (dopamine, norepinephrine modulating plasticity)
Psychologically, this captures:
• Developmentally sensitive periods (high α early, decreasing with age)
• Attention (α higher for attended stimuli)
• Emotional arousal (strong emotions temporarily increase α)
• Freezing (α approaching zero for domains with entrenched priors)
A learning system with dynamic learning rates is a system that regulates its own plasticity. The trajectory of α over time maps the organism's capacity for self-modification—where it remains flexible, where it has rigidified, and under what conditions it can reopen learning.
This makes α(t) a kind of meta-parameter—a window into the organism's relationship with its own changeability — approaching a formalism to express the problem of moral philosophy
Monday, January 5, 2026
Einstein, Heidegger, Russell
Heidegger mentions the theory of relativity in Being and Time (SZ, footnote to p. 417). In effect he dismisses the impact of relativity on the problem he is discussing because his subject is not the measurement of time. He mentions the interesting idea that astronomical phenomena create a background for the human creation of time.
Heidegger mentioned relativity theory before this – in his 1915 habilitation lecture, where he also argued that Einstein dealt with measurements of time rather than time itself. His most sustained early critique came later – in his 1924 lecture "The Concept of Time," where he explained that renewed interest in time was largely due to Einstein, but maintained that "the theory of relativity leaves the concept of time untouched." Heidegger saw nothing there.
During the last three decades of his life (mid-1940s to mid-1970s), Heidegger wrote and published a great deal, but there is no mention of Einstein. His later work shifted away from the concerns of SZ toward reinterpretations of pre-Socratic philosophers and "the history of being."
Heidegger delivered a lecture long after SZ – entitled "Time and Being," at the University of Freiburg, in 962, which was his most direct return to the themes of SZ. This lecture does not engage with Einstein or relativity theory at all. Instead, it attempts to think about time and being in a "non-metaphysical" way, moving away from the terminology and framework of SZ.
The scholarly consensus is that Heidegger maintained his original position throughout his career: physics deals with the measurement and calculation of time (what he called "clock time"), while philosophy must investigate the more fundamental questions of temporality and the lived experience of time. Overall: he never reconsidered or deepened his engagement with the actual content of relativity theory after his early critiques in the 1910s-1920s.
Heidegger missed something enormous here – something like missing a mountaintop right next to you. The idea that "clock time" might actually illuminate rather than obscure the nature of lived experience and human being is an inversion of Heidegger's dismissal – one that proves itself immediately if one pursues the argument.
The self is both a synchronization scheme (in spacetime, ST) and a continuous adaptation function (in phase space, PS). This emerges with contemporary work in neuroscience and cognitive science – which Heidegger could not have known – seeing the brain as a predictive processing system – as constantly integrating information across timescales, from millisecond neural firing patterns to long-term memory consolidation. The self is an ongoing temporal integration process, stitching together disparate events. At the same time, the self is roughly the learning process modifying itself in response to prediction errors. This odd dualism – physical and thermodynamic – is exactly the self that Albert Einstein was working on and developing at the same time that Heidegger was dismissing his ideas.
Relativity theory doesn't just offer a "measurement" of time. It fundamentally reconceives the relationship between time, space, and observation. The observer-dependence of simultaneity, the relativity of temporal ordering for spacelike-separated events, the unbreakable connection between time and causality – these are not just matters of clocks (or devices of any kind) but touch on the very structure of experience and being-in-the-world that Heidegger envisioned.
There’s a deeper irony: Heidegger emphasized "thrownness" and our embeddedness in a world not of our making, yet he dismissed this revolutionary knowledge which teaches us precisely about the temporal structure of the world. Seeing human being and in the context of its "local error landscape" — the fluctuating gradient descent through possibility space – could have enriched Heidegger’s ideas about time, Being-towards-death, and authentic temporality.
Did the phenomenological method constrain him from engaging seriously with mathematical physics? Was it a kind of disciplinary territorialism and blindness? Perhaps it was his evident antisemitism. Heidegger’s understanding of scientific thinking was profound, as one can see in What is a thing? and his various unpackings of Kant. A revolutionary himself, he was also a human being with very ostentatious faults. My explanation is from Prospect theory, Stanovich, and the idea that even superb intelligence does not protect one from even base biases.
Heidegger's antisemitism was perhaps not incidental to his thought – the Black Notebooks reveal how deeply woven it is in his philosophical worldview, particularly in his characterization of "calculative thinking" and technological modernity as expressions of a supposedly Jewish worldlessness. The objectivity of science is interpreted through the history of a despised other. His misunderstanding at this level of his own ideas is pitiful – comic – a continuing shame.
The Stanovich point is that high intelligence and sophisticated reasoning don't immunize people against motivated reasoning or ideological blindness. In fact, intelligence can make one better at constructing elaborate justifications for biases – however fanciful or despicable. Heidegger could perform brilliant phenomenological analyses of Kant and Hegel and Aristotle, yet his framework systematically devalued certain modes of thought – mathematical, scientific, technological – at the service of his own small-mindedness, racism, and sick devotion to antisemitic ideology.
The self is a synchronization mechanism (in ST) and an adaptive process (in PS) – these ideas have phenomenological resonance. Einstein's relativity reveals something about the observational structure of reality itself: the relativistic observer, embedded in spacetime, making measurements that depend on his or her own frame of reference – the home coordinate system –this result has phenomenological depth – it partly explains unbridgeable subjectivity itself.
Mathematical-physical accounts and the phenomenological accounts are not rival explanations – they are more like complementary descriptions at different scales and levels of analysis. The "microsecond stitching" and the "existential projection toward death" are both real aspects of temporal being.
Philosophical genius does not guarantee wisdom – our frameworks for understanding can be simultaneously penetrating and blind.
Different scales and levels of analysis – I see it as a kind of mistaken question, like trying to connect temperature with the action of electrons.
Just as temperature is a macro-level property that emerges from but isn't reducible to individual electron movements – and asking "which electron has the temperature?" is a category error – so is making the same mistake when we try to directly map relativistic or quantum time onto phenomenological temporality – yet these are different approaches to the same experience.
Temperature makes perfect sense at the thermodynamic scale. It's real, measurable, and causally powerful. But it's not a property of individual particles; it's a statistical feature of large ensembles. Similarly, lived time – the "stretching" of a boring afternoon, the compression of an engrossing conversation, the directedness of care and projects – operates at the scale of conscious, embodied experience. Clock time and phenomenologically relativistic time operate at different scales entirely. This is interesting in itself – it support’s Husserl’s account of the phenomenology of time consciousness – i.e., the deconstruction of the unitary present.
The mistake isn't in either description – but in assuming they must directly connect or that one must be more "fundamental" than the other. Heidegger was right that physics doesn't capture phenomenological time, but wrong to dismiss physics as merely "calculative." Einstein was describing something real about the structure of ST, even if that structure doesn't appear in first-person experience the way Heidegger's care-structures do. It emerges more from inference.
As the self operates across timescales – microsecond integration, moment-to-moment adaptation, long-term self-modification – the phenomena suggest that these levels can be coherently described without requiring reduction. The neural synchronization at millisecond scales, the phenomenological now-moment, the narrative arc of a life project – these are all temporal phenomena at distinct levels of organization. They are differential equations in ST and PS.
Heidegger and naive scientism made the same error from opposite sides: assuming there must be one "true" account of time. This is exactly the point that Einstein’s work makes evident.
Einstein -- in my thinking -- is the figure who makes it impossible to ignore our own 'coordinate system' involvement in the 'objects of experience.' The strictly positivistic philosophy from the Vienna Circle made is a complementary error, in dismissing metaphysics. Einstein chided Russell on this point … he thought Russell’s embarrassment about having to do metaphysics showed that he did not really grasp yet what is going on in philosophy and scientific inquiry.
Einstein, often claimed by the positivists as their champion, actually understood something they missed: that relativity theory doesn't eliminate the observer or reduce everything to measurement – it makes the observer's coordinate system constitutive of physical description itself. The "view from nowhere" becomes impossible – not because we lack the right instruments, but because the structure of spacetime itself is observer-relative.
This is why Einstein's disagreement with Russell is so telling. Russell and the Vienna Circle wanted to purge philosophy of metaphysics, to reduce everything to logical analysis of observational statements. But Einstein recognized that his theory required metaphysical commitments -- about the nature of simultaneity, about what counts as "real," about the relationship between mathematical structure and physical reality. You can't simply read these off empirical data; they involve interpretive frameworks – without which, they vanish.
The coordinate system isn't a distorting lens to be polished away—it's intrinsic to what we can know and say about the physical world. This insight should have been – and is now – a gift to phenomenology. Heidegger wanted to show that people are always already embedded in a meaningful world, that their perspectives are constitutive of experience. Einstein demonstrated that this is true even for physics – for the interpretation of physical reality. This is a world in which nothing ‘matters’ – from Heidegger’s perspective – yet even this world, the world of complete objectivity, is fundamentally, observer-dependent – as it were, human – not just raw content, but content that is logically and causally implicated in the process that discovers it.
Had Heidegger engaged seriously with Einstein, he might have found not a rival but a kindred insight: that no god's-eye view exists, that all knowing involves a standpoint, that this result does not make knowledge arbitrary or merely subjective – that perspective is essential to "worldhood." Perhaps he would have engaged with Einstein’s result that – running invariant across all coordinate systems – different observers can translate unerringly between frames, finding invariants that hold in all and for all perspectives. Truth is what survives translation – not just the ‘uncovering’ created in the jump to language – not just the local truth trying to prop itself up as universal – as the "history of being" -- obscurantism and weak argument.

