Identifiability and Learnability of Disentangled Latent Strategies
Identifiability and Learnability of Disentangled Latent Strategies
This workspace is the persistent scratchpad for the theory thread on when a disentangled latent strategy factorization is actually identifiable or learnable.
The key issue is not just whether a latent z helps predict y, but whether
the same z can be interpreted as the same high-level strategy across
examples. That distinction is central to deciding whether the model has learned
strategy semantics or only a locally useful predictive partition.
Navigation
- 00-context: the problem statement and project context.
- 01-core-questions: the main conceptual questions to answer.
- 02-metrics-and-definitions: candidate identifiability notions and how to measure them.
- 03-toy-models: toy mixtures and abstractions to formalize first.
- 04-proof-sketches: theorem sketches, impossibility examples, and recovery arguments.
- references: stable links to the main repo sources and adjacent notes.
Current Framing
This thread is related to the disentanglement methodology notes and to the active exploration of why latent usefulness does not automatically imply global strategy semantics.
Relevant context:
- methodology/cvae-disentangling
- methodology/concepts/token_weighted_excess_reconstruction
- methodology/cvae-objective-and-losses
path:.tasks/doing/2026-03-15-theory-latent-space-exploration/task.md- ideation/concepts/abstract-strategy-vs-solution
Update Rules
- Keep the core question explicit: learnability versus identifiability.
- Separate task-wise recovery from globally consistent recovery.
- Prefer definitions that can later be tied to empirical diagnostics in the
synthetic
exp2setting. - Put theorem-level ideas in
04-proof-sketches.mdonly when they clarify the definitions, not as a substitute for defining the problem.