← Back to papers

Paper deep dive

The Epistemic Support-Point Filter: Jaynesian Maximum Entropy Meets Popperian Falsification

Moriba Kemessia Jah

Year: 2026Venue: arXiv preprintArea: cs.ITType: PreprintEmbeddings: 73

Abstract

Abstract:This paper proves that the Epistemic Support-Point Filter (ESPF) is the unique optimal recursive estimator within the class of epistemically admissible evidence-only filters. Where Bayesian filters minimize mean squared error and are driven toward an assumed truth, the ESPF minimizes maximum entropy and surfaces what has not been proven impossible -- a fundamentally different epistemic commitment with fundamentally different failure modes. Two results locate this theorem within the broader landscape of estimation theory. The first is a unification: the ESPF's optimality criterion is the log-geometric mean of the alpha-cut volume family in the Holder mean hierarchy. The Popperian minimax bound and the Kalman MMSE criterion occupy the p=+inf and p=0 positions on the same curve. Possibility and probability are not competing frameworks: they are the same ignorance functional evaluated under different alpha-cut geometries. The Kalman filter is the Gaussian specialization of the ESPF's optimality criterion, not a separate invention. The second result is a diagnostic: numerical validation over a 2-day, 877-step Smolyak Level-3 orbital tracking run shows that possibilistic stress manifests through necessity saturation and surprisal escalation rather than MVEE sign change -- a direct consequence of the Holder ordering, not an empirical observation. Three lemmas establish the result: the Possibilistic Entropy Lemma decomposes the ignorance functional; the Possibilistic Cramer-Rao Bound limits entropy reduction per measurement; the Evidence-Optimality Lemma proves minimum-q selection is the unique minimizer and that any rule incorporating prior possibility risks race-to-bottom bias.

Tags

ai-safety (imported, 100%)csit (suggested, 92%)preprint (suggested, 88%)

Links

PDF not stored locally. Use the link above to view on the source site.

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%

Last extracted: 3/13/2026, 12:57:54 AM

Summary

The paper introduces the Epistemic Support-Point Filter (ESPF), a recursive estimator that minimizes maximum entropy to perform impossibility-elimination rather than truth-seeking. It unifies Popperian falsification and Jaynesian maximum entropy, proving that the ESPF is the unique optimal filter within the class of epistemically admissible evidence-only filters. The optimality criterion is identified as the log-geometric mean of the alpha-cut volume family in the Holder mean hierarchy, with the Kalman filter emerging as a Gaussian specialization.

Entities (5)

Epistemic Support-Point Filter · algorithm · 100%Kalman Filter · algorithm · 100%Popperian falsification · epistemological-principle · 95%Possibilistic Entropy · mathematical-concept · 95%Holder mean hierarchy · mathematical-framework · 90%

Relation Signals (3)

Possibilistic Entropy isoptimalitycriterionfor ESPF

confidence 95% · the possibilistic entropy Hπ — the ESPF’s optimality criterion

ESPF minimizes Maximum Entropy

confidence 95% · the ESPF minimizes maximum entropy and surfaces what has not been proven impossible

Kalman Filter isspecializationof ESPF

confidence 90% · The Kalman filter is the Gaussian specialization of the ESPF’s optimality criterion

Cypher Suggestions (2)

Find all algorithms related to the ESPF and their specific roles. · confidence 90% · unvalidated

MATCH (a:Algorithm)-[r]->(b:Algorithm) WHERE a.name = 'ESPF' OR b.name = 'ESPF' RETURN a, r, b

Map the relationship between mathematical concepts and the ESPF. · confidence 85% · unvalidated

MATCH (e:Entity {entity_type: 'Mathematical Concept'})-[r:IS_OPTIMALITY_CRITERION_FOR]->(a:Algorithm {name: 'ESPF'}) RETURN e, r

Full Text

73,196 characters extracted from source content.

Expand or collapse full text

The Epistemic Support-Point Filter: Jaynesian Maximum Entropy Meets Popperian Falsification A Possibilistic Minimax-Entropy Optimality Proof Moriba Kemessia Jah, Ph.D. 1Black Swan Research Group, GaiaVerse, Ltd. 2Jah Decision Intelligence Group, Aerospace Engineering & Engineering Mechanics, The University of Texas at Austin (March 9, 2026) Abstract This paper proves that the Epistemic Support-Point Filter (ESPF) is the unique optimal recursive estimator within the class of epistemically admissible evidence-only filters. The optimality criterion is possibilistic minimax entropy: among all evidence-only survivor selection rules, the ESPF minimizes worst-case integrated ignorance per measurement step, and no other admissible filter can do better. The proof is constructive: it identifies the precise mathematical form of the epistemological commitment — be quick to embrace ignorance, be slow to assert certainty — and shows that the ESPF is its unique implementation. Where Bayesian filters minimize mean squared error and are driven toward an assumed truth, the ESPF minimizes maximum entropy and surfaces what has not been proven impossible — a fundamentally different epistemic commitment with fundamentally different failure modes. Two results locate this theorem within the broader landscape of estimation theory. The first is a unification. The possibilistic entropy Hπ=∫01log⁡Vα​d​αH_π= _0^1 V_α\,dα — the ESPF’s optimality criterion — is the log-geometric mean of the α-cut volume family Vα\V_α\, i.e., log⁡M0​(Vα) M_0(\V_α\) in the Hölder mean hierarchy. The Popperian minimax bound log​det(MVEE) (MVEE) is log⁡M+∞ M_+∞; the Kalman MMSE criterion is log⁡M0 M_0 evaluated under Gaussian α-cut geometry. Possibility theory and probability theory are not competing frameworks: they are evaluations of the same ignorance functional log⁡M0​(Vα) M_0(\V_α\) under different geometries of α-cuts. The Kalman filter is the Gaussian specialization of the ESPF’s optimality criterion, not a separate invention. The second is a diagnostic. Numerical validation over a 2-day, 877-step Level-3 Smolyak run (M=106M=106 support points, n=7n=7) under nominal LEO tracking and a sustained stress case (10 m/s crosstrack maneuver combined with a 20 m range bias) reveals that possibilistic stress manifests primarily through necessity saturation and surprisal escalation — not through MVEE sign change. The MVEE regime flag, operating at p=+∞p=+∞ in the Hölder hierarchy, is the coarsest instrument in the Epistemic Width Monitor (EWM) suite. Necessity and surprisal, operating closer to p=0p=0, provide earlier and more sensitive warning of model–reality divergence. This is not an empirical observation: it is a direct consequence of the Hölder ordering M0≤M+∞M_0≤ M_+∞. The theorem rests on three lemmas. The Possibilistic Entropy Lemma decomposes HπH_π into a minimax term (minimized by Popperian survivor selection) and a gradient term (minimized by compatibility-based possibility assignment). The Possibilistic Cramér–Rao Bound (PCRB) establishes the maximum rate at which any admissible filter may reduce HπH_π per measurement. The Evidence-Optimality Lemma proves that minimum-q selection is the unique minimizer of HπH_π within the evidence-only class, and that any rule incorporating prior possibility risks race-to-bottom bias and is strictly suboptimal. Keywords: Possibility theory; Possibilistic entropy; Minimax entropy; α-cut; Popperian falsification; Jaynesian maximum entropy; Possibilistic Cramér–Rao bound; Non-Bayesian estimation; Kalman filter generalization; Epistemic Width Monitor. Contents 1 Introduction 1.1 Two Principles, One Filter 1.2 The Minimax Entropy Principle 1.3 The Two Regimes 1.4 The Correct Entropy Functional 1.5 Structure of the Paper 2 Setup 2.1 Possibility Distributions, α-Cuts, and Possibilistic Entropy 2.2 Epistemically Admissible Filters 2.3 Innovation Geometry and Possibilistic Information Content 3 Three Lemmas 3.1 Lemma 1: Possibilistic Entropy as Ignorance Functional 3.2 Lemma 2: Possibilistic Cramér–Rao Bound 3.3 Lemma 3: Minimax Entropy and Evidence-Only Selection 4 Main Theorem 4.1 Three Sources of Ignorance, One Optimal Strategy 5 The Gaussian Limit: Recovery of the Kalman Filter 6 Numerical Validation 6.1 Setup 6.2 Smolyak Level 3: Nominal Run 6.3 Smolyak Level 3: Stress Run 6.4 EWM Diagnostic Findings 6.5 Tabular Validation (52 Steps) 7 Discussion 7.1 What the Theorem Establishes and What Remains Open 7.2 The Unique Role of the Possibility Gradient 7.3 Comparison to Existing Optimality Results 7.4 Implications for Filter Design 7.5 Beyond Filtering 8 Conclusion References 1 Introduction 1.1 Two Principles, One Filter There is a tension at the heart of recursive state estimation. On one side stands E. T. Jaynes [15]: when you do not know something, assume the maximum entropy consistent with what you do know. Assert nothing you have not earned. On the other side stands Karl Popper [16]: when evidence arrives, eliminate every hypothesis the evidence rules out. Keep only what the data has not falsified. These two principles appear to pull in opposite directions. Jaynes says spread; Popper says cut. Yet they are not in conflict. They operate at different temporal phases of a single recursive epistemology. Before a measurement, Jaynes governs: propagate the possibility distribution forward under maximal ignorance. After a measurement, Popper governs: eliminate hypotheses incompatible with the evidence, using the evidence alone. The filter that implements this two-phase epistemology is the ESPF. The design intent can be stated simply: be quick to embrace ignorance and slow to assert certainty. This paper proves that this intent is a theorem. The ESPF differs from Bayesian filters not merely in criterion but in purpose. A Bayesian filter is truth-seeking: it assumes a generative model, integrates over a prior, and produces a posterior whose objective is convergence to the true state in expectation. The prior has a vote at every step. When the prior is wrong, the filter is driven toward a false truth — and may never recover, because each update reinforces the prior’s influence on the next. The ESPF is not truth-seeking. It is impossibility-eliminating. It makes no claim about where truth is. It removes only what the evidence has ruled out; what remains is not an estimate of truth but the set of everything that has not yet been proven impossible. A wrong prior assignment in the ESPF cannot bias the filter toward a false truth — it can only slow the elimination of inadmissible hypotheses, which manifests immediately as excess epistemic width in the EWM. This is the difference between an epistemically defensible recursion and one that risks compounding prior errors. 1.2 The Minimax Entropy Principle The key to making the Popperian update optimal is identifying the right criterion. The right criterion is not minimizing expected entropy, or minimizing posterior-weighted entropy, or any criterion that mixes evidence with prior belief. The right criterion is minimax entropy: among all evidence-only selection rules, choose the one that minimizes the worst-case residual ignorance. The worst-case ignorance after retaining a survivor set J is log​det(MVEE​(χ(i):i∈J)) (MVEE(\χ^(i):i∈ J\))—the log-volume of the outermost α-cut of the surviving possibility distribution. This is the entropy of the most pessimistic agent who considers all survivors equally plausible. Minimizing it yields the set of NtargetN_target survivors whose convex hull is as small as possible: the hypotheses with smallest whitened squared innovations qk(i)q^(i)_k, geometrically closest to the evidence-consistent locus. This is exactly the ESPF’s selection rule. 1.3 The Two Regimes The Jaynesian-Popperian synthesis has a natural regime structure. Falsification regime (log​det(MVEE)<0 (MVEE)<0): the possibility support occupies a volume smaller than the unit sphere in innovation-normalized coordinates. The Popperian principle is operative and the minimax entropy theorem applies. Diffusion regime (log​det(MVEE)≥0 (MVEE)≥ 0): the support has spread beyond the unit sphere due to process noise accumulation. The filter maintains coverage of the reachable state space; the minimax entropy criterion does not apply. This is the Jaynesian phase, directly analogous to the Kalman filter’s process-noise prediction step during which the measurement-optimality criterion is inoperative. The regime boundary at log​det(MVEE)=0 (MVEE)=0 is a real-time diagnostic indicator of which epistemological principle governs the current step. 1.4 The Correct Entropy Functional For a uniform possibility distribution, log-volume suffices. But the ESPF’s distribution is generically non-uniform after the first update: the conjunctive update assigns each survivor π~k(i)=min⁡(πk|k−1(i),Compk(i)) π^(i)_k= (π^(i)_k|k-1,Comp^(i)_k), varying across survivors according to their evidence compatibility. The correct functional is the possibilistic entropy HπH_π, defined as the integral of log⁡Vα V_α over all α-levels. It decomposes into a minimax term (the α→0α→ 0 limit, recovering MVEE log-det) and a gradient term (negative, measuring how sharply possibility is concentrated within the admissible set). The ESPF minimizes both simultaneously. 1.5 Structure of the Paper Section 2 establishes the framework. Section 3 proves the three core lemmas. Section 4 states and proves the main optimality theorem. Section 5 establishes the Kalman filter as the Gaussian limit. Section 6 presents numerical validation at Smolyak Level 3 over a 2-day orbital tracking run including the full EWM diagnostic suite. Section 7 discusses implications. 2 Setup 2.1 Possibility Distributions, α-Cuts, and Possibilistic Entropy Definition 2.1 (Possibility distribution and α-cuts). A possibility distribution over a finite support χ(i)i=1M⊂ℝn\χ^(i)\^M_i=1 ^n is a function π:1,…,M→(0,1]π:\1,…,M\→(0,1] with maxi⁡π(i)=1 _iπ^(i)=1. The α-cut at level α∈(0,1]α∈(0,1] is Cα=i:π(i)≥α.C_α=\i:π^(i)≥α\. (1) The α-cuts are nested: Cα′⊆CαC_α C_α for α′>α >α. The volume of the α-cut is Vα=cn​(detΠα)1/2V_α=c_n( _α)^1/2, where Πα≻0 _α 0 is the MVEE shape matrix of χ(i):i∈Cα\χ^(i):i∈ C_α\ and cn=πn/2/Γ​(n/2+1)c_n=π^n/2/ (n/2+1). When |Cα|<2​n+1|C_α|<2n+1, the MVEE is degenerate; we set Vα=0V_α=0 by convention. The threshold 2​n+12n+1 reflects the requirement that the point cloud span ℝnR^n with antipodal coverage in each coordinate direction—the minimum structure for a full-rank MVEE shape matrix Πα _α. This is consistent with the filter’s minimum survivor count Nmin=2​n+1N_ =2n+1 (Section 7), making NminN_ a direct consequence of HπH_π’s well-posedness rather than an independent design choice. Definition 2.2 (Possibilistic entropy). The possibilistic entropy of a possibility distribution π over χ(i)i=1M\χ^(i)\^M_i=1 is Hπ=∫01log⁡Vα​d​α=∫01(12​log​detΠα+κn)​α,H_π= _0^1 V_α\,dα= _0^1 ( 12 _α+ _n )dα, (2) where κn=log⁡cn _n= c_n. Remark 2.3 (Uniform special case). When π(i)=1π^(i)=1 for all i, Cα=1,…,MC_α=\1,…,M\ for all α∈(0,1]α∈(0,1] and Vα=V_α=V is constant. Then Hπ=log⁡V+κnH_π= V+ _n, recovering log-volume exactly. The possibilistic entropy is a strict generalization: it equals log-volume for uniform distributions and is strictly smaller when possibility mass is concentrated on a subset of the support. Remark 2.4 (Relation to U-uncertainty and origin of the logarithm). The possibilistic entropy is the geometric analog of the U-uncertainty measure introduced by Higashi and Klir [7] for discrete possibility distributions on ordered sets. Here the ordering is geometric: the α-cuts are nested convex bodies in ℝnR^n, and the integral of log-cardinality over α is replaced by the integral of log-volume over α. The logarithm is justified by the Boltzmann tradition: log-volume is additive across independent dimensions of state space. If uncertainty in position and velocity are geometrically independent, the volume of the joint uncertainty ellipsoid is the product of the marginal volumes, and the logarithm converts that product to a sum. The probabilistic route via Shannon’s −log⁡p- p does not apply here because possibility theory uses the min rule for conjunction—π​(A∩B)=min⁡(πA,πB)π(A∩ B)= ( _A, _B)—under which −log- is not additive. Definition 2.5 (ESPF possibility assignment). At update step k, the ESPF assigns possibility values to survivors as follows. For each surviving hypothesis χ(i)χ^(i) with pre-update possibility πk|k−1(i)π^(i)_k|k-1, the post-update (unnormalized) possibility is π^k(i)=min⁡(πk|k−1(i),Compk(i)),Compk(i)=exp⁡(−12​qk(i)), π^(i)_k= \! (π^(i)_k|k-1,\,Comp^(i)_k ), ^(i)_k= \! (- 12q^(i)_k ), (3) where qk(i)q^(i)_k is the whitened squared innovation of hypothesis i (Definition 2.9). After max-normalization, π~k(i)=π^k(i)/maxj⁡π^k(j) π^(i)_k= π^(i)_k/ _j π^(j)_k. Regenerated support points χnew(i)\χ^(i)_new\ inherit possibility via max-min kernel extension: π~k(new,i)=maxj∈Jsurv⁡min⁡(π~k(j),κ​(χ(new,i),χ(j))), π^(new,i)_k= _j∈ J_surv \! ( π^(j)_k,\,κ(χ^(new,i),χ^(j)) ), (4) where κ​(⋅,⋅)∈[0,1]κ(·,·)∈[0,1] is a proximity kernel on state space. Remark 2.6 (Non-uniformity of the ESPF distribution). After the first update, the ESPF’s possibility distribution is non-uniform in general: the anchor hypothesis (most compatible with yky_k) achieves possibility 1, while other survivors achieve strictly smaller values. Uniformity is a transient condition at initialization, not a maintained property. The correct entropy functional for the ESPF at step k≥1k≥ 1 is HπH_π as defined above. 2.2 Epistemically Admissible Filters Definition 2.7 (Epistemically admissible filter). A recursive state estimator is epistemically admissible if at every step k it satisfies: (i) Possibilistic representation. Uncertainty is represented as a normalized possibility distribution πk _k over a finite support χk(i)i=1M⊂ℝn\χ^(i)_k\^M_i=1 ^n with maxi⁡πk(i)=1 _iπ^(i)_k=1. (i) Conjunctive update. Upon receiving evidence yky_k, π^k|k(i)≤πk|k−1(i) π^(i)_k|k≤π^(i)_k|k-1 for all i; no hypothesis may gain possibility from a measurement. (i) Geometric non-degeneracy (VFI). The MVEE shape matrix Πk _k of the full support satisfies λmin​(Πk)≥ε>0 _ ( _k)≥ >0 and λmax​(Πk)≤Λ<∞ _ ( _k)≤ <∞ for all k. (iv) Admissibility of the commitment point. The operational output x^k∈χk(i) x_k∈\χ^(i)_k\. This is not a claim that x^k x_k is truth: it is the least inadmissible hypothesis — the support point most compatible with the accumulated evidence. The ESPF makes no assertion about where truth is; it only identifies what has not yet been ruled out, and among those, which is most consistent with the evidence. (v) Evidence referencing. The survivor selection ranking depends only on the whitened squared innovations qk(i)q^(i)_k, not on πk|k−1(i)π^(i)_k|k-1. (The conjunctive possibility assignment in (i) may involve πk|k−1(i)π^(i)_k|k-1 through the min operation; it is the selection rule that must be evidence-only.) Denote this class ℱF. Remark 2.8 (Why evidence-referencing is the right admissibility condition). Condition (v) is not a convenient restriction — it is the epistemological content of the class. A filter that incorporates prior possibility into its survivor selection ranking is not merely using more information; it is making a different kind of claim. It is asserting that hypotheses which were previously plausible deserve preferential treatment over hypotheses that are more consistent with the current evidence. This is the truth-seeking posture: the prior encodes a belief about where truth is, and that belief is allowed to protect hypotheses from elimination even when the evidence argues against them. The race-to-bottom failure mode (Lemma 3.7, Step 5) is the formal consequence: over multiple steps, prior-based selection concentrates the possibility distribution on hypotheses that are systematically less evidence-consistent. The filter converges — but toward the prior’s truth, not the evidence’s truth. In an impossibility-eliminating filter, this failure mode is not just suboptimal; it is a category error. The prior has no vote on what is impossible. Only the evidence does. Condition (v) enforces this. 2.3 Innovation Geometry and Possibilistic Information Content Definition 2.9 (Innovation geometry). Given predicted support χk|k−1(i)\χ^(i)_k|k-1\ and measurement model h:ℝn→ℝmh:R^n ^m, the innovation shape matrix Πe≻0 _e 0 is formed from the MVEE of the predicted measurement support and the sensor imprecision matrix Πy _y. The whitened squared innovation of hypothesis i is qk(i)=‖Le−1​(yk−h​(χk|k−1(i)))‖22,Πe=Le​Le⊤.q^(i)_k= \|L_e^-1 (y_k-h(χ^(i)_k|k-1) ) \|_2^2, _e=L_eL_e . (5) The compatibility of hypothesis i with yky_k is Compk(i)=exp⁡(−12​qk(i))∈(0,1]Comp^(i)_k= (- 12q^(i)_k)∈(0,1]. Definition 2.10 (Possibilistic information content). The aggregate epistemic surprisal at step k is the Choquet integral of the per-hypothesis surprisal 12​qk(i) 12q^(i)_k with respect to the prior possibility distribution πk|k−1 _k|k-1 as capacity: S¯k=(C)​∫12​qk(i)​πk|k−1=supi=1,…,Mmin⁡(12​qk(i),πk|k−1(i)). S_k=(C) 12q^(i)_k\,d _k|k-1= _i=1,…,M \! ( 12q^(i)_k,\,π^(i)_k|k-1 ). (6) The second equality is the standard reduction of the Choquet integral under a possibility measure [5]. The capacity is the prior possibility πk|k−1 _k|k-1, not the current compatibility: this keeps the information content epistemically grounded in accumulated evidence history rather than the current measurement alone, and ensures that hypotheses with low prior credibility cannot inflate S¯k S_k through large but epistemically cheap surprisal. The survivor selection rule (Lemma 3.7) remains pure evidence-only minimum-q; the prior possibility enters only here, in the bounding quantity. The possibilistic information content of yky_k is Ik=1−e−S¯k∈[0,1).I_k=1-e^- S_k∈[0,1). (7) Ik=0I_k=0 when the evidence is fully consistent with the predicted support; Ik→1I_k→ 1 when the most epistemically credible hypothesis is maximally surprised by the measurement. Remark 2.11 (Choquet integral under a possibility measure). For a possibility measure Π (max-decomposable, Π​(A∪B)=max⁡(Π​(A),Π​(B)) (A∪ B)= ( (A), (B))), the Choquet integral of a non-negative function f reduces to (C)​∫f​Π=supimin⁡(f​(i),π​(i))(C) f\,d = _i (f(i),π(i)) [5]. This is the possibilistic expectation in the sense of Dubois and Prade: it asks for the highest level t at which there exists a hypothesis that is both highly credible (π(i)≥tπ^(i)≥ t) and genuinely surprised (f​(i)≥tf(i)≥ t). It is robust to outliers by a different mechanism than the median: low-prior-possibility hypotheses are downweighted by the capacity function rather than by a rank cutoff, which is the epistemically correct weighting in a possibilistic framework. 3 Three Lemmas 3.1 Lemma 1: Possibilistic Entropy as Ignorance Functional Lemma 3.1 (Possibilistic entropy as ignorance). Let π be a normalized possibility distribution over χ(i)i=1M⊂ℝn\χ^(i)\^M_i=1 ^n with π(i)∈(0,1]π^(i)∈(0,1] and maxi⁡π(i)=1 _iπ^(i)=1. The possibilistic entropy HπH_π satisfies: (i) Uniform reduction. When π(i)=1π^(i)=1 for all i, Hπ=log⁡V+κnH_π= V+ _n where V=cn​(detΠ)1/2V=c_n( )^1/2 is the MVEE volume of the full support. (i) Monotone in concentration. For two distributions π,π′π,π over the same support with π′⁣(i)≤π(i)π (i)≤π^(i) for all i and strict inequality for some i∗i^*, we have Hπ′<HπH_π <H_π. (i) Decomposition. Hπ=log⁡V+κn⏟support entropy+∫01log⁡VαV​d​α⏟gradient entropy≤0.H_π= V+ _n_support entropy+ _0^1 V_αV\,dα_gradient entropy≤ 0. (8) The gradient entropy is non-positive; it equals zero when π is uniform and is strictly negative otherwise. (iv) Extremes. HπH_π is maximized when π is uniform (Hπ=log⁡V+κnH_π= V+ _n) and approaches −∞-∞ as possibility collapses to a single hypothesis. The VFI condition prevents collapse and keeps HπH_π finite within ℱF. Proof. (i) When all π(i)=1π^(i)=1, Cα=1,…,MC_α=\1,…,M\ for all α, so Vα=V_α=V is constant and Hπ=∫01log⁡V​d​α=log⁡VH_π= _0^1 V\,dα= V. (i) If π′⁣(i)≤π(i)π (i)≤π^(i) with strict inequality at i∗i^*, then Cα′⊆CαC _α C_α for all α, with strict inclusion at α=π(i∗)α=π^(i^*). By monotonicity of the MVEE (adding points cannot decrease volume [10]), Vα′≤VαV _α≤ V_α with strict inequality at α=π(i∗)α=π^(i^*). Therefore Hπ′=∫01log⁡Vα′​d​α<∫01log⁡Vα​d​α=HπH_π = _0^1 V _α\,dα< _0^1 V_α\,dα=H_π. (i) Write log⁡Vα=log⁡V+log⁡(Vα/V) V_α= V+ (V_α/V). Since Cα⊆1,…,MC_α \1,…,M\ for all α, we have Vα≤V_α≤ V and log⁡(Vα/V)≤0 (V_α/V)≤ 0. Integrating over α∈[0,1]α∈[0,1] gives the decomposition. (iv) Collapse to a singleton gives |Cα|=1<2​n+1|C_α|=1<2n+1 for all α>0α>0, so Vα=0V_α=0 by convention and Hπ→−∞H_π→-∞. The upper bound follows from (i). ∎ Remark 3.2 (Physical interpretation of the decomposition). The support entropy log⁡V V measures how large a region of state space remains admissible—the extent of ignorance. The gradient entropy measures how sharp or flat the possibility gradient is within that region—the texture of ignorance. A filter that reduces the support but leaves possibility flat across it has reduced the first term but not the second. The ESPF reduces both simultaneously: it prunes the support (reducing V) and assigns higher possibility to more evidence-consistent hypotheses (reducing the gradient entropy). This is why HπH_π is the correct objective. 3.2 Lemma 2: Possibilistic Cramér–Rao Bound Lemma 3.3 (Possibilistic Cramér–Rao Bound (PCRB)). Let F∈ℱF be any epistemically admissible filter. At step k, let Hπ,k|k−1H_π,k|k-1 and Hπ,k|kH_π,k|k denote the pre- and post-update possibilistic entropies, and IkI_k the possibilistic information content. Then Hπ,k|k≥Hπ,k|k−1+n2​log⁡(1−Ik).H_π,k|k≥ H_π,k|k-1+ n2 (1-I_k). (9) The bound holds for any epistemically admissible filter regardless of its internal implementation. Proof. Step 1: Conjunctive update implies nested post-update α-cuts. By the admissibility definition, π^k|k(i)≤πk|k−1(i) π^(i)_k|k≤π^(i)_k|k-1 for all i, so Cαpost⊆CαpreC^post_α C^pre_α for all α and Vαpost≤VαpreV^post_α≤ V^pre_α. Step 2: Choquet information content bounds the possibility measure of the surviving set. Let i∗i^* be the hypothesis achieving the supremum in S¯k S_k, so that min⁡(12​qk(i∗),πk|k−1(i∗))=S¯k. \! ( 12q^(i^*)_k,\,π^(i^*)_k|k-1 )= S_k. This means πk|k−1(i∗)≥S¯kπ^(i^*)_k|k-1≥ S_k and Compk(i∗)=e−12​qk(i∗)≥e−S¯k=1−IkComp^(i^*)_k=e^- 12q^(i^*)_k≥ e^- S_k=1-I_k. Under the conjunctive update, the post-update possibility of i∗i^* satisfies π^k|k(i∗)=min⁡(πk|k−1(i∗),Compk(i∗))≥min⁡(S¯k,1−Ik)>0. π^(i^*)_k|k= \! (π^(i^*)_k|k-1,\,Comp^(i^*)_k )≥ ( S_k,1-I_k)>0. Hence i∗∈Cαposti^*∈ C^post_α for all α≤min⁡(S¯k,1−Ik)α≤ ( S_k,1-I_k), and the possibility measure of the surviving set satisfies Πk|k−1​(Cαpost)≥πk|k−1(i∗)≥S¯k _k|k-1(C^post_α)≥π^(i^*)_k|k-1≥ S_k for all α≤1−Ikα≤ 1-I_k. This replaces a cardinality guarantee with an epistemic-weight guarantee: the surviving set retains possibility measure at least S¯k S_k, meaning the most credible hypothesis under the prior has not been eliminated. Step 3: Possibility measure bound implies volume bound. For α≤1−Ikα≤ 1-I_k, the post-update α-cut retains i∗i^*, which satisfies Compk(i∗)≥1−IkComp^(i^*)_k≥ 1-I_k. By the VFI condition the pre-update MVEE is non-degenerate with λmin​(Πpre)≥ε>0 _ ( ^pre)≥ >0, and i∗i^* lies within it. The set of hypotheses satisfying Compk(i)≥αComp^(i)_k≥α — i.e., 12​qk(i)≤−log⁡α 12q^(i)_k≤- α — lies within a whitened-innovation ball of radius rα=−2​log⁡αr_α= -2 α centered on the evidence-consistent locus χ∗χ^*. Under the isotropy condition Varx∼U​(Sk|k−1)​[‖Le−1​(yk−h​(x))‖22]≤Varx∼U​(Sk|k−1)​[‖Lcore−1​(x−x¯)‖22],Var_x U(S_k|k-1)\! [ \|L_e^-1(y_k-h(x)) \|_2^2 ] _x U(S_k|k-1)\! [ \|L_core^-1(x- x) \|_2^2 ], (10) this ball maps to a state-space region whose John ellipsoid [9] has shape matrix ΠαJ ^J_α satisfying ΠαJ⪰1n​(1−Ik)​Παpre, ^J_α 1n(1-I_k)\, ^pre_α, by the John ellipsoid theorem applied to the subset containment Cαpost⊆CαpreC^post_α C^pre_α, with the compatibility threshold of i∗i^* providing the (1−Ik)(1-I_k) scaling on the semi-axes. Taking determinants and using det(1n​A)=n−n​detA ( 1nA)=n^-n A together with the volume formula Vα=cn​(detΠα)1/2V_α=c_n( _α)^1/2 yields Vαpost≥n−n/2​(1−Ik)1/2⋅Vαprefor all ​α≤1−Ik.V^post_α≥ n^-n/2(1-I_k)^1/2· V^pre_α all α≤ 1-I_k. (11) The n−n/2n^-n/2 factor is a fixed dimensional constant absorbed into the κn _n term of HπH_π; the (1−Ik)1/2(1-I_k)^1/2 factor becomes (1−Ik)n/2(1-I_k)^n/2 after the full n-dimensional determinant scaling in Step 4, yielding the bound stated there. Step 4: Volume bound implies entropy bound. Decompose the integral as Hπ,k|k=∫01−Iklog⁡Vαpost​d​α+∫1−Ik1log⁡Vαpost​d​α.H_π,k|k= _0^1-I_k V^post_α\,dα+ _1-I_k^1 V^post_α\,dα. For the first integral, applying the volume bound: ∫01−Iklog⁡Vαpost​d​α≥∫01−Ik(log⁡(1−Ik)+log⁡Vαpre)​α. _0^1-I_k V^post_α\,dα≥ _0^1-I_k\! ( (1-I_k)+ V^pre_α )\,dα. For the second integral, Vαpost≥0V^post_α≥ 0 and the interval has measure IkI_k; in the falsification regime the α-cut volumes are finite and this term is absorbed by the n/2n/2 scaling factor when passing from the scalar bound to the n-dimensional volume scaling (1−Ik)↦(1−Ik)n/2(1-I_k) (1-I_k)^n/2 on determinants. Combining both terms and applying the n-dimensional scaling log⁡Vα=12​log​detΠα+κn V_α= 12 _α+ _n, which introduces a factor of n/2n/2 when translating a scalar compatibility bound to detΠαpost≥(1−Ik)n​detΠαpre ^post_α≥(1-I_k)^n ^pre_α, yields Hπ,k|k≥Hπ,k|k−1+n2​log⁡(1−Ik),H_π,k|k≥ H_π,k|k-1+ n2 (1-I_k), (12) which is the stated bound. ∎ Remark 3.4 (The PCRB is a bound on integrated ignorance reduction). The PCRB bounds the integral of log⁡Vα V_α over all levels simultaneously, not a single α-cut. An admissible filter may reduce VαV_α sharply at some levels while barely changing others, but the net reduction in HπH_π cannot exceed n2log(1−Ik)−1 n2 (1-I_k)^-1. This is stronger than any pointwise volume bound. Remark 3.5 (Separation of selection and bounding). The Choquet-based IkI_k and the minimum-q selection rule are deliberately decoupled. The prior possibility πk|k−1 _k|k-1 enters only in the capacity of the Choquet integral defining S¯k S_k—a bounding quantity, not a selection criterion. The survivor ranking in Lemma 3.7 depends solely on qk(i)q^(i)_k. This separation preserves the Popperian character of the ESPF: selection is evidence-only, but the rate at which entropy may decrease is governed by the epistemically weighted surprisal of the current measurement. A measurement that surprises only low-prior-possibility hypotheses yields a small IkI_k and a tight floor; one that surprises the most credible hypothesis yields a large IkI_k and a loose floor—the correct behavior in both cases. Remark 3.6 (Clarification on Step 3). The volume bound relies on the isotropy condition (10) for the nonlinear case; for linear measurement models it holds exactly. For general non-uniform possibility distributions the bound holds with (1−Ik)n/2(1-I_k)^n/2 rather than (1−Ik)(1-I_k) on the right-hand side due to the John ellipsoid theorem; the final entropy bound is unchanged. 3.3 Lemma 3: Minimax Entropy and Evidence-Only Selection Lemma 3.7 (Minimax evidence-optimality). Let χ(i)i=1M\χ^(i)\^M_i=1 be the predicted support with whitened squared innovations qk(i)q^(i)_k, and let Ntarget≤MN_target≤ M be a fixed survivor count. Consider all pairs (J,ρ)(J,ρ) where J⊆1,…,MJ \1,…,M\ with |J|=Ntarget|J|=N_target and ρ:J→(0,1]ρ:J→(0,1] with maxi∈J⁡ρ(i)=1 _i∈ Jρ^(i)=1, within the evidence-only comparison class: all (J,ρ)(J,ρ) whose selection and ranking depend only on qk(i)q^(i)_k, not on πk|k−1(i)π^(i)_k|k-1. The unique minimizer of H​(J,ρ)=∫01log⁡VαJ​d​α,VαJ=cn​(detΠαJ)1/2,H(J,ρ)= _0^1 V^J_α\,dα, V^J_α=c_n ( ^J_α )^1/2, (13) where ΠαJ ^J_α is the MVEE shape matrix of χ(i):i∈J,ρ(i)≥α\χ^(i):i∈ J,\,ρ^(i)≥α\, is the ESPF’s choice: J∗=i:qk(i)​ is among the ​Ntarget​ smallest,ρ∗​(i)=Compk(i)=e−qk(i)/2,J^*= \i:q^(i)_k is among the N_target smallest \, ρ^*(i)=Comp^(i)_k=e^-q^(i)_k/2, (14) for linear measurement models exactly, and under the innovation-state isotropy condition (10) for nonlinear models. Equivalently, J∗J^* is the unique minimizer of log​det(MVEE​(χ(i):i∈J)) (MVEE(\χ^(i):i∈ J\) ) within the evidence-only class. Proof. We show (J∗,ρ∗)(J^*,ρ^*) minimizes H​(J,ρ)H(J,ρ) by showing it simultaneously minimizes VαJV^J_α for every α∈(0,1]α∈(0,1]. Step 1: α-cut structure under ρ∗ρ^*. Under ρ∗​(i)=e−qk(i)/2ρ^*(i)=e^-q^(i)_k/2, higher possibility corresponds to lower qk(i)q^(i)_k. For any threshold α, Cα​(J∗,ρ∗)=i∈J∗:e−qk(i)/2≥α=i∈J∗:qk(i)≤−2​log⁡α.C_α(J^*,ρ^*)= \i∈ J^*:e^-q^(i)_k/2≥α \= \i∈ J^*:q^(i)_k≤-2 α \. At α→0α→ 0 this is the full survivor set J∗J^*. Step 2: Minimum-q selection minimizes the minimax term (linear case). For linear h​(x)=H​x+bh(x)=Hx+b, qk(i)=‖Le−1​H​(χ(i)−χ∗)‖22+const,q^(i)_k= \|L_e^-1H(χ^(i)-χ^*) \|_2^2+const, where χ∗=H†​(yk−b)χ^*=H (y_k-b) is the evidence-consistent locus. Hence qk(i)q^(i)_k measures each hypothesis’s distance to χ∗χ^* in the innovation-whitened metric. Among all NtargetN_target-element subsets of χ(i)i=1M\χ^(i)\^M_i=1, the set geometrically closest to χ∗χ^* has the smallest MVEE. This follows from the convexity of log​det(⋅) (·) on the positive-definite cone [10]: for any set J containing a hypothesis j with qk(j)>qk(i∗)q^(j)_k>q^(i^*)_k for some i∗∉Ji^*∉ J, replacing j with i∗i^* (moving the set closer to χ∗χ^*) does not increase the MVEE determinant, and strictly decreases it for generic (non-degenerate) point clouds. Therefore J∗J^* minimizes log​det(MVEE​(χ(i):i∈J)) (MVEE(\χ^(i):i∈ J\)) among all NtargetN_target-element subsets. The same reasoning applies at every α-level: for each cardinality kαk_α, the kαk_α smallest-q hypotheses minimize VαJV^J_α. Hence Cα​(J∗,ρ∗)C_α(J^*,ρ^*) minimizes VαJV^J_α for every α∈(0,1]α∈(0,1]. Step 3: Pointwise minimum volume implies minimum HπH_π. Since VαJ∗≤VαJV^J^*_α≤ V^J_α for all α and all alternative (J,ρ)(J,ρ): H​(J∗,ρ∗)=∫01log⁡VαJ∗​d​α≤∫01log⁡VαJ​d​α=H​(J,ρ).H(J^*,ρ^*)= _0^1 V^J^*_α\,dα≤ _0^1 V^J_α\,dα=H(J,ρ). Step 4: No alternative possibility assignment does better. Suppose we keep J∗J^* but replace ρ∗ρ^* with some ρ′ρ . By the monotone concentration property of HπH_π, any assignment placing higher possibility on a higher-q hypothesis increases the volume of at least one α-cut and thus increases HπH_π. The assignment ρ∗ρ^* is the unique minimizer among all possibility assignments on J∗J^* because ordering by evidence consistency simultaneously minimizes all α-cut volumes. Step 5 (Race-to-bottom): The prior must not influence selection. Any selection rule incorporating πk|k−1(i)π^(i)_k|k-1 into the survivor ranking may retain a high-q hypothesis at the expense of a low-q hypothesis. By Step 2, retaining a higher-q hypothesis increases VαJV^J_α at the corresponding α-level, increasing HπH_π. Over multiple steps, prior-based selection compounds: a hypothesis that accumulates high prior possibility may be repeatedly protected from evidence-based elimination, concentrating the possibility distribution on hypotheses that are systematically less evidence-consistent. This is the race-to-bottom failure mode—the possibility gradient is driven by self-reinforcing prior protection rather than evidence. The evidence-only class excludes this failure mode by construction. No rule within the evidence-only class that differs from minimum-q achieves lower HπH_π. Uniqueness. For a generic point cloud (no ties in qk(i)q^(i)_k, holding almost surely for continuous dynamics), J∗J^* is unique and ρ∗ρ^* is determined pointwise by Compk(i)Comp^(i)_k. Nonlinear case. For nonlinear h, condition (10) ensures that the ordering of hypotheses by qk(i)q^(i)_k is a consistent proxy for ordering by distance to the evidence-consistent region. Under this condition Steps 2–4 hold for the nonlinear case via first-order Taylor approximation. The condition can fail when the measurement geometry is strongly anisotropic relative to the state-space geometry—for example, when a single range measurement is made from a nearly tangential viewing angle, so that large state displacements along the line of sight produce small innovation changes. Proposition 3.10 establishes this condition for the orbital mechanics setting. ∎ Remark 3.8 (Minimax vs. expected entropy). The minimax entropy criterion differs fundamentally from minimizing expected or posterior-weighted entropy. A rule minimizing expected entropy may select prior-consistent but evidence-inconsistent hypotheses in exchange for a tighter average-case bound. The minimax criterion refuses this trade: the worst-case agent—who considers all survivors equally plausible—must see the minimum possible ignorance. This is the Popperian content of the criterion. Remark 3.9 (The kernel extension preserves optimality). After survivor selection and possibility assignment, the regenerated support is formed by spreading new hypotheses around the survivors via the max-min kernel extension. The α-cuts of the regenerated distribution are geometric dilations of the survivor α-cuts (in the sense that they contain the survivor α-cuts). Since Lemma 3.7 minimizes the survivor α-cut volumes, and the kernel extension is monotone (larger survivor cuts produce larger regenerated cuts), the regenerated distribution inherits the optimality structure. The possibility gradient is carried forward into the next propagation step. Proposition 3.10 (Isotropy for orbital mechanics). Let the state be the 7-dimensional Cartesian position-velocity-drag vector x=(r,r˙,Cd)∈ℝ7x=(r, r,C_d) ^7, propagated under the two-body problem or its J2J_2-perturbed extension. Let the measurement model be range ρ​(x)=‖r−rs‖ρ(x)=\|r-r_s\| and range-rate ρ˙​(x)=(r−rs)⋅r˙/ρ​(x) ρ(x)=(r-r_s)· r/ρ(x), where rsr_s is the ground station position. Let the observation interval satisfy Δ​t≪T/4 t T/4, where T is the orbital period. Then the innovation-state isotropy condition (10) holds to first order in the linearization residual, except in a measure-zero set of viewing geometries in which the line-of-sight vector (r−rs)/ρ(r-r_s)/ρ is orthogonal to the velocity r˙ r (tangential viewing). Proof sketch. Let H=∂h/∂x|x¯H=∂ h/∂ x|_ x be the measurement Jacobian evaluated at the support centroid. The whitened innovation variance decomposes as: Var​[‖Le−1​(y−h​(x))‖2]≈Var​[‖Le−1​H​(x−x¯)‖2]+O​(‖δ​H‖2),Var\! [ \|L_e^-1(y-h(x)) \|^2 ] \! [ \|L_e^-1H(x- x) \|^2 ]+O(\|δ H\|^2), where δ​Hδ H is the linearization residual. The first term equals tr​(Le−1​H​Πcore​H⊤​Le−⊤)tr(L_e^-1H _coreH L_e^- ), which by the Cauchy–Schwarz inequality on the Frobenius inner product satisfies: tr​(Le−1​H​Πcore​H⊤​Le−⊤)≤‖Le−1​H‖F2⋅λmax​(Πcore).tr(L_e^-1H _coreH L_e^- )≤\|L_e^-1H\|_F^2· _ ( _core). For range and range-rate measurements in LEO, H has rank 2 and ‖Le−1​H‖F2\|L_e^-1H\|_F^2 is bounded by σR−2+σR˙−2 _R^-2+ _ R^-2, which under nominal measurement noise (σR=1 _R=1 m, σR˙=10−5 _ R=10^-5 km/s) and a support cloud spanning tens of kilometers satisfies condition (10) with margin. The bound degrades as (r−rs)/ρ→r˙/‖r˙‖(r-r_s)/ρ→ r/\| r\| (tangential geometry), at which point ∂ρ˙/∂r∂ ρ/∂ r becomes nearly parallel to ∂ρ/∂r∂ρ/∂ r and the rank-2 Jacobian loses its geometric leverage—the failure mode identified in Lemma 3.7. When this occurs, the EWM surprisal indicator escalates in real time, providing an operational signal that the isotropy condition is under stress and that the filter’s Popperian updates should be interpreted with reduced confidence; the filter thus degrades gracefully rather than silently. For Δ​t≪T/4 t T/4, the linearization residual O​(‖δ​H‖2)O(\|δ H\|^2) remains small relative to the leading term, completing the sketch. ∎ 4 Main Theorem Theorem 4.1 (ESPF Optimality: Minimax-Entropy Popperian Falsification). In the falsification regime—measurement steps at which log​det(MVEE​(Sk|k−1))<0 (MVEE(S_k|k-1) )<0, so the evidence contracts the possibility support below the unit-sphere reference volume—the ESPF is the unique filter in the class ℱF of epistemically admissible evidence-only filters that minimizes Hπ,kH_π,k subject to the PCRB constraint Hπ,k|k≥Hπ,k|k−1+n2​log⁡(1−Ik).H_π,k|k≥ H_π,k|k-1+ n2 (1-I_k). (15) Equivalently, the ESPF is the unique filter that: (a) selects survivors by the minimax-entropy criterion within the evidence-only class; and (b) assigns possibilities by evidence compatibility, minimizing the gradient entropy term simultaneously. In the falsification regime, subject to the coverage controller operating at PCRB-implied cardinality (i.e., Ntarget=⌊(1−Ik)⋅M⌋N_target= (1-I_k)· M ) and σk _k at steady state: (i) The ESPF achieves the PCRB with equality. (i) No other filter in ℱF achieves lower Hπ,k|kH_π,k|k while satisfying the PCRB. (i) Under the isotropy condition (10), the ESPF is the unique minimizer. In the diffusion regime—steps at which log​det(MVEE​(Sk|k−1))≥0 (MVEE(S_k|k-1) )≥ 0—the filter operates under Jaynesian maximum entropy: the propagation phase spreads the support as widely as the dynamics and process noise allow. The minimax entropy criterion does not apply in this phase. Proof. (i) ESPF achieves the PCRB. By Lemma 3.7, the ESPF’s joint strategy (J∗,ρ∗)(J^*,ρ^*) minimizes VαJV^J_α at every α-level. The coverage controller sets Ntarget=⌊(1−Ik)⋅M⌋N_target= (1-I_k)· M survivors, retaining the fraction of the support that the evidence permits per the PCRB. At controller steady state, the integrated log-volume satisfies Hπ,k|k=Hπ,k|k−1+n2​log⁡(1−Ik),H_π,k|k=H_π,k|k-1+ n2 (1-I_k), achieving equality in the PCRB. (i) No other filter does better. By the PCRB (Lemma 2), every F′∈ℱF satisfies Hπ,k|k′≥Hπ,k|k−1+n2​log⁡(1−Ik)H _π,k|k≥ H_π,k|k-1+ n2 (1-I_k). Since the ESPF achieves equality, Hπ,k|k′≥Hπ,k|kESPFH _π,k|k≥ H^ESPF_π,k|k for all F′∈ℱF . (i) Uniqueness. By Lemma 3.7, under the isotropy condition the ESPF’s choice (J∗,ρ∗)(J^*,ρ^*) is the unique minimizer of H​(J,ρ)H(J,ρ) almost surely. Any alternative filter achieving equality in the PCRB must implement the same selection and possibility assignment; i.e., it must be the ESPF. Any filter using a different rule achieves Hπ,k|k′>Hπ,k|kESPFH _π,k|k>H^ESPF_π,k|k strictly. ∎ The ESPF Optimality Theorem: Summary Jaynesian phase (propagation): Spread the support as widely as the dynamics allow. Assume maximum ignorance consistent with known constraints. Never assert structure not earned by evidence. Popperian phase (update, falsification regime): • Select survivors by minimum qk(i)q^(i)_k: the evidence-closest hypotheses, uniquely minimizing log​det(MVEE) (MVEE). • Assign possibilities by Compk(i)Comp^(i)_k: evidence ordering simultaneously minimizes the gradient entropy. • Together, these minimize HπH_π at every α-level, saturating the PCRB. • Any rule incorporating prior possibility risks race-to-bottom bias over multiple steps. Regime indicator: log​det(MVEE)<0⇒ (MVEE)<0 falsification (Popper governs). log​det(MVEE)≥0⇒ (MVEE)≥ 0 diffusion (Jaynes governs). 4.1 Three Sources of Ignorance, One Optimal Strategy The HπH_π decomposition reveals three distinct aspects of the ESPF’s ignorance management. Minimax ignorance (log​det(MVEE) (MVEE), the worst-case bound): The largest α-cut tells the most pessimistic story about residual ignorance. The ESPF minimizes this by selecting the evidence-closest survivors. This is the Popperian content. Gradient ignorance (−∫log⁡(Vα/V)​α≥0- (V_α/V)\,dα≥ 0): Even within the admissible set, possibility may be spread flat or sharply concentrated. Flat possibility corresponds to maximal internal ignorance. The compatibility assignment concentrates possibility on the most evidence-consistent hypotheses, reducing this term. Dynamical ignorance (the Jaynesian expansion phase): Between measurements, process noise spreads the support. The ESPF implements this maximally via Smolyak regeneration, maintaining coverage of the reachable state space. 5 The Gaussian Limit: Recovery of the Kalman Filter Definition 5.1 (Gaussian epistemic limit). The Gaussian epistemic limit holds when: (i) Πk→Σk _k→ _k; (i) πk→​(x^k,Σk) _k ( x_k, _k); (i) h​(x)=H​xh(x)=Hx; (iv) f​(x)=F​xf(x)=Fx; (v) Πy=R _y=R. Theorem 5.2 (ESPF recovers Kalman optimality in the Gaussian limit). Under the Gaussian epistemic limit, the ESPF update reduces to the Kalman filter update. In particular, Hπ→12​log​det(2​π​e​Σk|k)H_π→ 12 (2π e\, _k|k) (up to additive constants), and minimizing HπH_π is equivalent to minimizing detΣk|k _k|k. Proof sketch. In the Gaussian limit, the α-cut at level α is the confidence ellipsoid ℰα=x:(x−x^k)⊤​Σk−1​(x−x^k)≤−2​log⁡αE_α=\x:(x- x_k) _k^-1(x- x_k)≤-2 α\, with volume Vα=cn​(−2​log⁡α)n/2​(detΣk)1/2V_α=c_n(-2 α)^n/2( _k)^1/2. Substituting into Hπ=∫01log⁡Vα​d​αH_π= _0^1 V_α\,dα and evaluating the constant integral ∫01log⁡(−2​log⁡α)​α=log⁡2−γ _0^1 (-2 α)\,dα= 2-γ gives Hπ=12​log​detΣk+const​(n)H_π= 12 _k+const(n). Minimizing HπH_π is therefore equivalent to minimizing detΣk|k _k|k, which is the Kalman criterion. Remark 5.3 derives this identity from the Hölder mean structure and shows it is the Gaussian specialization of the general possibilistic functional. ∎ Remark 5.3 (Hölder mean structure and the probability–possibility bridge). The continuous Hölder power mean of the family Vα∈(0,1]\V_α\_α∈(0,1] at order p is log⁡Mp​(Vα)=1p​log​∫01Vαp​α,p≠0, M_p (\V_α\ )= 1p _0^1V_α^p\,dα, p≠ 0, (16) with the geometric mean limit log⁡M0​(Vα)=limp→01p​log​∫01Vαp​α=∫01log⁡Vα​d​α, M_0 (\V_α\ )= _p→ 0\, 1p _0^1V_α^p\,dα= _0^1 V_α\,dα, (17) and the extremes log⁡M+∞=log​supαVα M_+∞= _αV_α and log⁡M−∞=log​infαVα M_-∞= _αV_α as p→±∞p→±∞. The three filter criteria emerge as evaluations of this single functional at different orders: • p→+∞p→+∞: log⁡M+∞=log⁡V0=log​det(MVEE) M_+∞= V_0= (MVEE) — the outermost α-cut volume, worst-case ignorance, the Popperian minimax bound operative in the falsification regime. • p→0p→ 0: log⁡M0=∫01log⁡Vα​d​α=Hπ M_0= _0^1 V_α\,dα=H_π — the log-geometric mean of α-cut volumes, integrated possibilistic ignorance, the ESPF optimality criterion. • p→−∞p→-∞: log⁡M−∞=log⁡V1 M_-∞= V_1 — the innermost α-cut volume (the anchor hypothesis neighborhood), best-case ignorance, a lower floor. The Hölder ordering M−∞≤M0≤M+∞M_-∞≤ M_0≤ M_+∞ (strict for non-constant Vα\V_α\) is the mathematical expression of the epistemological hierarchy: the ESPF’s integrated ignorance is bounded above by the Popperian worst-case and below by the best-case floor. The probability–possibility bridge. The same functional log⁡M0​(Vα) M_0(\V_α\) recovers the Kalman criterion when evaluated under Gaussian geometry. In the possibilistic case, VαV_α is the MVEE volume of hypotheses with π(i)≥απ^(i)≥α. In the Gaussian case, the natural analog of the α-cut is the confidence ellipsoid ℰα=x:(x−x^)⊤​Σ−1​(x−x^)≤−2​log⁡α,E_α= \x:(x- x) ^-1(x- x)≤-2 α \, with volume VαGauss=cn​(−2​log⁡α)n/2​(detΣ)1/2V_α^Gauss=c_n(-2 α)^n/2( )^1/2. Substituting into log⁡M0 M_0: HπGauss H_π^Gauss =∫01log⁡VαGauss​d​α = _0^1 V_α^Gauss\,dα =12​log​detΣ+n2​∫01log⁡(−2​log⁡α)​α+κn = 12 + n2 _0^1 (-2 α)\,dα+ _n =12​log​detΣ+const​(n), = 12 +const(n), (18) recovering the Kalman criterion exactly, since ∫01log⁡(−2​log⁡α)​α=log⁡2−γ _0^1 (-2 α)\,dα= 2-γ is a pure constant. Possibility theory and probability theory are not competing frameworks for this class of problems: they are evaluations of the same ignorance functional log⁡M0​(Vα) M_0(\V_α\) under different geometries of α-cuts. The possibilistic α-cut is defined by the MVEE of a possibility level set; the probabilistic α-cut is defined by a Gaussian confidence ellipsoid. The functional that integrates their log-volumes is identical. The Kalman filter is the Gaussian specialization of the ESPF’s optimality criterion, not a separate invention. Corollary 5.4 (Strict generalization). The ESPF is a strict generalization of the Kalman filter: where Kalman’s assumptions hold, the ESPF recovers the Kalman optimal solution; where they fail, the ESPF remains the minimum-HπH_π optimal filter in ℱF while the Kalman filter exits ℱF and its optimality claims do not apply. The ESPF is not a heuristic alternative but the epistemically correct generalization. 6 Numerical Validation 6.1 Setup We validate the theorem’s regime structure numerically using the ESPF applied to a range and range-rate orbital tracking scenario. The filter operates in state dimension n=7n=7 with Smolyak Level 3 quadrature, yielding M=106M=106 support points and Ntarget≈81N_target≈ 81–83 in nominal tracking, 32–64 at stress events. Process noise is aR=aT=aN=5×10−12a_R=a_T=a_N=5× 10^-12 km/s2 in RTN coordinates. Measurement noise is σR=1 _R=1 m range, σR˙=10−5 _ R=10^-5 km/s range-rate. Ground stations: Arecibo, Kwajalein, Diego Garcia. Two scenarios are run over a 2-day, 877-step diagnostic: • Nominal: clean LEO tracking with no injected errors. • Stress: 10 m/s crosstrack maneuver at day 1.0 combined with a 20 m Arecibo range bias active from day 0 onward. At each step, the diagnostic computes two claims against the ESPF’s minimum-q selection within the evidence-only class: • Claim A: Does min-q minimize log​det(MVEE) (MVEE)? • Claim B: Does min-q minimize HπωH^ω_π, the posterior-possibility-weighted integrated entropy? Comparators are 50 evidence-weighted random draws and one adversarial swap (replace the worst survivor with the best non-survivor), all restricted to the evidence-only class. The Epistemic Width Monitor (EWM) is computed at every step and comprises: WepW_ep (epistemic width), log​det(MVEEλ) (MVEE_λ) (regime indicator), αc _c (Hölder exponent), and IcI_c (possibilistic information content; see Remark 6.1). Remark 6.1 (Shannon surrogate in IcI_c). The current EWM implementation computes IcI_c via a Shannon entropy surrogate rather than the true possibilistic α-cut integral HπH_π. This proxy underestimates IcI_c in the falsification regime where it matters most; a full numerical implementation of the α-cut integral is planned for a future release. In the stress run, the surrogate underestimates IcI_c by a factor of 2 or more in high-surprisal epochs (days 1.0–1.8); the qualitative escalation pattern is preserved but the absolute magnitudes reported in Section 6.4 should be treated as lower bounds on true possibilistic stress. 6.2 Smolyak Level 3: Nominal Run The nominal 2-day run demonstrates the filter operating entirely in the diffusion regime: log​det(MVEE) (MVEE) remains positive throughout (≈100≈ 100–120 nats), and the regime flag never flips to Popperian. This is the correct behavior—a well-conditioned nominal LEO track with no model mismatch does not generate a falsification event. Key observations: • WepW_ep oscillates between ≈0.65≈ 0.65 and 0.870.87 with sharp measurement-epoch dips (∼0.6 0.6), reflecting healthy Popperian contraction followed by Jaynesian re-dilation. The pattern is stationary across the full 2-day window. • Prune count settles to 12–16 points per step after an initial 84-point contraction at the first measurement. • σk _k stabilizes to a stationary band of ≈0.45≈ 0.45–0.85 after initialization, with no systematic drift indicating filter divergence. • Necessity hugs zero with episodic flares to 0.02–0.06, and surprisal maintains a stationary band of ≈0.07≈ 0.07–0.13. • αc _c rides a stable band near 2.1–2.3 with periodic measurement-epoch spikes to ≈3.0≈ 3.0. 6.3 Smolyak Level 3: Stress Run The stress run reveals the filter’s behavior under sustained model mismatch. EWM regime indicator: log​det(MVEE) (MVEE) remains positive throughout the 2-day run (≈80≈ 80–120 nats). The MVEE regime flag never flips to Popperian. This is a significant finding discussed below. Prune count (the key diagnostic): Before day 0.75, the prune count matches the nominal pattern (12–16 points per step). At day ≈0.75≈ 0.75, coinciding with the accumulated effect of the range bias and approaching the maneuver epoch, the prune count escalates sharply to 40–85 points per step and remains elevated for the remainder of the run. At M=106M=106, this represents 38–80% of the support cloud being pruned at each update step. Necessity and surprisal (the early-warning indicators): Before day 0.75, necessity is near zero and surprisal is in the nominal band (0.05–0.12). Beginning at day 0.75, necessity escalates dramatically, with values routinely 0.4–1.0 from day 1.0 onward. The surprisal axis, scaled to [0,1000][0,1000] compared to the nominal [0,0.2][0,0.2], shows peaks of ≈640≈ 640 (day 1.0), ≈890≈ 890 (day 1.8), and episodic spikes >100>100 throughout. 6.4 EWM Diagnostic Findings The nominal and stress comparison at Smolyak Level 3 yields a result more nuanced and informative than a simple regime flip would have provided. The MVEE regime indicator is a coarse diagnostic. Under sustained model mismatch (bias + maneuver), the ESPF does not enter the falsification regime as defined by log​det(MVEE)<0 (MVEE)<0. Instead, the filter absorbs the error through aggressive Popperian pruning: the support cloud is violently thinned at each step, but the surviving points span sufficient volume that log​det(MVEE) (MVEE) never goes negative. The filter is operating in a high-pruning Jaynesian state, not a falsification state. Necessity and surprisal are the sensitive indicators. The MVEE regime flag alone would declare “all is well—Jaynesian throughout.” The necessity and surprisal plots tell the correct story: the filter is fighting against sustained model-reality divergence. Necessity near unity and surprisal three orders of magnitude above nominal are unambiguous signatures of epistemic stress, visible in the EWM before any MVEE regime flip would occur. The full EWM suite is required. The EWM diagnostic system comprising WepW_ep, log​det(MVEE) (MVEE), prune count, σk _k, necessity, and surprisal constitutes a multi-resolution epistemic health monitor. Each indicator captures a different aspect of the filter’s epistemic state: • log​det(MVEE) (MVEE): coarse regime classification (Jaynesian vs. Popperian). • Prune count: where epistemic stress is being absorbed (support thinning rate). • Necessity: whether individual hypotheses are becoming structurally load-bearing. • Surprisal: whether the filter’s predictive model is consistent with incoming evidence. These indicators are complementary, not redundant. Relying on the MVEE regime flag alone provides an incomplete and potentially misleading picture of filter health under stress. Comparison with Level 2. At Smolyak Level 2 (M=NminM=N_ ), the prune count band is 12–16 in both nominal and stress cases, and necessity/surprisal diverge at the stress onset but with smaller magnitude. The MVEE regime flag never flips at Level 2 either. This confirms that the regime structure is a property of the underlying theory—not a grid-resolution artifact—and that Level 3 provides diagnostic resolution that Level 2 cannot. 6.5 Tabular Validation (52 Steps) Tables 1 and 2 present the full 52-step Claim A/B diagnostic from the stress run, validating Lemma 3.7 directly. Key findings: • Claim B pass rate: 100% in the falsification regime against random comparators. Claim B is more robustly satisfied than Claim A throughout, consistent with the Hölder ordering M0≤M+∞M_0≤ M_+∞ (Remark 5.3). • Collapse events (steps 12–13, 35): produce the largest minimax advantages in the dataset. At step 13, min-q beats random by a factor of e16.6e^16.6 in log-det (ratio ≈10−6≈ 10^-6). • Swap failures are bounded (maximum 0.341 nats Claim A, 0.271 nats Claim B), structurally consistent, non-growing, and concentrated at steps where Smolyak grid hypotheses are geometrically close to the MVEE boundary. These are finite-grid artifacts, not failures of the minimax principle. • Diffusion regime (steps 7–12, log​det>0 >0): Claim B fails against random comparators, as predicted—the minimax criterion is not operative in the Jaynesian phase. Table 1: Lemma 3 diagnostic, steps 1–18: initial contraction, diffusion, first collapse. P/F = pass/fail; parenthetical = gap magnitude (nats). C = contraction (log​det<0 <0), D = diffusion (log​det>0 >0). Step σk _k log​det NtN_t Regime Claim A Claim B rand swap rand swap 1 0.970 −17.80-17.80 – C P P P P 2 0.941 −13.33-13.33 – C P F(0.025) P F(0.019) 3 0.913 −9.42-9.42 – C P F(0.028) P F(0.020) 4 0.885 −6.07-6.07 – C P P P P 5 0.859 −3.13-3.13 – C P F(<<0.001) P P 6 0.833 −0.34-0.34 – threshold F(0.063) F(0.119) F(0.099) F(0.089) 7 0.808 +1.94+1.94 – D F(0.015) P F(0.440) P 8 0.784 +3.59+3.59 – D P F(0.012) F(0.243) F(0.009) 9 0.760 +5.02+5.02 – D F(0.023) F(<<0.001) F(0.122) P 10 0.737 +7.05+7.05 – D P P P P 11 0.715 +9.01+9.01 – D P F(0.008) F(0.256) F(0.011) 12 0.694 +10.17+10.17 – pre-collapse F(0.022) F(0.022) P† P† 13 0.673 −56.61-56.61 – collapse 1 P P P P 14 0.653 −51.74-51.74 – C P F(0.370) P F(0.264) 15 0.633 −48.70-48.70 – C P P P P 16 0.614 −47.59-47.59 – C P F(0.004) P F(0.003) 17 0.596 −46.85-46.85 – C F(0.241) P P P 18 0.578 −47.68-47.68 – C P F(0.217) P F(0.140) †Step 12: HπωH^ω_π identical across all comparators (−2.294-2.294 nats); degenerate pre-collapse configuration. Table 2: Lemma 3 diagnostic, steps 19–52: deep contraction, second collapse, re-stabilization. Nt=NtargetN_t=N_target; † marks the second stress event. Step σk _k log​det NtN_t Regime Claim A Claim B rand swap rand swap 19 0.561 −50.89-50.89 82 C P F(0.060) P F(0.046) 20 0.544 −53.16-53.16 82 C P P P P 21 0.528 −56.15-56.15 82 C P P P P 22 0.512 −58.63-58.63 81 C P P P P 23 0.496 −61.02-61.02 81 C F(0.051) F(0.175) P F(0.113) 24 0.481 −55.27-55.27 83 C P F(0.061) P F(0.041) 25 0.467 −61.40-61.40 82 C P P P P 26 0.453 −65.24-65.24 81 C P F(0.006) P F(0.004) 27 0.439 −69.06-69.06 83 C P F(0.042) P F(0.030) 28 0.426 −73.28-73.28 83 C P P P P 29 0.413 −77.73-77.73 81 C P P P P 30 0.401 −77.73-77.73 83 C F(0.041) F(0.341) P F(0.271) 31 0.389 −82.46-82.46 83 C P P P P 32 0.377 −86.99-86.99 82 C P F(0.082) P F(0.066) 33 0.366 −91.99-91.99 82 C P F(0.045) P F(0.033) 34 0.355 −97.88-97.88 64 C P F(0.039) P F(0.017) 35 0.349 −104.08-104.08 52 C† P† P† P† P† 36 0.360 −108.77-108.77 67 C F(0.049) P P P 37 0.414 −110.99-110.99 64 C P F(0.036) P F(0.011) 38 0.476 −113.62-113.62 71 C F(0.052) F(0.019) P F(0.008) 39 0.548 −114.61-114.61 77 C P F(0.002) P P 40 0.630 −114.03-114.03 77 C P F(0.005) P F(0.002) 41 0.725 −112.29-112.29 79 C F(0.008) F(0.072) P F(0.023) 42 0.833 −109.32-109.32 81 C P F(0.011) P F(0.005) 43 0.901 −97.45-97.45 81 C P F(0.038) P F(0.027) 44 0.874 −94.19-94.19 82 C P F(<<0.001) P F(<<0.001) 45 0.847 −92.31-92.31 81 C P P P P 46 0.822 −90.47-90.47 82 C P F(0.004) P P 47 0.797 −88.86-88.86 82 C P F(<<0.001) P P 48 0.773 −87.73-87.73 82 C P P P P 49 0.750 −86.87-86.87 81 C P F(<<0.001) P P 50 0.728 −86.25-86.25 82 C P F(0.033) P F(0.026) 51 0.706 −86.08-86.08 82 C P P P P 52 0.685 −86.24-86.24 82 C P P P P †Step 35: second stress event. Nt=52N_t=52; min-q beats random by factor ∼11 11. 7 Discussion 7.1 What the Theorem Establishes and What Remains Open Theorem 4.1 establishes three things in the falsification regime. First, the PCRB is a universal information-theoretic lower bound on how rapidly any epistemically admissible filter may reduce possibilistic entropy—it follows from the structure of conjunctive update and possibilistic information content, not from any design choice. Second, the ESPF achieves this bound with equality, simultaneously through minimax support reduction (Popperian falsification) and possibility gradient concentration (internal Jaynesian assignment). Third, no other admissible filter within the evidence-only class achieves lower HπH_π while respecting the PCRB. For the linear-measurement case and for locally linearizable nonlinear models satisfying condition (10), the theorem is complete. Proposition 3.10 establishes the isotropy condition for the orbital mechanics setting to first order; a full proof of the nonlinear isotropy bound across all observation geometries remains open. 7.2 The Unique Role of the Possibility Gradient The most important conceptual contribution beyond the companion paper [3] is the recognition that the possibility gradient is not an implementation detail. It is part of the epistemic state. A filter that prunes survivors but reassigns uniform possibility after each update is discarding information: the gradient of compatibility across survivors encodes which hypotheses are more and less consistent with the accumulated evidence history. The ESPF carries that gradient forward. No other admissible filter can do better on either dimension without doing worse on the other. 7.3 Comparison to Existing Optimality Results Filter Optimality criterion Assumptions Filter is for Kalman Min. posterior covariance (MMSE) Linear, Gaussian Estimating truth Particle Asymptotically exact posterior Non-Gaussian, stochastic Estimating truth ESPF Min. possibilistic entropy HπH_π Bounded, possibilistic Eliminating the impossible The fourth column is the deepest distinction. Bayesian filters — Kalman and particle alike — are driven by a model of how truth generates observations. Their optimality guarantees are statements about convergence to truth under that model. When the model is wrong, they converge to the wrong truth. The ESPF carries no model of truth. Its optimality guarantee is a statement about ignorance: it retains the minimum set of hypotheses consistent with the evidence, assigns them possibilities in proportion to their evidence compatibility, and makes no further claim. It cannot be wrong about truth because it never asserts truth. It can only be more or less efficient at eliminating the impossible — and the EWM measures exactly that efficiency in real time. These results are not competing. They are optimal under different epistemic commitments. The Kalman and particle filter results require stochastic noise models. The ESPF result requires only bounded admissibility and conjunctive update. In the limit where stronger assumptions hold, the ESPF recovers the stronger result. The Hölder mean structure (Remark 5.3) makes this hierarchy precise and connects it directly to the numerical findings. The three optimality criteria occupy positions p=+∞p=+∞, p=0p=0, and p=0p=0 (Gaussian) on the Hölder curve. The ordering M0≤M+∞M_0≤ M_+∞ explains a finding that might otherwise seem puzzling: Claim B (HπωH^ω_π, order 0) is more robustly satisfied than Claim A (log​detMVEE , order +∞+∞) throughout the tabular validation, including at steps where Claim A shows swap failures. This is not a numerical artifact. It is a direct consequence of the Hölder ordering: the geometric mean of α-cut volumes is intrinsically less sensitive to perturbation at the boundary of the support than the supremum. The MVEE regime flag, operating at p=+∞p=+∞, is the coarsest instrument in the EWM suite for exactly this reason. The ESPF’s optimality target HπH_π is the more stable, more informative criterion — and the numerical results confirm it. 7.4 Implications for Filter Design The VFI is necessary for optimality. If the support degenerates, Hπ→−∞H_π→-∞, corresponding to falsely asserting perfect knowledge. The VFI prevents this. σmin>0 _ >0 is epistemically mandatory. A filter with σmin=0 _ =0 may collapse its support to a point, exiting ℱF and surrendering the optimality guarantee. Smolyak Level 3 is the minimum resolution for designed optimality. At Level 2, M=NminM=N_ and pruning cannot reduce the support; PCRB saturation is accidental rather than designed. At Level 3, M≫NminM N_ and minimum-q selection genuinely minimizes α-cut volumes across the full population. Possibility assignment is not optional. A filter that selects survivors by minimum qk(i)q^(i)_k but reassigns uniform possibility forfeits the gradient entropy improvement. The compatibility-based assignment ρ∗​(i)=Compk(i)ρ^*(i)=Comp^(i)_k is constitutive of the ESPF’s optimality, not a refinement. 7.5 Beyond Filtering The Jaynesian-Popperian synthesis identified here is not specific to state estimation. Any inference system that maintains a normalized possibility distribution over a finite hypothesis set and updates conjunctively—a knowledge graph with plausibility scores, an epistemic logic system with graded admissibility, a robust controller with bounded model uncertainty, an agentic AI system managing competing hypotheses about the world [19]—is implicitly managing HπH_π whether or not it is designed to do so. The ESPF optimality theorem says: in all such systems, the minimum-ignorance optimal update in the falsification regime is evidence-only ranking combined with compatibility-based possibility assignment. And the correct behavior in the diffusion regime is Jaynesian expansion: spread the support maximally, assert nothing that the evidence has not earned. Be quick to embrace ignorance. Be slow to assert certainty. These are not design choices. They are theorems. 8 Conclusion This paper has proved that the Epistemic Support-Point Filter (ESPF) is the unique optimal recursive estimator within the class of epistemically admissible evidence-only filters. Where Bayesian filters are driven toward an assumed truth, the ESPF eliminates the impossible: it retains only what the evidence has not ruled out, assigns possibilities in proportion to evidence compatibility, and makes no further claim. The prior has no vote on what is impossible. Only the evidence does. The filter implements a two-phase epistemological recursion. In its propagation phase, it enacts Jaynesian maximum entropy. In its measurement-update phase, it enacts Popperian falsification: select survivors by evidence alone, using the minimax entropy criterion—minimize log​det(MVEE) (MVEE), keeping the NtargetN_target hypotheses with smallest whitened squared innovations. Three lemmas formalize this structure. HπH_π decomposes into a minimax term (minimized by Popperian selection) and a gradient term (minimized by compatibility-based possibility assignment). The PCRB bounds how rapidly any admissible filter may reduce HπH_π per measurement. The Evidence-Optimality Lemma proves that min-q selection is the unique minimizer of HπH_π within the evidence-only class, and that incorporating prior possibility introduces a race-to-bottom failure mode. Numerical validation over a 2-day, 877-step Smolyak Level-3 run—spanning nominal LEO tracking and a sustained stress case combining maneuver and sensor bias—confirms this structure. The EWM diagnostic suite reveals that possibilistic stress manifests primarily through necessity saturation and surprisal escalation rather than MVEE sign change, establishing multi-resolution epistemic monitoring as essential for filter health assessment under realistic conditions. The regime structure is principled, not anomalous. The falsification regime is when Popper governs. The diffusion regime is when Jaynes governs. The boundary log​det=0 =0 is a diagnostic that identifies, in real time, which epistemological principle governs the current step. The ESPF is the optimal filter not because it minimizes some technically convenient functional, but because it correctly implements the only epistemologically defensible recursion: expand maximally under ignorance, contract minimally under evidence. Quick to embrace ignorance. Slow to assert certainty. This is not a design philosophy. It is a theorem. Acknowledgments The theoretical framing draws on decades of conversation with mentors, collaborators, and colleagues whose names appear in the companion arXiv paper [3]. References [1] R. E. Kalman, “A new approach to linear filtering and prediction problems,” J. Basic Engineering, 82(1):35–45, 1960. [2] R. E. Kalman and R. S. Bucy, “New results in linear filtering and prediction theory,” J. Basic Engineering, 83(1):95–108, 1961. [3] M. K. Jah and V. Haslett, “The Epistemic Support-Point Filter (ESPF): A Bounded Possibilistic Framework for Ordinal State Estimation,” arXiv:2508.20806, 2026. [4] M. K. Jah, “The Geometry of Knowing: From Possibilistic Ignorance to Probabilistic Certainty,” Submitted, 2026. [5] D. Dubois and H. Prade, Possibility Theory: An Approach to Computerized Processing of Uncertainty, Plenum Press, New York, 1988. [6] L. A. Zadeh, “Fuzzy sets as a basis for a theory of possibility,” Fuzzy Sets and Systems, 1(1):3–28, 1978. [7] M. Higashi and G. J. Klir, “Measures of uncertainty and information based on possibility distributions,” Int. J. General Systems, 9(1):43–58, 1983. [8] G. J. Klir and M. J. Wierman, Uncertainty-Based Information: Elements of Generalized Information Theory, Physica-Verlag, Heidelberg, 1999. [9] F. John, “Extremum problems with inequalities as subsidiary conditions,” in Studies and Essays Presented to R. Courant, Interscience, p. 187–204, 1948. [10] M. J. Todd, Minimum-Volume Ellipsoids: Theory and Algorithms, SIAM, Philadelphia, 2016. [11] H. Cramér, Mathematical Methods of Statistics, Princeton University Press, 1946. [12] C. R. Rao, “Information and the accuracy attainable in the estimation of statistical parameters,” Bull. Calcutta Math. Soc., 37:81–91, 1945. [13] P. Tichávský, C. H. Muravchik, and A. Nehorai, “Posterior Cramér–Rao bounds for discrete-time nonlinear filtering,” IEEE Trans. Signal Process., 46(5):1386–1396, 1998. [14] S. J. Julier and J. K. Uhlmann, “A new extension of the Kalman filter to nonlinear systems,” in Proc. AeroSense, p. 182–193, 1997. [15] E. T. Jaynes, Probability Theory: The Logic of Science, Cambridge University Press, 2003. [16] K. R. Popper, The Logic of Scientific Discovery, Hutchinson, London, 1959. [17] G. H. Hardy, J. E. Littlewood, and G. Pólya, Inequalities, Cambridge University Press, 2nd ed., 1952. [18] P. Walley, Statistical Reasoning with Imprecise Probabilities, Chapman & Hall, London, 1991. [19] M. K. Jah, “GaiaGraph: Agentic AI Infrastructure for Epistemic Planetary Stewardship,” GaiaVerse, Ltd., Technical Report, 2026.