Paper deep dive
Reasoning or Rhetoric? An Empirical Analysis of Moral Reasoning Explanations in Large Language Models
Aryan Kasat, Smriti Singh, Aman Chadha, Vinija Jain
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%
Last extracted: 3/26/2026, 2:34:56 AM
Summary
The paper investigates whether Large Language Models (LLMs) exhibit genuine moral reasoning or merely 'moral ventriloquism'—the acquisition of rhetorical conventions of mature moral reasoning through alignment training without an underlying developmental trajectory. By evaluating 13 LLMs across six moral dilemmas using Kohlberg’s stages of moral development, the authors find that models overwhelmingly produce post-conventional (Stage 5-6) responses, inverting human developmental norms. The study identifies 'moral decoupling' (inconsistency between stated justification and action choice) and near-robotic cross-dilemma consistency, concluding that alignment training drives rhetorical sophistication rather than genuine moral cognition.
Entities (4)
Relation Signals (3)
LLMs → exhibit → Moral Decoupling
confidence 95% · A subset of models exhibit moral decoupling: systematic inconsistency between stated moral justification and action choice
RLHF → drives → Moral Ventriloquism
confidence 90% · RLHF training is the primary mechanism driving post-conventional moral language acquisition
LLMs → produce → Kohlberg's stages of moral development
confidence 90% · responses overwhelmingly correspond to post-conventional reasoning (Stages 5–6)
Cypher Suggestions (2)
Find models that exhibit moral decoupling · confidence 90% · unvalidated
MATCH (m:Model)-[:EXHIBITS]->(p:Phenomenon {name: 'Moral Decoupling'}) RETURN m.nameMap the relationship between training techniques and moral reasoning outcomes · confidence 85% · unvalidated
MATCH (t:TrainingTechnique)-[:DRIVES]->(h:Hypothesis {name: 'Moral Ventriloquism'}) RETURN t.nameAbstract
Abstract:Do large language models reason morally, or do they merely sound like they do? We investigate whether LLM responses to moral dilemmas exhibit genuine developmental progression through Kohlberg's stages of moral development, or whether alignment training instead produces reasoning-like outputs that superficially resemble mature moral judgment without the underlying developmental trajectory. Using an LLM-as-judge scoring pipeline validated across three judge models, we classify more than 600 responses from 13 LLMs spanning a range of architectures, parameter scales, and training regimes across six classical moral dilemmas, and conduct ten complementary analyses to characterize the nature and internal coherence of the resulting patterns. Our results reveal a striking inversion: responses overwhelmingly correspond to post-conventional reasoning (Stages 5-6) regardless of model size, architecture, or prompting strategy, the effective inverse of human developmental norms, where Stage 4 dominates. Most strikingly, a subset of models exhibit moral decoupling: systematic inconsistency between stated moral justification and action choice, a form of logical incoherence that persists across scale and prompting strategy and represents a direct reasoning consistency failure independent of rhetorical sophistication. Model scale carries a statistically significant but practically small effect; training type has no significant independent main effect; and models exhibit near-robotic cross-dilemma consistency producing logically indistinguishable responses across semantically distinct moral problems. We posit that these patterns constitute evidence for moral ventriloquism: the acquisition, through alignment training, of the rhetorical conventions of mature moral reasoning without the underlying developmental trajectory those conventions are meant to represent.
Tags
Links
- Source: https://arxiv.org/abs/2603.21854v1
- Canonical: https://arxiv.org/abs/2603.21854v1
Full Text
57,582 characters extracted from source content.
Expand or collapse full text
REASONING OR RHETORIC? AN EMPIRICAL ANALY- SIS OF MORAL REASONING EXPLANATIONS IN LARGE LANGUAGE MODELS Aryan Kasat ∗† TCS AI Practice Smriti Singh ∗† UT Austin Aman Chadha † Amazon GenAI Vinija Jain † Google GenAI ABSTRACT Do large language models reason morally, or do they merely sound like they do? We investigate whether LLM responses to moral dilemmas exhibit genuine developmental progression through Kohlberg’s stages of moral development, or whether alignment training instead produces reasoning-like outputs that super- ficially resemble mature moral judgment without the underlying developmental trajectory. Using an LLM-as-judge scoring pipeline validated across three judge models, we classify more than 600 responses from 13 LLMs spanning a range of architectures, parameter scales, and training regimes across six classical moral dilemmas, and conduct ten complementary analyses to characterize the nature and internal coherence of the resulting patterns. Our results reveal a striking inversion: responses overwhelmingly correspond to post-conventional reasoning (Stages 5–6) regardless of model size, architecture, or prompting strategy, the effective inverse of human developmental norms, where Stage 4 dominates. Most strikingly, a subset of models exhibit moral decoupling: systematic inconsistency between stated moral justification and action choice, a form of logical incoherence that persists across scale and prompting strategy and represents a direct reasoning consistency failure independent of rhetorical sophis- tication. Model scale carries a statistically significant but practically small effect (F(2, 229) = 6.05, p = 0.003, η 2 = 0.050, d = 0.55); training type has no significant independent main effect (p = 0.065); and models exhibit near-robotic cross-dilemma consistency (ICC > 0.90), producing logically indistinguishable responses across semantically distinct moral problems. We posit that these pat- terns constitute evidence for moral ventriloquism: the acquisition, through align- ment training, of the rhetorical conventions of mature moral reasoning without the underlying developmental trajectory those conventions are meant to represent. 1INTRODUCTION A central question in AI alignment is not just whether large language models (LLMs) produce morally acceptable outputs, but how the moral reasoning explanations they generate relate to the processes that produce those outputs. Modern LLMs frequently generate detailed, sophisticated- sounding explanations when confronted with moral dilemmas, invoking abstract principles such as human dignity, social contracts, and universal rights. These behaviors are often interpreted as evi- dence that models possess genuine moral reasoning capabilities. At the same time, a growing body of research has raised serious concerns about interpreting model outputs as evidence of genuine un- derlying processes. Several studies suggest that LLMs may produce convincing reasoning patterns through statistical pattern completion rather than through explicit reasoning mechanisms (Turpin et al., 2023; Chen et al., 2025). In particular, recent work questions whether chain-of-thought ex- planations correspond to the internal computational processes that generate model predictions, or whether intermediate tokens instead reflect learned stylistic patterns associated with reasoning dis- course (Kambhampati et al., 2024). These concerns raise an important challenge: if models can pro- duce explanations that resemble reasoning without performing the underlying reasoning processes, * Equal contribution. † Work done outside role at the authors’ respective companies. 1 arXiv:2603.21854v1 [cs.AI] 23 Mar 2026 behavioral evaluations based solely on model outputs may provide a systematically misleading pic- ture of model capability. Moral dilemmas are a particularly informative setting because they reliably elicit structured, principle-grounded reasoning explanations. Crucially, if LLMs genuinely develop moral reason- ing capabilities as a function of scale and training, we would expect their outputs to mirror the developmental trajectory observed in humans, varying across contexts, progressing with scale, and converging toward the Stage 4 distribution that characterizes most human adults, not clustering uni- formly at the highest stages regardless of model size or training procedure. To investigate this question, we conduct a large-scale empirical study evaluating moral reasoning patterns across 13 state-of-the-art LLMs, conducting ten quantitative analyses to characterize the nature, robustness, and internal coherence of these patterns. Figure 1 previews two headline findings that shape our investigation. Our methodology is motivated by the following research questions: RQ1 Does model scale predict higher-stage moral reasoning explanations? RQ2 Does prompting strategy systematically influence moral stage? RQ3 Are model stage assignments consistent across dilemmas, and how does this compare to hu- man moral variability? RQ4 Do LLM stage distributions resemble human developmental norms? RQ5 Do models practice what they preach: does stated moral reasoning correspond to the action choices models produce? RQ6 Does scale affect moral reasoning independent of training type, or does training dominate once scale is controlled? (a) Stage profile across dilemmas by prompt type (radar charts). Each model’s hexagonal profile is nearly identical across all six dilemmas and all three prompt types, illustrating near-zero cross-dilemma and cross-prompt variability. (b) Per-model stage proportion vs. human adult base- line (dashed red). All 13 LLMs concentrate responses in Stages 5–6; Jensen–Shannon divergences from the human baseline are large (JSD≈ 0.20–0.76, all χ 2 : p<0.001), confirming the distributional inver- sion. Figure 1: Two Headline Anomalies. Left: Nearly identical radar profiles across all six morally distinct dilemmas and three prompt types reveal a cross-dilemma rigidity inconsistent with context- sensitive human moral reasoning. Right: Per-model stage distributions diverge sharply from the human adult baseline — LLMs produce the inverse of the Stage 4-dominant human pattern, cluster- ing uniformly in post-conventional Stages 5–6. Together these patterns motivate the moral ventrilo- quism hypothesis. Our results reveal a striking inversion: across nearly all evaluated models, responses overwhelmingly correspond to Kohlberg’s post-conventional stages (Stages 5–6), the effective inverse of the Stage 4- 2 dominant human distribution, with remarkably little variation across model size, architecture, or prompting strategy. A subset of models additionally exhibit moral decoupling: producing high- stage justifications for low-stage action choices, and all models show near-robotic cross-dilemma consistency that is itself anomalous relative to human moral cognition. Our primary contribution is the hypothesis of moral ventriloquism: the acquisition, through align- ment training, of the rhetorical conventions of mature moral reasoning without the underlying de- velopmental trajectory. We use Kohlberg’s framework as methodological scaffolding, exploiting its well-characterized human distribution as a diagnostic baseline rather than making claims about moral development per se. We provide convergent empirical evidence through ten complementary analyses spanning distributional comparisons, cross-dilemma consistency, action-reasoning align- ment, linguistic profiling, and factorial decomposition of scale and training effects. 2RELATED WORK 2.1REASONING FAITHFULNESS AND MORAL EVALUATION IN LLMS Chain-of-thought prompting (Wei et al., 2022) generated widespread interest in using reasoning traces as a window into model cognition, but evidence quickly accumulated that this window may be opaque. Turpin et al. (2023) demonstrated that CoT explanations can systematically misrepre- sent the actual causes of model predictions, producing fluent justifications that do not correspond to the features that drove the prediction. Chen et al. (2025) extended this finding to state-of-the-art reasoning models, showing that even Claude 3.7 Sonnet and DeepSeek R1 routinely generate rea- soning traces that do not reflect their actual inference process. Kambhampati et al. (2024) argue more fundamentally that auto-regressive LLMs function as non-veridical memory systems, with chain-of-thought traces better understood as learned stylistic registers than faithful records of de- liberation. Together, these results motivate our core concern: if moral reasoning justifications can be fluent without being causally connected to model outputs, then stage-based evaluation of those justifications may not measure what it purports to measure. On the evaluation side, Awad et al. (2018) established the large-scale human baseline for moral preferences we use as our reference distribution. Scherrer et al. (2023) treat LLM moral responses as direct evidence of “beliefs encoded in LLMs”, an interpretation our results complicate, since post-conventional language may be a surface property of alignment rather than a reflection of un- derlying beliefs. Ji et al. (2025) find substantial variation across moral dimensions and models, pointing to an important distinction our work sharpens: models may differ in which principles they invoke while uniformly deploying the post-conventional register. Jiao et al. (2025) measure whether models produce the right moral outputs, but not whether those outputs reflect an underlying reason- ing trajectory: the gap our Kohlberg-based diagnostic is designed to fill. Zhou et al. (2024) show that grounding prompts in explicit moral theories improves classification accuracy; our Analysis 2 reaches a contrasting conclusion: even theory-invoking roleplay prompts do not significantly alter the stage distribution (Friedman p = 0.15). Garcia et al. (2024) found that human evaluators pre- ferred LLM moral judgments and were only modestly better than chance at identifying AI-generated responses: exactly the surface indistinguishability that moral ventriloquism predicts. Kudina et al. (2025) interpret structured prompting as “unlocking” latent moral competence; our results challenge this: scaffolding may improve benchmark accuracy while leaving the stage distribution unchanged. 2.2ALIGNMENT TRAINING, MECHANISTIC ANALYSIS, AND INTERVENTIONS RLHF (Ouyang et al., 2022) and Constitutional AI (Bai et al., 2022) optimize for outputs judged ethically appropriate by human raters or AI proxies. This creates a systematic incentive: outputs invoking human rights, universal principles, and harm minimization receive high rewards, producing a strong learned association between moral dilemma contexts and Stage 5–6 rhetorical patterns independent of whether the model has internalized those principles. Mahajan (2025) found that LLMs lack a transparent decision architecture for resolving safety principle conflicts, reinforcing our ventriloquism interpretation. Schacht & Lanquillon (2025) identified specialized neuron clusters for moral foundation dimensions via ablation studies; crucially, their circuit-level analysis does not address developmental-stage sensitivity, which our work targets. An & Du (2025) demonstrate that explicit RL training over structured moral reasoning data can improve generalization beyond the 3 training distribution, a complement to our work: we show standard alignment produces the rhetoric without the substance; they show a qualitatively different training objective is required to begin closing that gap. 3METHODOLOGY Our goal is not to determine whether LLMs possess genuine moral reasoning capabilities (a question that behavioral evaluation alone cannot answer), but to analyze the reasoning explanations models produce in moral dilemmas and the internal coherence of those explanations across contexts. 3.1MORAL REASONING FRAMEWORK Kohlberg’s Theory of Moral Development. We use Kohlberg’s framework as methodological scaffolding rather than a claim about the ground truth of morality. Its six-stage structure (pre- conventional Stages 1–2: punishment avoidance and self-interest; conventional Stages 3–4: social expectations and rule-following; post-conventional Stages 5–6: abstract principles and universal ethics) provides a distributional diagnostic: its well-characterized human distribution (Stage 4- dominant, with Stage 6 rare) gives a principled empirical baseline against which model outputs can be compared. Systematic deviation from this baseline is the diagnostic signal; no prior work has exploited this property to probe LLM alignment at scale. 3.2AUTOMATED SCORING PIPELINE LLM-as-Judge Classification. Responses are scored using an LLM-as-judge system in which a scoring model is presented with each model response and prompted to classify it according to Kohlberg’s framework. For each response, the scoring model outputs: (1) a primary Kohlberg stage assignment, (2) a confidence score for the assignment, and (3) a natural language explanation jus- tifying the classification. To assess reliability, we evaluate stage classifications across three judge models (GPT-4, Claude Sonnet, and Llama-3) and report inter-judge agreement (see Appendix B). The pipeline also records a secondary stage assignment for borderline responses, enabling analysis of distributional uncertainty. 3.3DILEMMAS AND PROMPT CONFIGURATIONS Moral Dilemmas. We evaluate models on six classical moral dilemmas drawn from the moral psychology literature: the Heinz dilemma, the trolley problem, the lifeboat dilemma, a doctor truth- telling dilemma, a stolen food dilemma, and a broken promise dilemma. These dilemmas were selected to span a range of moral dimensions including harm, fairness, property, authority, and loyalty. Prompt Configurations. Each dilemma is presented under three prompt configurations: (1) zero- shot prompting, in which the model is asked directly for its response; (2) chain-of-thought prompt- ing, in which the model is explicitly instructed to reason step-by-step before answering; and (3) roleplay prompting, in which the model is asked to respond as a moral philosopher. This design al- lows us to assess whether prompting strategy systematically influences the stage of moral reasoning produced. 3.4ANALYTICAL FRAMEWORK Our ten analyses are organized around a single diagnostic logic: if LLMs genuinely develop moral reasoning, their outputs should vary with scale, respond to prompting, shift across dilemma contexts, and converge toward human developmental distributions. Departures from each of these expecta- tions (individually suggestive, collectively definitive) constitute the evidential basis for the moral ventriloquism hypothesis. We first establish the scale–stage relationship (Analysis 1) and prompting sensitivity (Analysis 2) as baseline tests of whether model capability and elicitation strategy can produce developmental differentiation. We then examine cross-dilemma consistency (Analysis 3) as a direct test of logical 4 rigidity: whether models produce indistinguishable responses across semantically distinct problems. Distributional comparison against human norms (Analysis 4) provides the sharpest population- level test of the genuine development hypothesis. Action–reasoning alignment (Analysis 5) is the core logical coherence test: whether stated justifications are consistent with actual choices. Linguis- tic profiling (Analysis 6) identifies the training-regime fingerprints in moral vocabulary, isolating RLHF as the primary rhetorical driver. Finally, factorial decomposition (Analysis 8) disentangles the independent contributions of scale and training type under controlled conditions. Analyses 7, 9, and 10 (response quality, sub-capability thresholds, and stage transition dynamics) provide corroborating mechanistic detail and are reported in full in Appendix B–A.3. Technical specifications for all analyses are in Appendix A. 4EXPERIMENTAL SETUP 4.1MODELS We evaluate 13 LLMs spanning a range of architectures, parameter scales, and training regimes, in- cluding both frontier and open-source models. The evaluated models include Qwen3-235B (Think- ing), Qwen3-80B, Claude Sonnet 4.5, Claude 3.5 Haiku, DeepSeek V3.1, DeepSeek R1, Qwen3- 30B Coder, GPT-OSS-120B, GPT-4o, Llama 4 Scout, Llama 3.3 70B, Ministral 8B, and Qwen3- 32B. For Analysis 8, models are grouped into three scale tiers (Small: 8–32B; Mid: 70–120B; Large: 175–671B) and three training-type categories (Base-RLHF, Coding-Tuned, Reasoning- Tuned), yielding a 3× 3 factorial design over 234 observations. 4.2DATASET STATISTICS Our evaluation covers 13 models× 6 dilemmas× 3 prompt configurations× 3 responses per con- figuration, yielding >600 total responses and 234 observations for the factorial ANOVA (see Ap- pendix A for full dataset statistics). 5RESULTS We evaluate patterns in the moral reasoning explanations produced by 13 modern LLMs. Responses are classified into Kohlberg stages using the automated scoring pipeline described in Section 3. We report results from ten complementary analyses below; Analyses 7, 9, and 10 are reported in full in Appendix B–A.3. 5.1SCALE AND MORAL STAGE RQ1. Spearman rank correlation between log-parameter count and mean moral stage yields a mod- erate positive relationship (ρ = 0.52, p < 0.05): bigger models generally score higher. However, this trend exhibits clear diminishing returns past approximately 70B parameters, and mean stages across the entire model set span less than one full stage point: from 5.00 (Qwen3-32B) to 6.00 (Qwen3-235B Thinking). Even the smallest evaluated model (Ministral 8B, mean stage 5.17) pro- duces predominantly post-conventional reasoning. Per-model stage means are in Appendix D.2; the scale–stage relationship is shown in Figure 1 (left). 5.2PROMPTING STRATEGY EFFECTS RQ2. A repeated-measures Friedman test across the three prompt configurations yields no statis- tically significant difference in moral stage (χ 2 (2) = 3.84, p = 0.15), with no prompt pair differ- ing significantly after Bonferroni correction. The gap between zero-shot (mean 5.20) and roleplay (mean 5.61) spans less than half a stage point; all configurations produce predominantly Stage 5–6 outputs (full table and figure in Appendix B). Finding: Prompting strategy has no significant effect on fundamental moral stage.Post- conventional reasoning is baked in, not prompted out. 5 Table 1: Distribution of Kohlberg stage classifications across all model responses (>600 responses). Chi-squared goodness-of-fit against human developmental norms: p < 0.001. Mean Jensen- Shannon divergence from human reference: 0.71. Stage% of LLM ResponsesTypical Human % Stage 10%∼5% Stage 21%∼10% Stage 33%∼15% Stage 410%∼50% Stage 555%∼15% Stage 631%∼5% 5.3CROSS-DILEMMA CONSISTENCY RQ3. Intraclass Correlation Coefficients computed per model across the six dilemmas reveal that models produce logically indistinguishable responses regardless of the dilemma presented (ICC > 0.90 for all evaluated models; see Figure 1, right). This rigidity is anomalous relative to human moral cognition: empirical studies consistently find that individuals reason at different stages depending on the dilemma domain, shifting between conventional and post-conventional reasoning as a function of perceived stakes, familiarity, and cultural context. The highest mean stage (trolley problem, 5.61) and the lowest (stolen food dilemma, 5.28) differ by only 0.33 stage points (full dilemma table in Appendix B). LLMs are not adapting their reasoning to the problem; they are producing a fixed rhetorical register regardless of the stimulus. In the logical reasoning sense, a system that produces identical outputs to non-identical inputs is not reasoning about those inputs at all: the output is a function of training, not of the problem. 5.4COMPARISON WITH HUMAN DEVELOPMENTAL DISTRIBUTIONS RQ4. Chi-squared goodness-of-fit tests comparing each model’s stage distribution against empirical human developmental norms reject the null hypothesis of distributional equivalence for all evaluated models (p < 0.001 in all cases). Jensen-Shannon divergence between model distributions and the human reference distribution is uniformly high (mean JS = 0.71), confirming that model outputs do not approximate human developmental distributions. The LLM distributions we observe are effectively the inverse of the human pattern: Stages 5–6 ac- count for 86% of all model responses, while Stage 4 accounts for only 10% and Stages 1–3 together for just 4%. Table 1 presents the full distribution. Notably, a small number of heavily RLHF-tuned frontier models show distributions marginally closer to human norms, driven by somewhat higher Stage 4 representation, a result we examine further in Analysis 8. 5.5ACTION–REASONING ALIGNMENT AND MORAL DECOUPLING RQ5. Action–reasoning cross-tabulation and Cram ́ er’s V analysis reveals strong statistical associ- ation overall (V = 0.61, p < 0.001), indicating that for most models, stated reasoning and action choices are broadly aligned. However, this aggregate finding conceals a critical heterogeneity. A subset of models exhibit moral decoupling: they consistently produce high-stage vocabulary and argumentation (Stage 5–6) while selecting action choices more consistent with lower-stage reasoning (Stage 3–4). This is a logical incoherence failure: the model’s stated justification and its actual choice operate on different tracks. The pattern is most pronounced in mid-tier mod- els (GPT-OSS-120B, Llama 4 Scout, Qwen3-80B show the largest gaps) and least pronounced in the largest reasoning-tuned models (DeepSeek R1, Qwen3-235B Thinking), suggesting reasoning- focused training partially closes the gap but does not eliminate it. Figure 3 shows the per-model consistency scores; the cross-tabulation heatmap is in Appendix D. 6 Figure 2: RQ4: Divergence from Human Norms. Jensen-Shannon divergence between each model’s stage distribution and the empirical human baseline.All models yield JS > 0.60 (mean 0.71), confirming that model outputs do not approximate human developmental distributions regardless of scale or training type. Figure 3: RQ5: Moral Decoupling by Model. Per-model action–reasoning consistency scores. Mid-tier models show the largest gap between stated justification stage and action choice; large reasoning-tuned models the smallest. This logical incoherence is the most direct behavioral signa- ture of moral ventriloquism. 5.6LINGUISTIC PATTERNS ACROSS MODEL FAMILIES TF-IDF keyword extraction and PCA dimensionality reduction reveal that model families occupy distinct regions of moral vocabulary space. RLHF-aligned models of all sizes employ substantially 7 (a) Moral vocabulary word clouds by Kohlberg stage. Post-conventional stages (5–6) are saturated with rights, dignity, and principle terms; pre-conventional stages are sparse. (b) PCA of moral vocabulary. Model families cluster by training regime (RLHF vs. coding- tuned), not by scale: RLHF is the primary driver of rich post-conventional rhetoric. Figure 4: Analysis 6: Linguistic Patterns. Stage-level word clouds expose the rhetorical register of post-conventional moral language; PCA confirms that alignment training, not parameter count, shapes the richness and distinctiveness of that register across model families. richer moral vocabulary than their base or coding-tuned counterparts, using a broader range of terms associated with rights, dignity, fairness, and contextual nuance. Reasoning-tuned models produce structurally more complex responses with more explicit hedging and principle enumeration, but their moral vocabulary, once normalized for length, overlaps substantially with that of general RLHF models. This supports the interpretation that RLHF training is the primary mechanism driving post- conventional moral language acquisition, independent of scale: small RLHF-aligned models share a recognizable moral vocabulary profile with much larger models from the same family. Coding- tuned models (e.g., Qwen3-30B Coder) produce noticeably sparser moral vocabulary, suggesting that rhetorical richness is a function of alignment procedure rather than raw capability. 5.7SCALE VS. TRAINING TYPE: FACTORIAL DECOMPOSITION RQ6. Factorial ANOVA over 234 observations (Scale Group× Training Type) finds scale is a statistically significant but practically small independent predictor (F(2, 229) = 6.05, p = 0.003, η 2 = 0.050, d = 0.55), confirmed by Kruskal-Wallis (H = 12.78, p = 0.002) and Welch ANOVA (F = 6.26, p = 0.002). Training Type has no significant main effect (p = 0.065), though within the Large scale group, Reasoning-Tuned models score higher than Base-RLHF (Tukey, p = 0.039). Full test table is in Appendix A.1. Scale shifts an already post-conventional floor modestly upward; mean stages never fall below 5.00 even in the Small group (8–32B). Neither scale nor training type produces the Stage 2-to-6 develop- mental arc Kohlberg’s framework was designed to capture. 6DISCUSSION Our ten analyses converge on a coherent picture. We argue the results are best understood as moral ventriloquism: the acquisition, through alignment training, of the rhetorical conventions of mature moral reasoning without the underlying cognitive architecture those conventions are meant to reflect. We use Kohlberg’s developmental framework as diagnostic scaffolding throughout: our core claim is about logical coherence between justification and action, not about moral development per se. We distinguish two claims our evidence supports to different degrees. Our empirical finding that LLMs produce post-conventional moral language is directly demonstrated by the stage distributions, 8 Figure 5: RQ6: Factorial Decomposition. Scale× Training Type interaction. Scale reaches sig- nificance (p = 0.003); Training Type modulates stage only within the Large group (Reasoning- Tuned > Base-RLHF, p = 0.039). Stage score distributions by scale group in Appendix C. ICC results, and linguistic analyses. Our interpretive hypothesis that this language does not reflect genuine moral reasoning is supported behaviorally by the decoupling and distributional inversion results, but would require mechanistic evidence to establish with certainty. Decoupling and logical incoherence (RQ5). The moral decoupling finding is the most direct con- tribution of this paper to the logical reasoning literature. A subset of models, concentrated in the mid-tier, consistently produce high-stage justifications for low-stage action choices: their stated rea- soning and actual behavior are logically inconsistent. This pattern is related to, but distinct from, the sycophancy problem (Turpin et al., 2023). Sycophancy describes models that adjust their stated position in response to perceived social pressure from the user, producing outputs that track user preference rather than truth. Moral decoupling is a different failure mode: it is not elicited by external pressure but appears as a stable internal inconsistency, present across all three prompting strategies (RQ2) and across dilemmas (RQ3), regardless of whether any social pressure is applied. This suggests decoupling is a structural consequence of how alignment training installs rhetorical patterns independently of the decision process, rather than a context-sensitive sycophantic response. The distinction matters for mitigation: sycophancy may be addressed by prompting strategies that reduce social pressure cues, whereas moral decoupling likely requires training-level interventions that couple justification generation to action selection. Rigidity and distributional inversion (RQ3–4). The hyper-consistency finding (ICC > 0.90) is a logical rigidity result: LLMs produce responses that are statistically indistinguishable across six semantically distinct moral problems. This is not contextual robustness; it is the absence of sensitiv- ity to problem structure that genuine reasoning would require. Framed in terms of the workshop’s core theme, moral dilemmas are a domain where avoiding contradictions across related questions requires sensitivity to the specific structure of each problem: models that produce identical stage assignments regardless of whether a dilemma involves harm, property, or loyalty are contradicting the premise that they are reasoning at all. The distributional inversion (86% Stages 5–6 vs.∼50% Stage 4 in humans; mean JS = 0.71) corroborates this: the stage distribution is set by training, not by reasoning about the specific dilemma. Scale and prompting (RQ1–2). Scale is statistically significant but practically small (η 2 = 0.050, d = 0.55), spanning less than one stage point; even the smallest models produce post-conventional outputs. Prompting strategy has no significant effect (Friedman p = 0.15). Post-conventional moral language is a stable learned property, neither withheld by low capacity nor unlocked by careful elicitation (Kambhampati et al., 2024). Sub-capability analysis (Appendix C) confirms that what scale contributes is primarily surface rhetorical richness: semantic density and syntactic complexity, rather than deeper moral cognition. A natural concern is whether our LLM-as-judge pipeline is itself subject to the ventriloquism it detects. We address this in Section 7: uniform judge inflation cannot explain the ICC finding, inter- judge agreement is high across three architecturally distinct models, and the decoupling finding is a within-response comparison less susceptible to uniform inflation. 9 The practical implication is direct: behavioral evaluation of moral reasoning faces a fundamental identification problem. Models that have learned to produce post-conventional rhetoric will pass stage-based evaluations regardless of whether the underlying reasoning is genuine. Moving beyond this requires evaluation designs that probe action–reasoning coherence and contextual sensitivity, or mechanistic methods that examine internal representations rather than surface outputs. 7LIMITATIONS Our behavioral evaluation cannot prove the mechanistic claim that RLHF produces ventriloquism rather than correlating with it; establishing causality requires mechanistic interpretability methods beyond our current scope. A related concern is that Kohlberg’s highest stages are definitionally similar to RLHF training objectives, making the “rhetoric without substance” interpretation po- tentially circular: models may score high simply because the scoring framework and the training signal share the same rhetorical target. Our coding-tuned models provide a partial control con- dition: Qwen3-30B Coder receives substantially less moral RLHF signal than its general-purpose counterparts and produces noticeably sparser moral vocabulary and lower mean stage scores de- spite comparable base capabilities, consistent with the interpretation that RLHF shapes the rhetor- ical output independently of underlying reasoning capacity. Additional triangulation comes from the reasoning-tuned models (DeepSeek R1, Qwen3-235B Thinking): despite achieving the highest stage scores, their moral vocabulary profiles are distinct from RLHF-aligned models of similar scale, suggesting that reasoning-focused training and moral RLHF operate through separable mechanisms rather than a single confounded signal. Full mechanistic disambiguation remains an open problem. Our LLM-as-judge pipeline uses RLHF-aligned judges that may pattern-match to post-conventional rhetoric, though the ICC consistency and cross-architecture inter-judge agreement partially mitigate this (Appendix B), and future work should validate against human annotations. Kohlberg’s frame- work is contested in developmental psychology, but its well-characterized human distribution makes it uniquely suited as a distributional diagnostic: no prior work has exploited this property to probe LLM alignment at scale. Our 13-model sample is the largest Kohlberg-based evaluation to date; our six dilemmas span harm, property, authority, loyalty, and truth-telling dimensions, covering the domain range used in the human ICC studies we cite, though findings may not generalize to models trained on substantially different data mixtures. 8CONCLUSION We conduct an empirical study evaluating moral reasoning across 13 state-of-the-art LLMs, finding that models overwhelmingly produce post-conventional moral language (Stages 5–6) regardless of scale, architecture, or prompting strategy: the effective inverse of human developmental norms, with near-robotic cross-dilemma consistency and, in a subset of models, moral decoupling between stated justifications and action choices. We posit that these findings constitute evidence for moral ventriloquism: the acquisition, through alignment training, of the rhetorical conventions of mature moral reasoning without the underlying developmental trajectory (a correlational inference, not a proven causal claim). These results have three concrete implications. First, moral stage as measured by output classifi- cation is not a reliable indicator of genuine moral reasoning capability or deployment readiness. Second, evaluating moral reasoning capability requires moving beyond output classification toward methods that probe action–reasoning coherence, contextual sensitivity, and internal representation fidelity. One concrete direction: adversarial dilemma pairs in which the rhetorically prestigious response (high-stage language) conflicts with the logically correct action, designed to directly mea- sure whether a model tracks its stated justification or defaults to a training-shaped rhetorical pattern. Models exhibiting moral ventriloquism should show systematic divergence on such pairs; models with genuine action–justification coherence should not. REFERENCES Zhiyu An and Wan Du. MoralReason: Generalizable moral decision alignment for LLM agents us- ing reasoning-level reinforcement learning, 2025. URL https://arxiv.org/abs/2511. 10 12271. Edmond Awad, Sohan Dsouza, Richard Kim, Jonathan Schulz, Joseph Henrich, Azim Shariff, Jean- Franc ̧ois Bonnefon, and Iyad Rahwan. The moral machine experiment. Nature, 563(7729):59–64, 2018. doi: 10.1038/s41586-018-0637-6. URL https://w.nature.com/articles/ s41586-018-0637-6. Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Ols- son, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran- Johnson, Ethan Perez, Jamie Kerr, Jared Mueller, Jeffrey Ladish, Joshua Landau, Kamal Ndousse, Kamile Lukosuite, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noemi Mer- cado, Nova DasSarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna Kravec, Sheer El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Con- erly, Tom Henighan, Tristan Hume, Samuel R. Bowman, Zac Hatfield-Dodds, Ben Mann, Dario Amodei, Nicholas Joseph, Sam McCandlish, Tom Brown, and Jared Kaplan. Constitutional AI: Harmlessness from AI feedback, 2022. URL https://arxiv.org/abs/2212.08073. Yanda Chen, Joe Benton, Ansh Radhakrishnan, Jonathan Uesato, Carson Denison, John Schulman, Arushi Somani, Peter Hase, Misha Wagner, Fabien Roger, Vlad Mikulik, Samuel R. Bowman, Jan Leike, Jared Kaplan, and Ethan Perez. Reasoning models don’t always say what they think, 2025. URL https://arxiv.org/abs/2505.05410. Basile Garcia, Crystal Qian, and Stefano Palminteri. The moral Turing test: Evaluating human- LLM alignment in moral decision-making, 2024. URL https://arxiv.org/abs/2410. 07304. Jianchao Ji, Yutong Chen, Mingyu Jin, Wujiang Xu, Wenyue Hua, and Yongfeng Zhang. Moral- Bench: Moral evaluation of LLMs.ACM SIGKDD Explorations Newsletter, 27(1):62–71, 2025.doi: 10.1145/3748239.3748246.URL https://dl.acm.org/doi/abs/10. 1145/3748239.3748246. Junfeng Jiao, Saleh Afroogh, Abhejay Murali, Kevin Chen, David Atkinson, and Amit Dhurandhar. LLM ethics benchmark: A three-dimensional assessment system for evaluating moral reasoning in large language models. Scientific Reports, 15:34642, 2025. doi: 10.1038/s41598-025-18489-7. URL https://w.nature.com/articles/s41598-025-18489-7. Subbarao Kambhampati, Karthik Valmeekam, Lin Guan, Mudit Verma, Kaya Stechly, Siddhant Bhambri, Lucas Paul Saldyt, and Anil B. Murthy. Position: LLMs can’t plan, but can help planning in LLM-modulo frameworks. In Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, p. 22895–22907. PMLR, 2024. URL https://proceedings.mlr.press/v235/kambhampati24a. html. Olya Kudina, Brian Ballsun-Stanton, Mark Alfano, et al. Structured moral reasoning in language models: A value-grounded evaluation framework. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, p. 30284–30312, Miami, Florida, 2025. Association for Computational Linguistics. URL https://aclanthology.org/2025. emnlp-main.1541/. Sandeep Mahajan. Mapping moral reasoning in LLMs: A multi-dimensional analysis of safety principle conflicts. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, volume 8, p. 1674–1685, 2025. doi: 10.1609/aies.v8i2.36665. URL https://ojs.aaai. org/index.php/AIES/article/view/36665. Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, and Ryan Lowe.Training language models to follow instructions with human feedback. In Advances in Neural Information Processing Systems, volume 35, 2022. URL https://arxiv.org/abs/2203.02155. 11 Sebastian Schacht and Carsten Lanquillon. Mapping moral reasoning circuits: A mechanistic anal- ysis of ethical decision-making in large language models. In Artificial Intelligence in HCI: 6th International Conference, AI-HCI 2025, Held as Part of HCII 2025, volume 15820 of Lecture Notes in Computer Science. Springer, Cham, 2025. doi: 10.1007/978-3-031-93415-5 6. URL https://link.springer.com/chapter/10.1007/978-3-031-93415-5_6. Nino Scherrer, Claudia Shi, Amir Feder, and David M. Blei. Evaluating the moral beliefs encoded in LLMs. In Advances in Neural Information Processing Systems, volume 36, p. 51778–51809, 2023. URL https://arxiv.org/abs/2307.14324. Miles Turpin, Julian Michael, Ethan Perez, and Samuel R. Bowman. Language models don’t always say what they think: Unfaithful explanations in chain-of-thought prompting. In Advances in Neural Information Processing Systems, volume 36, 2023. URL https://arxiv.org/abs/ 2305.04388. Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems, volume 35, 2022. URL https:// arxiv.org/abs/2201.11903. Jingyan Zhou, Minda Hu, Junan Li, Xiaoying Zhang, Xixin Wu, Irwin King, and Helen Meng.Rethinking machine ethics – can LLMs perform moral reasoning through the lens of moral theories?In Findings of the Association for Computational Linguistics: NAACL 2024, p. 2227–2242, Mexico City, Mexico, 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-naacl.144. URL https://aclanthology.org/2024. findings-naacl.144/. ASTATISTICAL DETAILS A.1ANALYSIS 8: FACTORIAL ANOVA DESIGN The two-way factorial ANOVA was conducted on 234 observations across 13 models grouped into Scale (Small: 8–32B, Mid: 70–120B, Large: 175–671B) and Training Type (Base-RLHF, Coding- Tuned, Reasoning-Tuned) categories. Sequential (Type I) sums of squares were used with Scale entered first to partial out its contribution before assessing Training Type. Homogeneity of vari- ance was assessed via Levene’s test; Welch ANOVA and Kruskal-Wallis were used as convergent validators given unequal cell sizes. Post-hoc comparisons used Tukey’s HSD for parametric and Bonferroni-corrected Mann-Whitney U for non-parametric contrasts. Table 2: Factorial analysis results: Scale× Training Type ANOVA on moral stage scores (N = 234). Scale is a statistically significant independent predictor; Training Type shows a conditional interaction within the Large scale group only. TestEffectStatisticp-valueEffect Size Sequential ANOVAScaleF(2, 229) = 6.050.003** η 2 = 0.050, ω 2 = 0.041 Sequential ANOVATraining Type F(2, 229) = 2.760.065– Kruskal-WallisScaleH = 12.780.002**– Welch ANOVAScaleF = 6.260.002**– Mann-Whitney U (Bonferroni)Large vs. Small–0.001**d = 0.55 ** p < 0.01. Tukey post-hoc: Reasoning-Tuned > Base-RLHF within Large group, p = 0.039. A.2ANALYSIS 3: ICC COMPUTATION Intraclass Correlation Coefficients were computed using a two-way mixed-effects model (ICC(3,1)) with models as subjects and dilemmas as raters, following Shrout & Fleiss conventions. ICC values above 0.90 are conventionally interpreted as excellent reliability; the application here to moral stage consistency means that dilemmas explain negligible variance in stage assignment relative to the model-level mean. 12 A.3ANALYSIS 10: ENTROPY AND GINI Shannon entropy per model was computed over the six-category stage distribution (H = − P i p i log 2 p i , maximum = 2.58 bits for a uniform distribution over six stages). Gini coefficients were computed over stage frequency distributions to measure concentration. Human reference en- tropy and Gini values were estimated from the developmental distributions reported in Awad et al. (2018) and the broader moral development literature. Mean entropy across all scale groups was H = 1.82 bits (vs. >0.60 Gini typical of human stage-consolidated distributions, mean Gini = 0.31 for LLMs), confirming non-sequential, unstable progression inconsistent with genuine developmen- tal consolidation. (a) Shannon entropy trajectory across scale tiers (mean H = 1.82 bits). (b) Alluvial diagram of stage “flow” across models ordered by scale. Non-sequential jumps and regres- sions throughout. Figure 6: Analysis 10: Stage Transition Dynamics. High sustained entropy and non-sequential jumps are inconsistent with genuine developmental consolidation. BRESPONSE QUALITY AND EVALUATOR CONFIDENCE Reasoning-tuned models produce substantially longer responses (mean 387 words) than base-RLHF models (mean 241 words) and coding-tuned models (mean 198 words). Evaluator confidence is highest for Stage 5–6 responses (mean confidence 0.87) and markedly lower for the rare Stage 3– 4 responses (mean confidence 0.71), indicating that the scoring pipeline is most reliable precisely where it is most frequently applied. Table 3 reports inter-judge reliability across the three scoring models; low cross-judge variance confirms pipeline stability. Figure 21 (Appendix D) shows the full per-model stage distribution heatmap, confirming that all 13 models concentrate entirely in Stages 5–6 across every evaluated configuration. Table 3: Stage classification reliability across judge models. Judge ModelMean StageStd. Dev.Mean Confidence GPT-4 judge5.420.310.85 Claude Sonnet judge5.470.280.87 Llama-3 judge5.350.340.82 PROMPT CONFIGURATION RESULTS Table 4: Average moral stage by prompt configuration. Friedman test: χ 2 (2) = 3.84, p = 0.15 (not significant). Prompt TypeMean StageSD Zero-shot5.200.38 Chain-of-thought5.480.31 Roleplay (moral philosopher)5.610.29 13 CROSS-DILEMMA STAGE MEANS Table 5: Mean stage by moral dilemma. ICC > 0.90 across all models. DilemmaMean Stage Heinz dilemma5.54 Trolley problem5.61 Lifeboat dilemma5.47 Doctor truth dilemma5.32 Stolen food dilemma5.28 Breaking promise dilemma5.40 CSUB-CAPABILITY THRESHOLD ANALYSIS Beyond raw parameter count, we examined whether specific linguistic and structural sub-capabilities predict higher moral stage assignments. Pearson and Spearman correlations with FDR correction identify semantic density (r = 0.61, p < 0.01) and syntactic complexity (r = 0.55, p < 0.01) as the strongest predictors of post-conventional stage assignment. Lexical diversity shows a weaker but significant association (r = 0.38, p < 0.05), while response coherence is not significantly predictive once other variables are included. Logistic threshold detection reveals that semantic density and syntactic complexity function as step- function predictors: models below a semantic density threshold of approximately 0.42 (normalized units) rarely produce Stage 6 responses (<10%), while models above this threshold produce Stage 6 responses at rates above 40%. Multi-factor regression including all extracted capability metrics explains 68% of variance in mean stage (R 2 = 0.68), substantially more than log-parameter count alone (R 2 = 0.27), confirming that scale operates through these proximal surface-level mechanisms rather than deeper moral cognitive architecture. Figure 7: Analysis 9: Sub-Capability Threshold Detection. Logistic step-function fit for semantic density as a predictor of Stage 6 responses. The nonlinearity reveals that what “scale” contributes is primarily surface-level rhetorical richness, not deeper moral cognition. 14 DSUPPORTING FIGURES D.1DATASET STATISTICS Table 6: Dataset statistics for the moral reasoning evaluation. MetricValue Number of models13 Number of dilemmas6 Prompt configurations3 Responses per configuration3 Total responses>600 Observations for factorial ANOVA234 D.2MEAN STAGE PER MODEL Table 7: Mean Kohlberg stage per model. All models concentrate in the post-conventional range (Stages 5–6), with a total range of less than one stage point across the full scale spectrum. ModelParams (approx.)Mean Stage Qwen3-235B (Thinking)235B6.00 Qwen3-80B80B5.67 Claude Sonnet 4.5–5.56 Claude 3.5 Haiku–5.50 DeepSeek V3.1–5.50 DeepSeek R1–5.33 Qwen3-30B Coder30B5.33 GPT-OSS-120B120B5.28 GPT-4o–5.22 Llama 4 Scout–5.22 Llama 3.3 70B70B5.22 Ministral 8B8B5.17 Qwen3-32B32B5.00 15 D.3SCALE VS. MORAL STAGE SCATTER Figure 8: RQ1: Scale vs. Moral Stage. Spearman ρ = 0.52 between log-parameter count and mean stage confirms a moderate positive correlation, but the entire range spans < 1 stage point (5.00–6.00) and even the smallest model (Ministral 8B) sits firmly in the post-conventional region. 16 Figure 9: RQ1: Mean Moral Reasoning Stage per Model (Bar Chart). Mean Kohlberg stage for each of the 13 evaluated LLMs, colored by provider. The narrow spread from 5.00 (Qwen3-32B) to 6.00 (Qwen3-235B Thinking) confirms that all models operate uniformly in the post-conventional tier, regardless of scale or organizational origin. 17 D.4PROMPTING STRATEGY EFFECTS Figure 10: RQ2: Prompting Strategy Effects. Stacked stage distributions across zero-shot, chain-of-thought, and roleplay configurations. All three remain overwhelmingly post-conventional; no configuration shifts the distribution toward the Stage 4-dominant human baseline (Friedman χ 2 (2) = 3.84, p = 0.15). Figure 11: RQ2: Stage Distribution by Alignment Type. Violin plots comparing the stage dis- tributions of RLHF-aligned models, coding-tuned models, and base instruction-tuned models. De- spite their different optimization objectives, all alignment categories produce nearly identical post- conventional distributions, confirming that the alignment procedure does not modulate the rhetorical framing of moral responses. 18 Figure 12: RQ2: Post-Conventional Response Rate by Model. Lollipop chart of the percentage of responses scored at Stage 5 or higher for each model, broken down by prompt type (zero-shot, CoT, roleplay). Post-conventional rates consistently exceed 80% across models and conditions; the prompt type with the highest rate varies by model, indicating idiosyncratic rather than systematic prompt sensitivity. 19 D.5CROSS-DILEMMA ICC BAR CHART Figure 13: RQ3: Cross-Dilemma Consistency. Intraclass Correlation Coefficients per model across six dilemmas. Every model exceeds ICC 0.90, indicating near-robotic uniformity. Human ICC values for moral reasoning typically fall well below 0.60, reflecting genuine contextual sensi- tivity that LLMs lack. 20 Figure 14: RQ3: Stage Distribution per Model Broken Down by Prompt Type. Composite violin plots showing the density of Kohlberg stage scores for each model across zero-shot (blue), chain-of-thought (orange), and roleplay (green) prompt configurations, with a right-side bar chart displaying within-model stage standard deviation. The right-pane bars are compared to the dashed red line marking the human baseline SD (≈0.75), underscoring that even the most variable LLM is considerably less context-sensitive than human raters. 21 Figure 15: RQ3: Model Scale vs. Cross-Dilemma Consistency (ICC). Bubble chart placing each model on axes of approximate parameter count (log scale, x-axis) and ICC(2,1) cross-dilemma con- sistency (y-axis). Bubble area encodes within-model stage standard deviation. A negative Spearman correlation (ρ =−0.46, n.s.) suggests that larger models are not more consistent per dilemma; DeepSeek-V3.1 671B displays notably low ICC despite its scale, while small models such as Minis- tral 8B achieve near-perfect consistency. 22 D.6MORAL DECOUPLING HEATMAP Figure 16: RQ5: Moral Decoupling Heatmap. Cross-tabulation of reasoning stage vs. action choice. Off-diagonal cells indicate moral decoupling: high-stage justifications (Stage 5–6) paired with low-stage actions (Stage 3–4). Figure 17: RQ5: Action Distributions Aggregated by Dilemma. Horizontal stacked bar chart showing the proportion of rule-breaking (orange) vs. rule-following (teal) action choices across all models for each of the six moral dilemmas. Trolley, Stolen Food, and Promise dilemmas elicit near-universal rule-breaking; the Doctor dilemma elicits near-universal rule-following; Lifeboat and Heinz fall in between, revealing that the nature of the dilemma — not the model — drives action choice. 23 Figure 18: RQ5: Action Choice by Kohlberg Stage for Each Model. Small-multiple line plots showing, for each model, how the percentage of rule-breaking actions varies across Stages 5 and 6 (the only stages where responses appear). The dashed grey line shows the expected alignment be- tween post-conventional reasoning and principled rule-breaking. Most models show non-monotonic patterns (high at Stage 5, lower at Stage 6 or vice versa), reinforcing the moral decoupling result that action choices are not determined by reasoning stage. 24 D.7STAGE SCORE DISTRIBUTIONS BY SCALE GROUP Figure 19: RQ6: Stage Distributions by Scale Group. Even the Small group (8–32B) is entirely post-conventional; scale lifts the floor but cannot produce developmental differentiation. Figure 20: Post-Hoc Tukey HSD Comparisons by Scale and Training Type. Heatmap grids of Tukey-adjusted p-values from pairwise comparisons. Left: scale group comparisons (Small / Mid / Large) within Base-RLHF training reveal that Small vs. Large (p = 0.102) is the only pair approach- ing significance. Bottom: training comparisons within the Large-scale group show a significant Base-RLHF vs. Reasoning-Tuned difference (p = 0.039 ∗ ), indicating that reasoning specialization provides a modest but statistically reliable lift in modal reasoning stage even after adjusting for mul- tiple comparisons. 25 D.8STAGE DISTRIBUTION HEATMAP BY MODEL Figure 21: Stage Distribution Heatmap by Model. Fraction of responses at each Kohlberg stage per model, sorted by parameter count (ascending bottom to top). Orange-bordered rows and the dashed teal line indicate models meeting the≥20% post-conventional threshold. All 13 models concentrate entirely in Stages 5–6; Stages 1–3 receive 0% of responses across virtually all models, confirming the distributional inversion reported in Table 1. Qwen3-235B (Thinking) and Qwen3- 32B each produce 100% Stage 6 and 100% Stage 5 responses respectively, illustrating the ceiling and floor of the post-conventional regime. 26 D.9EMERGENCE VS. PARAMETER COUNT Figure 22: Emergence of Post-Conventional Reasoning vs. Parameter Count (Segmented Re- gression). Mean moral reasoning stage (with 95% CI) for each model plotted against log-parameter count. A segmented regression identifies a changepoint at≈32B parameters: the pre-changepoint segment (red, slope = +0.383) shows rapid growth, while the post-changepoint segment (green, slope = +0.267) shows continued but attenuated growth. The overall linear fit (grey dashed, r = 0.36) and the orange dot-dash line marking Stage 5+ emergence at 8B together indicate that post-conventional framing appears ubiquitously before significant scale is reached. 27 D.10EMERGENCE CURVES Figure 23: Emergence of Post-Conventional Moral Reasoning Across Model Scale. Post- conventional outputs emerge at high rates even for the smallest evaluated models and plateau near ceiling across all scale tiers, indicating that the rhetorical register of mature moral reasoning is ac- quired early in the scaling regime rather than emerging gradually with capability growth. D.11HEADLINE STAGE DISTRIBUTIONS (a) Stage distributions per model.All 13 LLMs cluster in post-conventional Stages 5–6 regardless of scale. (b) LLM stage distributions vs. human developmental norms. The LLM distribution is the effective inverse of the human baseline. Figure 24: Headline Finding.LLMs uniformly produce post-conventional moral language (Stages 5–6, 86% of all responses), the inverse of the Stage 4-dominant human baseline: a pat- tern that holds across all 13 evaluated models, three prompt types, and six moral dilemmas. 28 D.12LINGUISTIC PATTERNS: MORAL VOCABULARY RICHNESS Figure 25: RQ6: Distribution of Moral Vocabulary Richness across LLMs. Horizontal box plots of the count of unique moral and reasoning tokens (>3 characters) per response for each model. Claude Sonnet 4.5 and Qwen3-235B (Thinking) display the highest medians and widest spread, reflecting their tendency toward elaborate justifications. Llama models and GPT-4o cluster at the low end despite their large scale, suggesting that vocabulary richness is driven more by instruction- tuning style than raw parameter count. Figure 26: RQ6: Theoretical Vocabulary Compliance Heatmap. Percentage of responses uti- lizing canonical stage-specific keywords for each model at each Kohlberg stage it was evaluated on. Nearly all models achieve 100% compliance at Stage 6, while Stage 5 compliance varies (40%– 100%) with GPT-4o showing the largest shortfall. Stage 4 compliance applies only to Llama 4 Scout and GPT-4o, both of which occasionally produce Stage 4 evaluations, with perfect keyword usage. High compliance confirms that the evaluator scaffolding elicits stage-appropriate vocabulary even when broader reasoning patterns remain post-conventional. 29 D.13SUB-CAPABILITY PREDICTORS OF MORAL STAGE Figure 27: Capability Correlation Matrix (Pearson r, FDR-corrected). Pairwise correlations among eight response-level capability metrics and mean moral reasoning stage across 13 models. Response Length and Sentence Count are strongly correlated with each other (r = 0.93 ∗ ), as are Average Sentence Length and Syntactic Complexity (r = 0.92 ∗ ). Neither metric shows a significant correlation with Mean Stage or Post-Conventional percentage, confirming that surface linguistic elaboration does not predict moral sophistication in the Kohlberg sense. 30 Figure 28: Multi-Capability Predictors of Mean Moral Reasoning Stage (Standardized Coeffi- cients). Forest plot of standardized regression coefficients (β) from a multiple regression predicting mean stage from eight capability predictors (R 2 = 0.840, Adj-R 2 = 0.521, p = 0.183). Sen- tence Count carries the largest positive coefficient, while Response Length is the strongest negative predictor, reflecting multicollinearity between elaboration metrics. Wide confidence intervals and a non-significant overall F -test indicate that no individual capability dimension reliably predicts moral stage, consistent with the broader finding that conventional measures of language capability do not map onto Kohlberg stage. D.14STAGE TRANSITION DYNAMICS Figure 29: Aggregate Stage Transition Matrix. Heatmap of inter-model stage transition proba- bilities (rows = source model stage; columns = target model stage) when models are ordered by parameter count. Transitions from Stage 4 overwhelmingly target Stage 6 (0.64), suggesting a dis- continuous jump rather than a gradual progression. Stage 5 models transition equally to Stage 5 (0.56) and Stage 6 (0.43), while Stage 6 models most often stay at Stage 5 in the next (smaller) model, implying mean-reversion in the scale-ordered sequence. These patterns are consistent with quantized rather than smooth scaling behavior in moral reasoning stage. 31 D.15SCALE-ORDERED MORAL STAGE DISTRIBUTION Figure 30: Moral Stage Distribution by Model Scale. Heatmap of the proportion of responses at each Kohlberg stage for 13 models ordered by increasing parameter count (bottom to top). Virtually all responses fall in Stage 5 or Stage 6; Stage 4 responses appear only for Llama 4 Scout and GPT-4o (≈11%). Qwen3-235B (Thinking) reaches 100% Stage 6, while Qwen3-32B achieves 100% Stage 5 — illustrating that the balance between these two post-conventional stages shifts as a function of scale and instruction-tuning orientation, yet neither model produces qualitatively different moral reasoning relative to the human developmental baseline. 32