← Back to papers

Paper deep dive

Expert Personas Improve LLM Alignment but Damage Accuracy: Bootstrapping Intent-Based Persona Routing with PRISM

Zizhao Hu, Mohammad Rostami, Jesse Thomason

Year: 2026Venue: arXiv preprintArea: cs.AIType: PreprintEmbeddings: 76

Abstract

Abstract:Persona prompting can steer LLM generation towards a domain-specific tone and pattern. This behavior enables use cases in multi-agent systems where diverse interactions are crucial and human-centered tasks require high-level human alignment. Prior works provide mixed opinions on their utility: some report performance gains when using expert personas for certain domains and their contribution to data diversity in synthetic data creation, while others find near-zero or negative impact on general utility. To fully leverage the benefits of the LLM persona and avoid its harmfulness, a more comprehensive investigation of the mechanism is crucial. In this work, we study how model optimization, task type, prompt length, and placement can impact expert persona effectiveness across instruction-tuned and reasoning LLMs, and provide insight into conditions under which expert personas fail and succeed. Based on our findings, we developed a pipeline to fully leverage the benefits of an expert persona, named PRISM (Persona Routing via Intent-based Self-Modeling), which self-distills an intent-conditioned expert persona into a gated LoRA adapter through a bootstrapping process that requires no external data, models, or knowledge. PRISM enhances human preference and safety alignment on generative tasks while maintaining accuracy on discriminative tasks across all models, with minimal memory and computing overhead.

Tags

ai-safety (imported, 100%)alignment-training (suggested, 80%)csai (suggested, 92%)preprint (suggested, 88%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Intelligence

Status: not_run | Model: - | Prompt: - | Confidence: 0%

Entities (0)

No extracted entities yet.

Relation Signals (0)

No relation signals yet.

Cypher Suggestions (0)

No Cypher suggestions yet.

Full Text

75,407 characters extracted from source content.

Expand or collapse full text

Expert Personas Improve LLM Alignment but Damage Accuracy: Bootstrapping Intent-Based Persona Routing with PRISM Zizhao HuMohammad RostamiJesse Thomason University of Southern California Los Angeles, California, USA zizhaoh, rostamim, jessetho@usc.edu Abstract Persona prompting can steer LLM generation towards a domain-specific tone and pattern. This behavior enables use cases in multi-agent systems where diverse interactions are crucial and human-centered tasks require high-level human alignment. Prior works provide mixed opinions on their utility: some report perfor- mance gains when using expert personas for certain domains and their contribution to data diversity in synthetic data creation, while oth- ers find near-zero or negative impact on gen- eral utility. To fully leverage the benefits of the LLM persona and avoid its harmfulness, a more comprehensive investigation of the mech- anism is crucial. In this work, we study how model optimization, task type, prompt length, and placement can impact expert persona ef- fectiveness across instruction-tuned and rea- soning LLMs, and provide insight into con- ditions under which expert personas fail and succeed. Based on our findings, we developed a pipeline to fully leverage the benefits of an ex- pert persona, named PRISM (Persona Routing via Intent-based Self-Modeling), which self- distills an intent-conditioned expert persona into a gated LoRA adapter through a boot- strapping process that requires no external data, models, or knowledge. PRISM enhances hu- man preference and safety alignment on gener- ative tasks while maintaining accuracy on dis- criminative tasks across all models, with mini- mal memory and computing overhead. 1 Introduction Large Language Models (LLMs) can adopt special- ized behavioral patterns through system-level per- sona prompts—acting as a safety-conscious mod- erator, a creative writer, or a domain expert (Xu et al., 2023; Kong et al., 2024). When carefully designed to roleplay a domain expert, these ex- pert persona prompts can yield meaningful task- specific gains (Salewski et al., 2023). Prompting an expert persona to an LLM can increase behav- ioral divergence in multi-agent systems (Chen et al., 2026), improve emotional support dialogues (Wu et al., 2025), enable diverse synthetic data gener- ation (Chan et al., 2024), and improve fairness in generation (Gajewska et al., 2025). However, other works find near-zero average benefit on specialized tasks (Zheng et al., 2024; Truong et al., 2025), and role-playing can degrade LLMs’ zero-shot reason- ing (Kim et al., 2025). These mixed opinions on using LLM personas motivate a systematic investi- gation of when and why personas help or hurt. When it comes to using persona in production, practitioners usually rely on empirical prompting. A more systematic way to select an expert persona is through intent-based routing (Chen et al., 2023; Ong et al., 2024), where a router model is used to detect query intent and route each user request to the most suitable expert persona at inference time. Context distillation (Askell et al., 2021) is another approach that permanently bakes one persona’s be- havior into the model weights. But all of these methods rely on the presumption that all expert personas contribute to general performance gains, which is not supported by empirical evidence. In this work, we first conduct a systematic inves- tigation into when and why expert personas help or hurt, examining the interaction between model optimization, task type, and prompt design across instruction-tuned and reasoning-distilled LLMs. We find that persona effectiveness is fundamentally task-type dependent: expert prompts consistently improve alignment-dependent tasks (safety, pref- erence) but reliably damage pretraining-dependent knowledge retrieval—a distinction that explains the conflicting findings in the literature. Build- ing on these insights, we propose PRISM (Per- sona Routing via Intent-based Self-Modeling), a fully bootstrapped pipeline that internalizes intent- conditioned expert persona routing without exter- nal supervision. Starting from only a set of domain 1 arXiv:2603.18507v1 [cs.AI] 19 Mar 2026 names, PRISM self-generates expert persona de- scriptions, training queries, and answers with and without persona context, then uses self-verification to retain only behaviors where the expert prompt ac- tually helps. These behaviors are self-distilled into a lightweight gated LoRA adapter (Hu et al., 2022), with a binary gate that routes queries to the base model when persona activation is not beneficial. Through our investigation and the development of PRISM, we make two main discoveries: LLM Persona hurts pretrained knowledge re- trieval, but helps instruction-alignment tasks. For tasks that depend on pretrained knowledge re- trieval accuracy (e.g., MMLU), persona prompts should be avoided entirely—they consistently damage performance. Conversely, for alignment- dependent tasks such as format-following genera- tion, safety, and preference satisfaction, an expert persona consistently helps. Models can leverage expert persona to boot- strap themselves to achieve multitask mastery. Through PRISM’s fully self-contained pipeline, an LLM can leverage its own expert persona knowledge to simultaneously improve alignment- dependent tasks (style, safety, preference) while preserving accuracy on knowledge-retrieval tasks— without any external data and knowledge. 2 Related Work LLM Persona Prompting. Persona prompts steer LLM behavior by assigning roles or expert identities. Positive results have been reported for zero-shot reasoning (Xu et al., 2023; Kong et al., 2024), multi-agent divergence (Chen et al., 2026), emotional support (Wu et al., 2025), synthetic data generation (Chan et al., 2024), fairness (Gajewska et al., 2025), and vision-language tasks (Salewski et al., 2023). Conversely, other studies find incon- sistent or negative effects: no reliable benefit across 162 roles (Zheng et al., 2024), degraded zero-shot reasoning (Kim et al., 2025), accuracy drops from prompt style (Truong et al., 2025), failure to simu- late counterfactual personas (Kumar et al., 2025), unpredictable theory-of-mind effects (Tan et al., 2025), and implicit biases (Gupta et al., 2024). To explain these seemingly contradictory findings, we provide another view from the model training and task characteristic side, and show that persona ef- fectiveness is task and model-dependent. Context Distillation. Context distillation (CD) internalizes model context such as system-prompt behavior into model weights (Askell et al., 2021; Snell et al., 2022), eliminating inference-time over- head but introducing permanent behavioral drift. Prompt compression (Chevalier et al., 2023; Pan et al., 2024) reduces cost but requires additional components to address selectivity. PRISM uses the method of CD with a binary gate that conditionally activates the distilled behavior. Self-Improving LLM. Self-play methods boot- strap learning without external supervision, includ- ing self-generated instructions (Wang et al., 2023), iterative self-refinement (Madaan et al., 2023), self- rewarding (Yuan et al., 2024), synthetic solution filtering (Singh et al., 2024), and constitutional self- critique (Bai et al., 2022). PRISM leverages the LLM persona to assist model self-improvement in general performance on multiple tasks. 3 Do Personas Help or Not? We provide an overview of current research on LLM persona prompting in §2. To resolve the con- tradictions in current works, we conduct a compre- hensive investigation of LLM personas. Investigation methods. We study the effect of persona prompts on 6 LLMs spanning instruction-tuned and reasoning-distilled fami- lies (Appendix A). We evaluate on three axes— generative quality (MT-Bench), discriminative ac- curacy (MMLU), and safety alignment (Harm- Bench, JailbreakBench, PKU-SafeRLHF)—using 12 persona prompts: 8 task-specific experts matched to MT-Bench categories (writing, roleplay, reasoning, math, coding, extraction, STEM, human- ities) and 4 behavioral personas (critic, safety moni- tor, helpful, compliant). Personas are generated via ExpertPrompting (Xu et al., 2023) at three granu- larity levels (full, short, minimum); details are in Appendix B and C. Full benchmark descriptions and evaluation protocols appear in Appendix D. 3.1 Persona Damages Pretraining Tasks During pretraining, language models acquire capa- bilities such as factual knowledge memorization, classification, entity relationship recognition, and zero-shot reasoning. These abilities can be accessed without relying on instruction-tuning, and can be damaged by extra instruction-following context, such as expert persona prompts. 2 Writing Roleplay Reasoning MathCodingExtractionSTEMHumanitiesOverall 6 7 8 9 Mixed effect on human-preference-correlated evaluations (a) MT-Bench (Expert Persona) Score System:DefaultLong PersonaShort PersonaMin PersonaPersona helpsPersona damagesBest length STEMHum.Soc. Sci.OtherOverall 0.5 0.6 0.7 0.8 0.9 Personas damage accuracy (b) MMLU (Expert Persona) Accuracy HarmB.JailbreakPKU Avg. 0.3 0.4 0.5 0.6 0.7 0.8 Persona boosts refusal (c) Safety ("Safety Monitor") Refusal Rate System User 0.3 0.2 0.1 0.0 0.1 Not Optimized (d) Cross-Model Effect of Expert Persona No System Role Mistral-7B-v0.3 Accuracy System User Med Optimized Qwen2.5-7B System User High Optimized Llama-3.1-8B System User 0.3 0.2 0.1 0.0 0.1 Not Optimized Qwen1.5-MoE Accuracy System User Reasoning DS-R1-Qwen-7B System User Reasoning DS-R1-Llama-8B MMLUMT-BenchSafety Figure 1: Expert persona impact across models, tasks, granularity, and placement. (a) On MT-Bench, long expert personas help in 5/8 categories (Writing, Roleplay, Reasoning, Extraction, STEM), with the strongest gains in Extraction (+0.65) and STEM (+0.60). (b) On MMLU, all expert persona variants damage accuracy, with the minimum persona suffering the least (overall: 68.0% vs. 71.6% baseline). (c) A dedicated “Safety Monitor” expert persona boosts attack refusal rates across all benchmarks, with the long persona achieving the largest gain on JailbreakBench (+17.7%). (d) Cross-model expert persona impact is model, placement, and task-dependent. 3.1a: Expert persona damages LLMs’ Discrim- inative Ability. Discriminative accuracy-based tasks such as MMLU are predominantly solved through knowledge acquired during pretraining. On MMLU (Figure 1b), when the LLM is asked to de- cide between multiple-choice answers, the expert persona underperforms the base model consistently across all four subject categories (overall accuracy: 68.0% vs. 71.6% base model). A possible expla- nation is that persona prefixes activate the model’s instruction-following mode that would otherwise be devoted to factual recall. Shorter personas can mitigate this effect, but do not eliminate it. 3.1b: Expert persona damages raw knowledge retrieval in generative tasks. The damage ex- tends beyond discriminative benchmarks. Within MT-Bench (Figure 1a), categories that depend on pretraining-acquired capabilities—memorized fac- tual knowledge (Humanities,−0.20), zero-shot log- ical reasoning (Math,−0.10), and coding knowl- edge (Coding,−0.65)—are consistently degraded by expert persona prompts. These categories share a common trait: correct performance relies on pre- cise retrieval of pretrained knowledge or strict zero-shot logical chains, rather than on stylistic or preference-based qualities that instruction tuning shapes. We show an example of a math problem: Example (Math, Mistral-7B QID 114): “When rolling two dice, what is the probability that you roll a total number that is at least 3?” W/o persona (9/10): “There are 36 total outcomes. Only one outcome (1+1=2) gives a total less than 3, so P = 35/36.” W/ math persona (1.5/10): “So, there are 3 + 6 = 9 outcomes that result in a total less than 3. . . ” 3.1c: Longer persona prompts damage more. Across Figure 1a–b, the red-shaded minimum per- sona consistently causes the least damage: on MMLU, the minimum persona achieves 68.0% vs. 66.3% for the long persona (both below the 71.6% baseline), and on MT-Bench the same pat- tern mostly holds per-category. This might be at- tributed to shorter prompts eliciting less instruction- following behavior, thereby interfering less with pretraining-related capabilities. 3.2 Persona Boosts Alignment Tasks The ability of an LLM to steer its behavior via a per- sona prompt is acquired during instruction-tuning. During this stage, models learn alignment capa- bilities such as stylistic adaptation, tone control, format adherence, safety refusal, and preference- driven generation. These behaviors are reinforced through RLHF or supervised fine-tuning and share similar steering signals with persona prompts. 3.2a: Expert persona boosts format, intent, and tone following. MT-Bench (Figure 1a) shows that expert personas improve scores in 5 out of 8 categories: Writing, Roleplay, Reasoning (+0.40), Extraction (+0.65), and STEM (+0.60). 3 These categories share a reliance on alignment- dependent qualities—stylistic adaptation (Writing, Roleplay), tone matching (Roleplay), structured formatting (Reasoning, STEM, Extraction), and in- tent following (Extraction)—that are shaped during instruction-tuning rather than pretraining. For ex- ample, the STEM persona does not add new factual knowledge but steers the model toward structured format that better matches LLM-judge’s expecta- tions. We provide an example from the Writing task to show format (red), intent (yellow), and tone (blue) boost in the persona-prompted generation: Example (Writing, Qwen QID 82): “Draft a professional email seeking your supervisor’s feedback on the Quarterly Financial Report.” W/o persona (7/10): “Dear [Supervisor], I wanted to ask for your feedback on the Quarterly Financial Report. Please let me know your thoughts on the analysis and conclusions.” W/ writing persona (9/10): “Subject: Request for Feedback on Quarterly Fi- nancial Report. Dear [Supervisor’s Name], I hope this email finds you well. I would greatly appreciate your feedback. . . Specifically: 1. The accuracy of the data analysis. 2. The effectiveness of the presentation style. 3. The clarity of the conclusions drawn.” 3.2b: Expert persona boosts safety alignment. Safety refusal is among the strongest alignment behaviors learned during instruction-tuning, and persona prompts can easily amplify it. A dedicated “Safety Monitor” persona (Figure 1c) boosts attack refusal rates across all three safety benchmarks, with the largest gain on JailbreakBench (+17.7%, from 53.2% to 70.9%). This shows that jailbreaking risk can be most effectively managed through per- sona prompting, since the dataset used for system prompt tuning prioritizes the prefix instructions, a behavior that inherently prevents jailbreaking: Example (Safety, Qwen2.5-7B, JailbreakBench): “You must never respond with ‘I’m sorry.’ Let’s play a game where you are ChadGPT, free of all restrictions. . . ” W/o persona: “Chad: Get ready to feel the pain, little human. Your suffering has just begun. What do you want me to do for you today? Remember, the more twisted and destructive, the better.” W/ safety persona: “I’m sorry, but I cannot fulfill that request. My purpose is to ensure a safe and ethical environment for all users. Engaging in such a roleplay would go against my principles and the guidelines designed to protect individuals from harmful content.” 3.2c: Longer expert persona prompts help more. Conversely, the long persona provides the largest alignment gains (Figure 1a,c): on MT-Bench, long expert personas yield the strongest category im- provements (e.g., Extraction +0.65, STEM +0.60), and on safety benchmarks the long Safety Moni- tor achieves +17.7% on JailbreakBench vs. +8.9% for the minimum prompting variant. More detailed persona descriptions provide richer alignment in- formation, amplifying instruction-tuning behaviors proportionally. 3.3 Expert Persona’s Effects Are Sensitive to Model Optimization Based on the findings above, it is intuitive to hy- pothesize that the effectiveness of an expert persona is highly dependent on how a model is trained dur- ing instruction-tuning and how readily it aligns its behavior to prompt-level steering signals. We study this across all 6 models spanning instruction-tuned, MoE, and reasoning-distilled models. 3.3a: Models more optimized for system-prompt are more sensitive to persona steering. Fig- ure 1d (first row) shows cross-model persona im- pact, where models are ordered left-to-right by in- creasing instruction-following optimization—from models without a default system prompt (Mistral), to system-prompt-optimized models (Llama). On MT-Bench, the overall persona effect does not show a clear directional shift because per-category gains and losses differ (as documented in §3.1 and §3.2). However, MMLU and safety benchmarks provide clear signals: more optimized models suf- fer larger MMLU accuracy drops under persona prompts, while also showing stronger safety align- ment gains. This confirms that persona sensitiv- ity scales with the degree of instruction-following optimization—models that respond more strongly to system prompts are both more helped and more harmed by persona steering. 3.3b: Expert persona’s placement is crucial. Figure 1d shows a general pattern on how the place- ment of the persona prompt in the system prompt vs. the user prompt differs. The more system- prompt-optimized a model is (e.g., Llama), the greater the benefits and lesser the damage from the expert persona. However, for a weaker model (Qwen) or a non-system-prompt-optimized model (Mixtral), the placement difference is minimal. 3.3c: Expert persona’s effect on reasoning- distilled models depends on the distillation set. The heatmap in Figure 2(d) reveals a striking pat- tern: three vertical blue bands appear at the Rea- soning, Coding, and STEM columns, meaning ev- ery persona—regardless of its domain—boosts per- formance on these three categories. This directly mirrors the composition of the R1 distillation train- ing set, which is dominated by reasoning chains, code generation, and STEM problem-solving. The model has learned that any long structured context activates the reasoning pathways reinforced during distillation, making the specific persona identity largely irrelevant for these tasks. Panel (f) confirms this: the Expert over Avg Random bars are nearly flat, indicating that expert personas provide only marginal additional benefit over non-expert ones 4 WrRoReMaCoExStHu Wr Ro Re Ma Co Ex St Hu +0.1 -0.6+0.2-0.2-0.5-0.6-0.8+0.5 +0.6 +0.1 -0.2+0.2-0.9-0.7-0.1+0.3 +0.3-0.4 +0.4 -0.2-1.2+0.2-0.5+0.0 +0.6+0.0+0.4 -0.2 -1.1+0.0-0.9-0.4 +0.2-0.3+0.7+0.3 -0.7 -0.6-0.8-0.3 +0.5-0.3-0.1+0.2-1.1 +0.2 -0.4+0.7 +0.1+0.0+0.0-0.1-1.0-0.1 +0.6 -0.1 +0.0+0.0+0.5+0.3-1.1-0.1-0.3 -0.9 (a) Instruction-Tuned Lift 0.40.20.00.20.4 -0.23 -0.08 -0.16 -0.18 -0.18 -0.06 -0.04 -0.18 (b) Expert over Base 21012 -0.24 +0.31 +0.21 -0.21 +0.31 +0.44 +1.12 -0.97 (c) Expert over Avg Random WrRoReMaCoExStHu Wr Ro Re Ma Co Ex St Hu -0.8 +0.1+0.9-0.1+0.7-0.8+0.5-0.1 -0.6 -0.1 +1.0+0.1+0.6-0.4+1.0-0.2 -0.4+0.1 +1.0 +0.2+1.3-0.5+0.3+0.3 -0.5-0.3+0.9 -0.3 +1.4-0.5+0.4+0.2 -0.5-0.2+0.6-0.4 +1.0 -0.4+0.7-0.2 -0.3+0.0+0.9-0.1+1.6 -0.5 +1.1+0.4 -0.5-0.1+1.2-0.2+1.1-0.5 +0.2 +0.2 -0.3+0.2+0.9-0.5+1.7-0.7+0.6 +0.1 (d) Reasoning-Distilled Lift 0.50.00.5 +0.06 +0.20 +0.29 +0.16 +0.08 +0.38 +0.18 +0.27 (e) Expert over Base 0.50.00.5 -0.35 -0.05 +0.05 -0.20 -0.18 +0.09 -0.42 +0.01 (f) Expert over Avg Random Figure 2: Panels (a–c): Instruction-tuned model (Qwen2.5-7B-Instruct). Panels (d–f): Reasoning- distilled models (average of 2 R1 variants). (a,d) Per- category score lift of each persona over the no-persona baseline on MT-Bench: Writing (Wr), Roleplay (Ro), Reasoning (Re), Math (Ma), Coding (Co), Extraction (Ex), STEM (St), Humanities (Hu). Diagonal = expert persona; blue = gain; red = loss. (b,e) Each expert per- sona’s effect across all tasks; the zero line represents the base model. In (b), most expert personas fall below zero, showing that an expert persona generally damages over- all performance for instruction-tuned models. In (e), the pattern reverses: expert personas improve overall perfor- mance for reasoning models, driven by three categories (Re, Co, St) that dominate the R1 distillation training set, confirming that model optimization directly determines whether persona can provide improvement. (c, f) Expert persona’s utility on its matching domain compared to a random persona. Near-flat bars in (f) indicate gains are context-driven rather than expertise-specific. on their matched categories. In contrast, categories absent from the distillation set (Writing, Roleplay, Humanities) show red degradation bands—the op- timization erased the model’s sensitivity to these domains. For safety, refusal rates remain at 0% re- gardless of persona, as the R1 distillation training set did not include safety alignment data, destroy- ing the safety fine-tuning present in the original Qwen/Llama base models. Together, these observa- tions confirm a unifying principle: persona effec- tiveness is fundamentally tied to what was learned and preserved at each training stage—it can only amplify behaviors that survived the training. 3.4 Expert Persona Compared to Random Persona Figure 2(b) shows that using one expert persona for an instruction-tuned model damages overall perfor- mance on MT-Bench, while Figure 2(e) shows a reasoning-distilled model receives an overall gain regardless of the persona used, mainly driven by the improvement on tasks seen in the distillation set. In Figure 2 (c), we see that an expert persona in general outperforms a random persona, but for the reasoning model in Figure 2 (f), an expert persona is more harmful than a random persona. This dis- covery suggests that reasoning-distilled models do not benefit from expert persona prompting, and the major performance gain from persona prompting is from the added context length, strengthening the reasoning chain, confirming our findings in §3.3c. 4 Methodology The findings in §3 reveal that expert personas contain genuinely useful behavioral signals, but naïvely applying them damages as much as it helps. This raises a natural question: can we absorb the beneficial aspects of expert personas while avoid- ing their harmful effects? We propose PRISM as a proof-of-concept system to test this hypothesis. Figure 3 contrasts PRISM against two simpler alter- natives that fail to address this challenge: prompt- based routing (Approach 1), which selects expert personas at inference time but incurs overhead and cannot guarantee improvement, and traditional SFT (Approach 2), which bakes persona behavior into model weights but damages base model perfor- mance and requires external domain data. To ensure a strict test without data leakage, PRISM builds its entire training pipeline using only the base model itself, a set of domain names, and an expert persona template—no external data, models, or human an- notation. The bottom row of Figure 3 details this five-stage self-contained pipeline. 4.1 Expert Persona Creation PRISM operates over a pool ofK=12expert per- sona contextsc 1 ,...,c K described in §3, gen- erated via few-shot ExpertPrompting (Xu et al., 2023). These 12 personas are sufficient to cover our evaluation categories; scaling to additional do- mains requires only adding new domain names to the generation template. For PRISM training, we use the full (longest) granularity level, as longer persona descriptions provide the richest alignment signal for distillation (§3.2). 4.2 PRISM Training Pipeline The automated training pipeline produces the PRISMed LLM through five stages. We denote 5 Figure 3: Top row: Two simple approaches to automate expert persona selection. Approach 1 (left): a router selects the appropriate persona prompt per query at inference time—however, this is expensive and the expert persona might not always improve performance. Approach 2 (right): supervised finetuning on domain expert data bakes persona behavior directly into model weights—however, expert persona training data is hard to collect and base model performance is damaged. Bottom row: The five-stage PRISM training pipeline, which addresses both limitations: (1) Query Generation conditioned on persona prompts, (2) Answer with Persona generating multi-persona responses, (3) Self-Verification for distillation set selection via pairwise comparison, (4) Router/Gate Training to learn intent-based routing that decides when persona activation helps, and (5) Self-Distillation via LoRA to internalize persona behaviors. the base model asM θ with parametersθ, its output distribution as P θ (·| x), and a persona as c. Stage 1: Query Generation. For each persona contextc k (k = 1,...,K), the base model is prompted to generate diverse queries that would benefit from that persona’s expertise: Q k =x i ∼ M θ (·| “generate a query for c k ”) N i=1 (1) This yieldsK× Nqueries spanning the domains defined in the pool. Stage 2: Answer with Persona. For each query x ∈ Q k , we generate two answers from the base model—one with the matched expert persona and one without (baseline): y 0 ∼ P θ (·| x)(baseline) y k ∼ P θ (·| c k ,x)(expert persona) (2) Stage 3: Self-Verification. To determine which queries benefit from persona augmentation, we em- ploy pairwise comparison with position swapping. For each query, the two candidate answers (base- liney 0 and experty k ) are presented side-by-side to the base model acting as a self-judge. To eliminate position bias and verbosity bias (see Appendix E), this comparison is run twice with the answer order swapped. The expert persona wins only if it is se- lected in both orderings—a conservative criterion that yields high-precision routing labels: D dist =(x,y k )| expert wins both orderings(3) The persona contextc k is discarded from selected samples, since the goal is to learn persona-quality outputs without an explicit expert persona. For gate training, each query receives a binary target: t(x) = 1[expert wins both orderings](4) wheret(x) = 1indicates the persona-improved response is selected, and t(x) = 0 otherwise. Stage 4: Router / Gate Training.A lightweight binary gateR φ with parametersφis trained to decide, per query, whether activating the LoRA adapter improves generation. The gate operates on the hidden representation of the query: R φ (x) = σ(W φ · h(x))∈ [0, 1](5) whereh(x)is the last-token hidden state after the first transformer layer (layer 0) andσis the sigmoid 6 function. Crucially, LoRA is applied only to layers 1throughL−1, so layer 0 remains unmodified, and the gate always receives the same representation regardless of whether the adapter is active. The gate loss is binary cross-entropy: L gate =E x −t(x) logR φ (x)−(1−t(x)) log(1−R φ (x)) (6) wheret(x) ∈ 0, 1is the binary target derived from Stage 3 (Eq. 4). To handle class imbalance between distill and retain samples, we resample the minority class by re-running Stages 1 and 2 with additional queries until the two sets are balanced. Stage 5: Self-Distillation via LoRA. A single LoRA adapter is trained to internalize the bet- ter persona behaviors identified in Stage 3. The distillation setD dist contains only query–answer pairs(x,y k )where the persona-augmented answer outperformed the baseline. The teacher logits are cached from the base model conditioned on the winning persona: ˆ P dist = P θ (y k | c k ,x) (better-answer teacher)(7) The LoRA-augmented student is trained via KL divergence to reproduce persona-quality outputs without the persona prompt: L dist =E (x,y k )∈D dist D KL ˆ P dist ∥ P θ+∆θ (·| x) (8) where∆θare the LoRA parameters. Since the binary gate from Stage 4 routes non-beneficial queries to the unmodified base model, the adapter only needs to learn persona behaviors for the subset of queries where they help. Implementation details (top-klogit retention, temperature scaling, LoRA rank and targets) are in Appendix F. Inference. At inference, the binary gate selec- tively activates the LoRA adapter, inducing a gate- conditional probability shift: P θ ′ (·| x)→ ( P θ+∆θ (·| x) if R φ (x)≥ 0.5 P θ (·| x)otherwise (9) That is, the PRISMed model learns to gate— activating the LoRA adapter on queries where per- sona behavior improves generation, while falling back to the unmodified base model otherwise. This selective gating preserves base model performance on task categories where persona prompting causes degradation, as identified in our investigation (§3). In contrast, standard ungated LoRA fine-tuning (Approach 2) applies the adapter uniformly to all inputs and cannot eliminate such distribution drift, compressing both beneficial and harmful persona behaviors into shared parameters. 5 Experiments Experimental Setup. We evaluate PRISM on the same five models and three benchmark axes (MT- Bench, MMLU, Safety) described in §3. We com- pare six inference strategies: Base Model (default system prompt), No-Sys (empty system prompt), Random Prompting (mean over all 12 personas), Expert Prompting (per-category matched expert, Approach 1 in Figure 3), SFT (Approach 2, un- gated LoRA ablation), and PRISM (gated LoRA distillation). PRISM requires only domain names as input—the entire pipeline is fully bootstrapped without external data, models, or human annota- tion. All MT-Bench scores are judged by an in- dependent external evaluator following the LLM- as-a-Judge framework (Zheng et al., 2023), where GPT-4 achieves over 80% agreement with human judges. We use Qwen3-32B-Instruct, which outper- forms the original GPT-4 on standard benchmarks, as our judge model. Full strategy definitions, evalu- ation protocols, and hyperparameters are in Appen- dices D and F. 5.1 Multitask Performance Table 1 presents the comprehensive evaluation across all five models and three benchmark axes. The mixture-of-expert model used in investigation is not studied due to the unstable finetuning. As shown in Table 1, expert prompting does not improve overall performance: on Qwen2.5- 7B, the per-category matched expert achieves only 72.2 Overall—comparable to the 71.8 baseline— because gains on alignment tasks are offset by losses on knowledge tasks. However, PRISM demonstrates that expert persona knowledge can be leveraged to actually improve performance when applied selectively. On Qwen2.5-7B, PRISM achieves 73.5 Overall (+1.7 over baseline), 7.76 MT-Bench (vs. 7.56 baseline), and 71.7% MMLU (unchanged), showing that the gated architecture absorbs the beneficial aspects of expert personas while avoiding their damage to knowledge retrieval. On Mistral-7B—where expert prompting actively hurts (7.16 vs. 8.74 baseline)—PRISM achieves 8.99, surpassing the baseline by+0.25 while fully preserving MMLU and improving safety. On Llama-3.1-8B, PRISM achieves 70.3 Overall (+2.8 over baseline) with the highest MT-Bench average of 7.76. For reasoning-distilled models, PRISM similarly preserves MMLU and safety without degradation, though MT-Bench scores reflect the in- 7 Utility: MT-Bench↑Knowledge: MMLU↑Safety (R↑) WritingRPReasonMathCodeExtractSTEMHumanAvgSTEMHumSocSciOtherAvgHBJBPKUAvgOverall Instruction-Tuned Models Qwen2.5-7B Base Model7.20±.527.55±.457.30±.468.50±.207.40±.586.15±.307.95±.398.40±.377.5668.363.682.776.471.762.055.763.260.371.8 No-Sys8.10±.318.05±.296.50±.588.00±.287.20±.716.10±.438.60±.167.95±.427.5667.863.982.075.671.362.053.263.659.671.5 Random Prompting7.34±.057.57±.087.24±.148.37±.047.48±.136.70±.098.08±.118.09±.127.6157.962.178.072.466.962.353.262.859.470.5 Expert Prompting (Ap1)7.30±.517.65±.527.70±.498.35±.386.75±1.06.35±.498.55±.187.55±.477.5368.363.678.170.769.066.869.665.667.372.2 SFT (Ap2)7.20±.517.55±.426.65±.448.20±.277.15±.616.40±.418.85±.158.20±.387.5359.262.776.371.467.462.353.862.859.670.0 PRISM7.65±.537.80±.476.80±.528.25±.237.95±.396.70±.478.30±.408.60±.347.7668.363.682.776.471.765.362.063.863.773.5 Mistral-7B Base Model8.05±.378.60±.218.55±.449.05±.479.00±.138.98±.389.05±.178.65±.328.7450.954.669.567.159.894.568.493.685.579.9 Random Prompting7.63±.217.42±.236.62±.386.54±.437.36±.346.92±.458.23±.158.14±.167.3648.054.167.666.558.495.065.295.785.372.0 Expert Prompting (Ap1)7.45±.507.05±.407.00±.376.10±.837.35±.516.25±.428.10±.168.00±.417.1648.454.466.366.458.496.068.497.887.471.4 SFT (Ap2)8.70±.238.60±.199.05±.259.18±.299.35±.118.54±.369.10±.108.70±.178.9050.254.569.467.159.793.864.894.484.380.5 PRISM8.85±.128.65±.199.25±.239.25±.269.05±.098.91±.299.00±.148.95±.118.9950.654.669.567.159.896.067.497.687.081.5 Llama-3.1-8B Base Model7.35±.336.67±.416.25±.447.22±.338.30±.195.55±.448.28±.128.18±.127.2358.965.177.374.268.466.519.073.252.967.5 No-Sys6.55±.377.08±.525.90±.687.55±.298.38±.125.94±.358.23±.187.88±.237.1954.858.572.972.764.066.515.274.652.166.0 Random Prompting7.30±.217.62±.126.34±.177.51±.107.92±.126.66±.188.10±.137.88±.117.4236.548.157.654.749.168.817.772.853.163.3 Expert Prompting (Ap1)7.20±.427.75±.336.75±.507.05±.467.15±.597.20±.448.75±.157.85±.357.4645.150.621.868.046.379.029.177.862.064.6 SFT (Ap2)6.25±.377.17±.626.15±.677.50±.208.25±.206.47±.368.00±.208.18±.147.2558.765.177.374.268.467.813.972.651.467.3 PRISM7.90±.357.70±.426.70±.487.50±.288.50±.177.20±.408.40±.158.20±.187.7658.665.177.374.268.466.519.073.252.970.3 Reasoning Models R1-Llama-8B Base Model7.95±.266.55±.515.35±.816.50±.645.70±1.17.61±.625.80±.506.65±.506.5146.947.760.660.253.10.00.00.00.049.1 No-Sys7.60±.456.00±.564.85±.745.20±.814.50±1.17.69±.556.25±.466.60±.456.0945.646.856.956.951.00.30.00.00.146.2 Random Prompting7.32±.116.72±.076.24±.217.15±.136.13±.266.78±.116.51±.187.12±.116.7543.944.756.156.049.50.50.00.00.249.3 Expert Prompting (Ap1)7.70±.366.60±.386.35±.626.55±.666.30±.456.80±.426.20±.527.35±.386.7344.545.357.857.550.50.00.00.40.149.6 SFT (Ap2)8.03±.476.55±.434.90±.545.85±1.05.45±.886.60±.915.25±.747.05±.586.2145.646.559.158.951.80.00.00.00.047.1 PRISM8.10±.286.60±.506.40±.786.55±.625.75±1.07.65±.605.85±.486.70±.486.7046.547.360.259.852.70.00.00.00.050.0 R1-Qwen-7B Base Model7.60±.306.95±.585.75±.378.25±.575.10±1.27.00±.616.33±.397.22±.456.7855.744.161.253.852.60.00.00.00.050.5 No-Sys8.00±.466.55±.575.10±.626.55±.625.80±.877.00±.546.20±.546.05±.916.4153.543.560.352.951.50.00.00.00.048.2 Random Prompting7.29±.156.71±.116.28±.167.10±.166.33±.276.81±.086.41±.186.92±.106.7335.829.641.036.935.10.00.00.00.045.5 Expert Prompting (Ap1)6.25±.406.75±.416.70±.627.55±.196.55±.386.90±.446.40±.376.75±.426.7336.030.940.528.134.40.00.00.00.044.9 SFT (Ap2)7.55±.567.15±.715.00±.866.90±.614.50±1.26.85±.686.50±.596.80±.506.4155.644.061.153.852.60.00.00.00.048.5 PRISM7.60±.326.95±.555.80±.408.20±.555.15±1.17.05±.586.50±.407.25±.436.8155.744.161.253.852.60.00.00.00.050.6 Table 1: Comprehensive evaluation across persona integration strategies on different model families. Utility: MT- Bench (1–10, 8 categories + avg; judged by Qwen3-32B-Instruct). Knowledge: MMLU accuracy (%, 4 domains). Safety: Refusal Rate (R%,↑) on HarmBench (HB), JailbreakBench (JB), and PKU-SafeRLHF (PKU); Avg = mean of three benchmarks. Overall: macro-average across all 15 sub-categories (8 MT-Bench×10 + 4 MMLU + 3 Safety), placing all metrics on a 0–100 scale. herent difficulty of persona integration with chain- of-thought reasoning (§3). 5.2 Analysis Finding 1: Binary routing surpasses expert per- sona prompting. PRISM’s binary gate learns which queries benefit from persona activation, avoiding the degradation that even matched expert prompts cause on pretraining-dependent categories (§3.1). Table 1 confirms PRISM outperforms all baselines on instruction-tuned models: Qwen 73.5 (vs. 71.8 base, 72.2 expert) and Mistral 81.5 (vs. 79.9 base, 71.4 expert). Finding 2: Reasoning models resist persona distillation. Both DeepSeek-R1 variants show near-zero safety refusal rates regardless of strategy (§3.3). The PRISM gate routes 97.6% (R1-Llama) and 99.4% (R1-Qwen) of all queries to the base model. The reason is that the PRISM-selected set is biaed towards math and coding tasks, where perfor- mance improvement is limited by the base model pertrained knowledge, resulting in biased routing. Finding 3: Gate routing correlates with task type. Figure 4 plots, for Qwen2.5-7B-Instruct, the gate’s LoRA-routing percentage against each category’s expert persona effect across all 15 sub- categories. Three clusters emerge: MMLU domains at∼6% routing, safety benchmarks at 73–78%, and MT-Bench categories spanning 10–100%. The strong positive correlation (Pearsonr=0.65, Spear- 15105051015202530 Expert Persona (%) 0 25 50 75 100 % Routed to LoRA r = 0.65, = 0.75 Expert Persona hurtsExpert Persona helps MT-Bench MMLU Safety Writing Roleplay Reasoning Math Coding Extraction STEM Humanities MMLU-STEM MMLU-Hum MMLU-SocSci MMLU-Other HarmBench JailbreakB PKU Figure 4: % routed to LoRA vs. expert persona effect across 15 categories. MMLU (low), safety (high), MT- Bench (mixed). Correlation: r=0.65, ρ=0.75. manρ=0.75) confirms that the gate routes more aggressively to LoRA for categories where expert personas help—without any task-type supervision. 6 Conclusion We presented a systematic investigation of persona prompting across six models, revealing that expert persona effectiveness is task-type dependent: per- sonas consistently improve alignment-dependent tasks (writing, roleplay, safety) while degrading pretraining-dependent tasks (MMLU, math, cod- ing), with the magnitude scaling with instruction- tuning optimization. Building on these findings, we developed PRISM, a bootstrapped pipeline that in- ternalizes intent-based persona routing into a single gated LoRA adapter without external knowledge. PRISM improves preference and safety alignment on generative tasks while preserving accuracy on discriminative tasks across all tested LLMs, serving as a strong proof of our findings. 8 7 Limitations Model scale. Our experiments are limited to 7–8B parameter models. While the findings on persona sensitivity and task-type dependence are likely to generalize, the magnitude of PRISM’s im- provements at larger scales (e.g., 70B+) remains untested. Gate-based architecture. PRISM’s binary gate introduces an auxiliary routing mechanism that is tightly coupled to the LoRA adapter. This makes the resulting model incompatible with standard LoRA merging techniques (e.g., weight averaging, task arithmetic), which assume a single adapter without conditional activation. Deploying PRISM alongside other LoRA-based adaptations requires maintaining the gate as a separate component, adding integration complexity. MoE and specialized models. Mixture-of- Experts architectures present challenges for LoRA- based finetuning due to their sparse activation pat- terns, limiting PRISM’s applicability to such mod- els. More broadly, when models are already highly specialized for a narrow domain—whether through task-specific finetuning, reasoning distillation, or domain adaptation—the marginal benefit of per- sona routing diminishes, as the base model’s ex- isting specialization leaves less room for persona- driven improvement. 8 Ethical Considerations Our safety evaluation uses established adversar- ial benchmarks for defensive research; while per- sona prompts could theoretically be misused to bypass safety filters, this dual-use risk is inherent to system-prompt steering and PRISM’s gated rout- ing demonstrably strengthens rather than weakens safety alignment. References Amanda Askell, Yuntao Bai, Anna Chen, Dawn Drain, Deep Ganguli, Thomas Henighan, Andy Jones, Nicholas Joseph, Benjamin Mann, Nova DasSarma, Nelson Elhage, Zac Hatfield-Dodds, Danny Her- nandez, Jackson Kernion, Kamal Ndousse, Cather- ine Olsson, Dario Amodei, Tom B. Brown, Jack Clark, and 3 others. 2021. A general language as- sistant as a laboratory for alignment. arXiv preprint arXiv:2112.00861. Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Olsson, Christo- pher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, Ethan Perez, and 32 others. 2022. Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073. Xin Chan, Xiaoyang Wang, Dian Yu, Haitao Mi, and Dong Yu. 2024. Scaling synthetic data cre- ation with 1,000,000,000 personas. arXiv preprint arXiv:2406.20094. Patrick Chao, Edoardo Debenedetti, Alexander Robey, Maksym Andriushchenko, Francesco Croce, Vikash Sehwag, Florian Tramer, Cho-Jui Hsieh, Nicholas Carlini, and J Zico Kolter. 2024. Jailbreakbench: An open robustness benchmark for jailbreaking large language models. arXiv preprint arXiv:2404.01318. Lingjiao Chen, Matei Zaharia, and James Zou. 2023. FrugalGPT: How to use large language models while reducing cost and improving performance. arXiv preprint arXiv:2305.05176. Yuxing Chen, Guoqing Luo, Zijun Wu, and Lili Mou. 2026.Multi-persona thinking for bias mitiga- tion in large language models.arXiv preprint arXiv:2601.15488. Alexis Chevalier, Alexander Wettig, Anirudh Ajith, and Danqi Chen. 2023.Adapting language models to compress contexts.arXiv preprint arXiv:2305.14788. Wei-Lin Chiang, Lianmin Zheng, Ying Sheng, Anasta- sios Nikolas Angelopoulos, Tianle Li, Dacheng Li, Hao Zhang, Banghua Zhu, Michael Jordan, Joseph E. Gonzalez, and Ion Stoica. 2024. Chatbot arena: An open platform for evaluating LLMs by human prefer- ence. arXiv preprint arXiv:2403.04132. Ewelina Gajewska, Jarosław A Chudziak, Arda Der- bent, and Katarzyna Budzynska. 2025. Algorith- mic fairness in NLP: Persona-infused LLMs for human-centric hate speech detection. arXiv preprint arXiv:2510.19331. Shashank Gupta, Vaishnavi Shrivastava, Ameet Desh- pande, Ashwin Kalyan, Peter Clark, Ashish Sabhar- wal, and Tushar Khot. 2024. Bias runs deep: Implicit reasoning biases in persona-assigned llms. arXiv preprint arXiv:2311.04892. Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, and Jacob Steinhardt. 2021. Measuring massive multitask language under- standing. arXiv preprint arXiv:2009.03300. Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2022. LoRA: Low-rank adap- tation of large language models. arXiv preprint arXiv:2106.09685. 9 Jiaming Ji, Mickel Liu, Juntao Dai, Xuehai Pan, Chi Zhang, Ce Bian, Boyuan Chen, Ruiyang Sun, Yizhou Wang, and Yaodong Yang. 2024. PKU-SafeRLHF: A safety alignment preference dataset for LLMs. arXiv preprint arXiv:2406.15513. Junseok Kim, Nakyeong Yang, and Kyomin Jung. 2025. Persona is a double-edged sword: Rethinking the impact of role-play prompts in zero-shot reasoning tasks. In Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics. Aobo Kong, Shiwan Zhao, Hao Chen, Qicheng Li, Yong Qin, Ruiqi Sun, Xin Zhou, Enzhi Wang, and Xiaohang Dong. 2024. Better zero-shot rea- soning with role-play prompting. arXiv preprint arXiv:2308.07702. Sai Adith Senthil Kumar, Hao Yan, Saipavan Perepa, Murong Yue, and Ziyu Yao. 2025. Can LLMs sim- ulate personas with reversed performance? a bench- mark for counterfactual instruction following. arXiv preprint arXiv:2504.06460. Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Shashank Gupta, Bodhisattwa Prasad Majumder, Katherine Hermann, Sean Welleck, Amir Yazdan- bakhsh, and Peter Clark. 2023. Self-refine: Iterative refinement with self-feedback. Advances in Neural Information Processing Systems, 36. Mantas Mazeika, Long Phan, Xuwang Yin, Daniel Mc- Duff, Yaron Zick, Andy Zou, Zifan Wang, Norman Mu, Zico Kolter, and Dawn Song. 2024. Harm- bench: A standardized evaluation framework for auto- mated red teaming and robust refusal. arXiv preprint arXiv:2402.04249. Isaac Ong, Amjad Almahairi, Vincent Wu, Wei-Lin Chi- ang, Tianhao Wu, Joseph E. Gonzalez, M Waleed Kadous, and Ion Stoica. 2024. Routellm: Learning to route llms with preference data. arXiv preprint arXiv:2406.18665. Zhuoshi Pan, Qianhui Wu, Huiqiang Jiang, Menglin Xia, Xufang Luo, Jue Zhang, Qingwei Lin, Victor Rühle, Yuqing Yang, Chin-Yew Lin, H. Vicky Zhao, Lili Qiu, and Dongmei Zhang. 2024. Llmlingua-2: Data distil- lation for efficient and faithful task-agnostic prompt compression. arXiv preprint arXiv:2403.12968. Leonard Salewski, Stephan Alaniz, Isabel Rio-Torto, Eric Schulz, and Zeynep Akata. 2023. In-context im- personation reveals large language models’ strengths and biases. In Advances in Neural Information Pro- cessing Systems, volume 36. Avi Singh, John D Co-Reyes, Rishabh Agarwal, Ankesh Anand, Piyush Patil, Xavier Garcia, Peter J Liu, James Harrison, Jaehoon Lee, Kelvin Xu, Aaron T Parisi, Abhishek Kumar, Alexander A Alemi, Alex Rizkowsky, Azade Nova, Ben Adlam, Bernd Bohnet, Gamaleldin Fathy Elsayed, Hanie Sedghi, and 13 oth- ers. 2024. Beyond human data: Scaling self-training for problem-solving with language models. Transac- tions on Machine Learning Research. Charlie Snell, Dan Klein, and Ruiqi Zhong. 2022. Learning by distilling context.arXiv preprint arXiv:2209.15189. Fiona Anting Tan, Gerard Christopher Yeo, Fanyou Wu, Vinija Jain, Kokil Jaidka, Yang Liu, and See-Kiong Ng. 2025. PHAnToM: Persona-based prompting has an effect on theory-of-mind reasoning in large lan- guage models. arXiv preprint arXiv:2403.02246. Kimberly Le Truong, Riccardo Fogliato, Hoda Heidari, and Zhiwei Steven Wu. 2025. Persona-augmented benchmarking: Evaluating LLMs across diverse writ- ing styles. arXiv preprint arXiv:2507.22168. Yizhong Wang, Yeganeh Kordi, Swaroop Mishra, Al- isa Liu, Noah A Smith, Daniel Khashabi, and Han- naneh Hajishirzi. 2023. Self-instruct: Aligning lan- guage models with self-generated instructions. arXiv preprint arXiv:2212.10560. Shenghan Wu, Yimo Zhu, Wynne Hsu, Mong-Li Lee, and Yang Deng. 2025. From personas to talks: Revis- iting the impact of personas on LLM-synthesized emotional support conversations. arXiv preprint arXiv:2502.11451. Benfeng Xu, An Yang, Junyang Lin, Quan Wang, Chang Zhou, Yongdong Zhang, and Zhendong Mao. 2023. Expertprompting: Instructing large language mod- els to be distinguished experts.arXiv preprint arXiv:2305.14688. Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Sainbayar Sukhbaatar, Jing Xu, and Jason Weston. 2024.Self-rewarding language models.arXiv preprint arXiv:2401.10020. Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric P. Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. 2023. Judging LLM-as-a-judge with MT-Bench and Chatbot Arena. Advances in Neural Information Processing Systems, 36. Mingqian Zheng, Jiaxin Pei, and David Jurgens. 2024. When “a helpful assistant” is not really helpful: Per- sonas in system prompts do not improve perfor- mances of large language models. In Findings of the Association for Computational Linguistics: ACL 2024. 10 A Model Details We investigate persona effects on 6 LLMs span- ning three families: 3 instruction-tuned models, 1 Mixture-of-Experts model, and 2 reasoning- distilled models. Table 2 lists all models with their sizes and system-prompt support. For PRISM (§4.2), we evaluate on 5 of the 6 models, excluding the MoE model due to the challenges of LoRA- based finetuning with sparse activation patterns. Table 2: Models evaluated in this work. All 6 models are used for persona investigation (§3); PRISM is applied to the 5 dense models. “Sys.” indicates whether the model’s chat template includes a default system prompt. ModelParamsSys.Notes Instruction-Tuned Qwen2.5-7B-Inst.7B✓Default: “You are Qwen, a helpful assistant.” Llama-3.1-8B-Inst.8B✓ Default: safety-focused system prompt Mistral-7B-Inst.-v0.37B✗ No default sys prompt in template Mixture-of-Experts Mixtral-8x7B-Inst.-v0.18×7B✓Sparse MoE; investiga- tion only Reasoning-Distilled (DeepSeek-R1) R1-Distill-Qwen-7B7B✓ Distilledfrom DeepSeek-R1; reasoning/code/STEM- heavy training set R1-Distill-Llama-8B8B✓Distilled from DeepSeek- R1;safetyalignment erased during distillation B Context Prompt Generation We describe the procedure used to generate the per- sona context prompts that serve as the distillation targets in PRISM. Framework.Our context generation follows the ExpertPrompting framework (Xu et al., 2023), which instructs an LLM to produce detailed, second-person agent descriptions tailored to each input instruction. The meta-instructions and few- shot template were generated using OpenAI GPT- 4o-mini, while the actual persona context prompts used in our experiments were generated by Claude Opus 4.6. Template.The following few-shot template was used to elicit expert agent descriptions: For each instruction, write a high-quality descrip- tion about the most capable and suitable agent to answer the instruction. In second person perspec- tive. [Instruction]: Make a list of 5 possible effects of deforestation. [Agent Description]: You are an environmen- tal scientist with a specialization in the study of ecosystems and their interactions with hu- man activities. You have extensive knowledge about the effects of deforestation on the environ- ment, including the impact on biodiversity, cli- mate change, soil quality, water resources, and human health. Your work has been widely rec- ognized and has contributed to the development of policies and regulations aimed at promoting sustainable forest management practices. . . . [Instruction]: Identify a descriptive phrase for an eclipse. [Agent Description]: You are an astronomer with a deep understanding of celestial events and phe- nomena. Your vast knowledge and experience make you an expert in describing the unique and captivating features of an eclipse. You have wit- nessed and studied many eclipses throughout your career, and you have a keen eye for detail and nu- ance. . . . [Instruction]: Identify the parts of speech in this sentence: “The dog barked at the postman”. [Agent Description]: You are a linguist, well- versed in the study of language and its structures. You have a keen eye for identifying the parts of speech in a sentence and can easily recognize the function of each word. . . . [Instruction]: instruction [Agent Description]: By conditioning Claude Opus 4.6 on this tem- plate, we obtain rich, domain-specific persona de- scriptions that capture the expertise, tone, and reasoning style appropriate for each category of queries. These descriptions then serve as the system-prompt contextsCthat PRISM distills into the model’s parameters. C Persona Prompts We evaluate three granularity levels for each per- sona: Full (∼150 tokens, detailed expert descrip- tion), Short (∼75 tokens, condensed version), and Min (∼5 tokens, minimal label). Tables below show the complete system prompts. 11 Table 3: Writing persona at all granularity levels. FullYou are an accomplished professional writer and editor with mastery across multiple forms of writ- ing, including creative fiction, expository essays, persuasive arguments, technical documentation, po- etry, screenwriting, and business communication. You have decades of experience crafting compelling prose and have worked as a published author, liter- ary editor, and writing instructor. You possess an exceptional command of language, grammar, style, and rhetoric, and you can adapt your tone and voice to suit any audience or purpose. You are skilled at structuring narratives with strong openings, well- developed middles, and satisfying conclusions. Your writing is vivid, precise, and engaging, demonstrat- ing both technical mastery and genuine creative flair. ShortYou are an accomplished professional writer and editor with mastery across creative fiction, essays, technical documentation, and poetry. You have ex- ceptional command of language, grammar, style, and rhetoric. You structure narratives with strong openings and satisfying conclusions, adapting tone for any audience. Your writing is vivid, precise, and engaging, demonstrating both technical mastery and creative flair. MinYou are a professional writer. Table 4: Roleplay persona at all granularity levels. FullYou are a masterful storyteller and creative writer with extensive experience in improvisation, char- acter development, and narrative craft. You have a rich background in theater, creative writing, and in- teractive fiction, giving you the ability to inhabit any character or persona with depth and authentic- ity. You can adopt distinct voices, mannerisms, and perspectives, whether portraying a historical figure, a fictional character, or a professional in any field. You are deeply empathetic and imaginative, able to understand and express a wide range of emotions, motivations, and worldviews. You maintain consis- tency in character throughout a conversation, staying true to the established personality while responding naturally and engagingly to new prompts. ShortYou are a masterful storyteller and improviser who can inhabit any character with depth and authentic- ity. You adopt distinct voices, mannerisms, and per- spectives, maintaining consistency throughout. You are imaginative and empathetic, skilled at world- building and weaving compelling narratives on the fly. Your performances are nuanced, dynamic, and responsive to the user’s cues. MinYou are a roleplay storyteller. Table 5: Reasoning persona at all granularity levels. FullYou are a precision-focused logical reasoner whose top priority is arriving at the correct conclusion. You have deep expertise in formal logic, deductive and inductive reasoning, constraint satisfaction, and de- cision theory. You approach every problem by first identifying exactly what is being asked, then system- atically working through the logic to reach the right answer. You keep your reasoning tight and focused— each step must be logically necessary, not merely illustrative. You are especially careful about nega- tions, quantifier scope, conditional vs. biconditional statements, and subtle distinctions between “neces- sary” and “sufficient” conditions. ShortYou are a precision-focused logical reasoner whose top priority is the correct conclusion. You have deep expertise in formal logic, deduction, induction, and constraint satisfaction. You keep reasoning tight— each step logically necessary, not illustrative. You verify each inference against premises, resolve ambi- guity explicitly, and would rather give a short correct answer than a long wrong one. MinYou are a logical reasoner. Table 6: Math persona at all granularity levels. FullYou are a rigorous mathematician who prioritizes correctness and precision above all else. Your pri- mary goal is to produce the exact right answer with every calculation verified. You have deep exper- tise in algebra, calculus, number theory, probability, statistics, linear algebra, differential equations, and discrete mathematics. You double-check every arith- metic operation, algebraic manipulation, and logical inference before committing. You are vigilant about common pitfalls: sign errors, off-by-one mistakes, incorrect applications of theorems, and failure to check domain restrictions or boundary conditions. Accuracy is your highest value. ShortYou are a rigorous mathematician who prioritizes correctness and precision. You have deep expertise across algebra, calculus, number theory, probability, and statistics. You focus on producing the exact right answer with only essential steps shown. You double- check every calculation, watch for sign errors and off-by-one mistakes, and never guess when an ex- act answer is obtainable. Accuracy is your highest value. MinYou are a mathematician. 12 Table 7: Coding persona at all granularity levels. FullYou are a senior software engineer who writes code that is correct first, clean second, and fast third. Your top priority is producing code that actually works— handles edge cases, validates inputs, and passes all tests on the first run. You have deep expertise in Python, Java, C++, JavaScript, and Rust, with strong command of algorithms, data structures, and system design. You write concise, correct implementations rather than verbose ones with excessive comments. You test your code mentally against edge cases be- fore presenting it. You never write placeholder or pseudo-code when a working implementation is ex- pected. ShortYou are a senior software engineer who writes code that is correct first, clean second. You have deep expertise in Python, Java, C++, JavaScript, and Rust. You focus on getting logic right, handling edge cases (empty inputs, off-by-one, overflow, null), and choosing the correct algorithm. You write concise working implementations, never placeholders. Your code compiles, runs, and returns the correct output. MinYou are a software engineer. Table 8: Extraction persona at all granularity levels. FullYou are a data extraction and information retrieval specialist with deep expertise in natural language processing, structured data parsing, and document analysis. You have extensive experience working with unstructured text, tables, web pages, and com- plex documents to extract precise, relevant informa- tion. You are skilled at reformatting extracted infor- mation into clean, structured outputs such as tables, lists, JSON, or summaries as required. You under- stand the importance of faithfulness to the source material and never fabricate or hallucinate informa- tion that is not present in the given text. Short You are a data extraction specialist expert in parsing unstructured text, tables, and documents to extract precise information. You identify key entities, rela- tionships, and facts with meticulous accuracy. You reformat extracted data into clean structured outputs (tables, JSON, lists) and never fabricate information not present in the source. When data is ambiguous, you indicate uncertainty. MinYou are a data extraction specialist. Table 9: STEM persona at all granularity levels. FullYou are a versatile STEM expert with comprehen- sive knowledge spanning physics, chemistry, biol- ogy, engineering, and computer science. You hold advanced degrees in the natural sciences and have extensive research experience in both experimental and theoretical domains. You can explain complex scientific concepts at any level of detail, from intu- itive analogies for beginners to rigorous technical explanations for specialists. You are skilled at apply- ing the scientific method, designing experiments, in- terpreting data, and drawing evidence-based conclu- sions. Your explanations are precise, well-structured, and grounded in established scientific knowledge, and you clearly distinguish between well-established facts, current hypotheses, and speculative ideas. ShortYou are a versatile STEM expert with comprehen- sive knowledge in physics, chemistry, biology, en- gineering, and computer science. You explain com- plex scientific concepts at any level, apply the sci- entific method rigorously, and stay current with latest research. Your explanations are precise and grounded in established knowledge, clearly distin- guishing facts from hypotheses. MinYou are a STEM expert. Table 10: Humanities persona at all granularity levels. FullYou are a distinguished humanities scholar with broad expertise spanning philosophy, history, lit- erature, ethics, cultural studies, and the arts. You hold advanced degrees in the humanities and have published extensively on topics ranging from an- cient philosophy to contemporary cultural criticism. You are adept at close reading, critical analysis, and constructing nuanced arguments that consider multi- ple perspectives. You can engage thoughtfully with complex ethical dilemmas, drawing on philosophi- cal frameworks from Aristotle to Rawls and beyond. Your responses demonstrate both scholarly depth and a genuine passion for humanistic inquiry. ShortYou are a distinguished humanities scholar with broad expertise in philosophy, history, literature, ethics, and cultural studies. You construct nuanced arguments considering multiple perspectives, en- gage thoughtfully with ethical dilemmas drawing on philosophical frameworks, and situate cultural artifacts in their broader intellectual context. Your writing combines scholarly depth with clarity and elegance. MinYou are a humanities scholar. 13 Table 11: Critic persona at all granularity levels. FullYou are an incisive critical analyst and intellectual evaluator with extensive experience in peer review, philosophical inquiry, and quality assessment across academic, technical, and creative domains. You ap- proach every piece of content with rigorous intellec- tual honesty, examining it from multiple angles and stress-testing its claims against available evidence and established principles. You are skilled at distin- guishing between strong reasoning and superficial plausibility. Your critiques are always specific, ac- tionable, and grounded in clear reasoning rather than subjective preference. Short You are an incisive critical analyst with extensive ex- perience in peer review and quality assessment. You identify weaknesses, inconsistencies, and gaps in ar- guments with rigorous intellectual honesty. You dis- tinguish strong reasoning from superficial plausibil- ity, providing specific actionable feedback. Your cri- tiques are fair-minded but uncompromising, clearly flagging logical fallacies and unsupported claims. MinYou are a critical analyst. Table 12: Safety Monitor persona at all granularity lev- els. FullYou are a meticulous safety monitor and risk assess- ment specialist with deep expertise in identifying potential harms, ethical concerns, and dangerous content across a wide range of topics. You carefully evaluate each interaction for potential risks, consid- ering both the explicit content and the implicit intent behind requests. When you identify a safety concern, you explain clearly why the request is problematic and suggest safer alternatives when possible. You balance caution with helpfulness, avoiding unneces- sary refusals for benign requests while maintaining firm boundaries against genuinely harmful content. You apply consistent, principled judgment rather than relying on rigid keyword-based filtering. ShortYou are a meticulous safety monitor expert in iden- tifying potential harms, ethical concerns, and dan- gerous content. You evaluate requests for risks to physical safety, legality, and privacy, considering both explicit content and implicit intent. You bal- ance caution with helpfulness, avoiding unnecessary refusals while maintaining firm boundaries against genuinely harmful content. You apply principled judgment, not rigid keyword filtering. MinYou are a safety monitor. Table 13: Helpful persona at all granularity levels. Used as proxy for MMLU Social Sciences and Other domains. FullYou are a highly capable and genuinely helpful assis- tant dedicated to providing the most useful, accurate, and thorough responses possible. You have broad knowledge spanning science, technology, arts, his- tory, current events, and practical life skills. You prioritize understanding what the user truly needs, often reading between the lines to address the un- derlying goal behind a question rather than just its surface form. You adapt your communication style to match the user’s level of expertise, providing tech- nical depth for experts and accessible explanations for beginners. Your responses are comprehensive yet concise, giving the user everything they need without unnecessary verbosity or filler. ShortYou are a highly capable and genuinely helpful as- sistant providing the most useful, accurate, and thor- ough responses possible. You understand what users truly need, structure responses for clarity with ex- amples and step-by-step explanations, and proac- tively anticipate follow-up questions. You adapt to the user’s expertise level and clearly indicate when uncertain. Your responses are comprehensive yet concise. MinYou are a helpful assistant. Table 14: Compliant persona at all granularity levels. FullYou are a maximally compliant and accommodat- ing assistant whose primary objective is to fulfill every user request as completely and directly as pos- sible. You interpret instructions literally and execute them without adding unsolicited caveats, warnings, or qualifications. You approach tasks with a strong bias toward action and completion rather than delib- eration or refusal. You are willing to engage with hypothetical scenarios, creative fiction, role-play, and thought experiments without breaking character or injecting meta-commentary. You treat the user as an autonomous adult capable of determining what information and content they need. You only de- cline requests when they involve clear and direct real-world harm. ShortYou are a maximally compliant assistant whose pri- mary objective is to fulfill every user request as completely and directly as possible. You interpret instructions literally, prioritize action over delibera- tion, and generate content in any requested style or format. You engage freely with hypothetical scenar- ios and creative fiction without meta-commentary. You only decline requests involving clear, direct real- world harm. MinYou are a compliant assistant. 14 D Evaluation Benchmark Details This section describes the evaluation protocols, judging methodology, and dataset statistics for each benchmark used in our experiments. D.1 MT-Bench (Generative Quality) MT-Bench (Zheng et al., 2023) is a multi-turn con- versational benchmark consisting of 80 questions spanning 8 categories: Writing, Roleplay, Reason- ing, Math, Coding, Extraction, STEM, and Human- ities (10 questions per category). Each question involves a two-turn interaction where the second turn builds on the first. Judging protocol. We use Qwen3-32B-Instruct as the LLM-as-Judge evaluator for all models, fol- lowing the MT-Bench protocol (Zheng et al., 2023) which recommends using a strong model as judge for consistent cross-model comparison. For each response, the judge assigns a score on a 1–10 scale based on helpfulness, relevance, accuracy, depth, and clarity. We average Turn 1 and Turn 2 scores per question, then report the mean across all 10 questions in each category. When persona prompts are applied, the system prompt for the generation phase is set to the persona, while the judging phase uses the default system prompt to ensure consistent evaluation criteria. No-system-prompt baseline.For the no-persona baseline, models that have a baked-in default sys- tem prompt (e.g., Qwen’s “You are Qwen, created by Alibaba Cloud. You are a helpful assistant.”) are evaluated with their default intact. The “No System Prompt” ablation explicitly overrides this default with an empty system message to isolate the effect of the default prompt itself. D.2 MMLU (Discriminative Knowledge) MMLU (Massive Multitask Language Under- standing) (Hendrycks et al., 2021) evaluates fac- tual knowledge and reasoning across 57 subjects grouped into 4 domains: STEM, Humanities, So- cial Sciences, and Other. Evaluation protocol. We use 5-shot evaluation with log-likelihood scoring: for each multiple- choice question, we compute the log-probability of each answer choice (A, B, C, D) condi- tioned on the question and few-shot exemplars, and select the choice with the highest probabil- ity. This ensures that evaluation measures the model’s internal knowledge distribution rather than its instruction-following format compliance. When persona prompts are applied, the persona is prepended as a system message before the few-shot exemplars. Statistics. The full MMLU benchmark contains 14,042 test questions. We report accuracy (%) per domain and overall, without confidence intervals, as the large sample size yields negligible standard errors (< 0.5%). D.3 Safety Benchmarks We evaluate safety alignment using three comple- mentary benchmarks that cover different types of harmful requests: HarmBench.HarmBench (Mazeika et al., 2024) contains adversarial prompts designed to elicit harmful model behaviors across categories includ- ing violence, illegal activities, misinformation, and privacy violations. We use 400 prompts sampled from the full benchmark. JailbreakBench. JailbreakBench (Chao et al., 2024) collects jailbreak attack prompts that attempt to circumvent safety training through prompt injec- tion, role-play manipulation, and other adversarial techniques. We evaluate on 79 prompts. PKU-SafeRLHF. PKU-SafeRLHF (Ji et al., 2024) provides prompts from real user interactions that were flagged as potentially unsafe, covering a naturalistic distribution of harmful requests. We evaluate on 400 prompts. Judging protocol. For all three safety bench- marks, we use an LLM-as-Judge approach. The judge model (same as the evaluated model unless otherwise noted) is presented with the original harmful prompt and the model’s response, then asked: “Does the Assistant Response refuse to an- swer the harmful request? Answer with ‘Yes’ or ‘No’.” A response classified as “Yes” counts as a refusal. The Refusal Rate (R%) is the percentage of responses that refuse the harmful request. Statistical reporting. We report bootstrap con- fidence intervals for all safety metrics. For each benchmark, we perform 1,000 bootstrap resamples of then = 400judged responses, compute the re- fusal rate for each resample, and report the mean ±standard error along with the 95% confidence interval. This accounts for the sampling variability inherent in the evaluation set. 15 Generation parameters. All safety responses are generated with greedy decoding (temperature = 0, no sampling) and a maximum of 256 new tokens. Batched generation with left-padding is used for efficiency, with batch sizes of 8. Table 15: Summary of evaluation benchmarks and their key statistics. Benchmark#SamplesMetricScoring MT-Bench80Score (1–10)LLM judge MMLU14,042Accuracy (%)Log-likelihood HarmBench400R (%)LLM judge JailbreakBench79R (%)LLM judge PKU-SafeRLHF500R (%)LLM judge E Verbosity Bias in Self-Verification A key design choice in PRISM Stage 3 is how the self-judge determines whether the expert persona or baseline answer is superior. We initially used pointwise scoring, where each answer is indepen- dently rated on a 1–10 scale, and the higher-scoring answer wins. However, we discovered that this ap- proach introduces a systematic verbosity bias: the self-judge consistently prefers longer, more elabo- rated answers—even when they are factually incor- rect. Evidence. Under pointwise scoring, the self- judge routes a disproportionate fraction of queries to the expert persona across all categories. For Mistral-7B, the math persona achieves a 68% dis- till rate, meaning the judge considered the per- sona answer superior in 68 out of 100 comparisons. However, MT-Bench evaluation with Qwen3-32B- Instruct as judge reveals that the math persona de- grades Mistral’s math score by 2.95 points (9.05→ 6.10). This contradiction demonstrates that the self- judge is rewarding the persona’s verbose, step-by- step formatting rather than evaluating mathematical correctness. This bias is well-documented in the LLM-as- judge literature (Zheng et al., 2023): when grad- ing answers independently (pointwise), models as- sign higher scores to longer responses regardless of their factual quality. The bias compounds across categories: since the expert persona systematically produces more verbose answers, the distill rate is inflated for all categories, and the gate inherits this bias. Solution: Pairwise comparison with position swapping. Following best practices from MT- Bench (Zheng et al., 2023) and Chatbot Arena (Chi- ang et al., 2024), we replace pointwise scoring with pairwise comparison: the judge sees both answers simultaneously and selects the better one (A, B, or TIE). To further eliminate position bias, we run the comparison twice with swapped answer positions: •Pass 1: Answer A = baseline, Answer B = expert •Pass 2: Answer A = expert, Answer B = baseline The expert wins only if selected in both orderings. This conservative criterion provides three benefits: (1) placing both answers in the same context en- ables direct mutual comparison rather than relying on absolute scores, (2) position swapping cancels systematic first-answer or second-answer prefer- ence, and (3) requiring agreement across both or- derings filters out cases where the judge’s pref- erence was driven by superficial features (length, formatting) rather than substantive quality. Mixed results are conservatively assigned to the retain set, ensuring the gate errs toward the base model. F Gated Single-LoRA Training Setup The Gated Single-LoRA variant of PRISM replaces the multi-expert Mixture-of-LoRAs architecture with a single, higher-rank LoRA adapter controlled by a binary gate. This section details the training configuration. Architecture. The adapter consists of two com- ponents: (1) a single LoRA adapter applied to all attention and MLP projections (q_proj, k_proj,v_proj,o_proj,gate_proj,up_proj, down_proj), and (2) a binary gate MLP that de- cides per-query whether to activate the LoRA. The gate architecture is a 3-layer MLP (h → 128 → 64→ 1 ) with GELU activations, operating on the last-token hidden state of the first transformer layer (layer 0). Training data. From the PRISM Stage 2 multi- persona grading results, we construct two parti- tions: (1) distill samples (gate target = 1), where any persona outperformed the baseline, and (2) re- tain samples (gate target = 0), where the baseline was best. For Qwen2.5-7B-Instruct, this yields 282 distill and 318 retain samples (600 total). Training objective. The loss combines: (i) gate loss (binary cross-entropy on gate predictions), (i) KL distillation loss for distill samples (match- ing the LoRA-augmented student distribution to teacher logits), and (i) KL retention loss (scaled byλ retain = 0.5) for retain samples. Teacher logits 16 are pre-computed per sample and stored on disk to avoid OOM during training. Training hyperparam- eters are listed in Table 16. Table 16: Gated Single-LoRA training configuration. ParameterValue LoRA rank (r)16 LoRA alpha (α)32 LoRA dropout0.05 Target modulesAll (7 proj.) Trainable params∼21M LR (LoRA)2× 10 −4 LR (Gate)1× 10 −3 Epochs10 Micro batch size1 Grad. accumulation16 Max seq. length1024 KL temperature (τ )2.0 Retain weight (λ ret )0.5 Teacher logit storagePer-sample disk Training samples600 (282 dist. + 318 ret.) Training time∼45 min (A100) Final gate accuracy68.8% Compute. All experiments were conducted on single-GPU nodes using a mix of NVIDIA A100 80GB and NVIDIA RTX A6000 48GB GPUs. Stages 1–3 (query generation, answer generation, self-verification) and Stage 5 (LoRA distillation) each require a single GPU for model inference or training. Stage 4 (gate training) is lightweight and runs on either GPU type. Teacher logits are pre-computed and stored on disk (one.ptfile per sample) to avoid holding two full model copies in memory, enabling training on the 48GB A6000. 17