Paper deep dive
SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection
Kexian Tang, Jiani Wang, Shaowen Wang, Kaifeng Lyu
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 96%
Last extracted: 3/26/2026, 2:48:05 AM
Summary
The paper introduces SPA (Scaling Prompt-engineered Augmentation), a simple yet highly effective baseline for knowledge injection in large language models. SPA utilizes a set of seven human-curated prompt templates based on cognitive science and educational psychology (Concept Learning, Critical Thinking, and Generative Learning) to generate large-scale synthetic data from small, domain-specific corpora. The authors demonstrate that SPA consistently outperforms complex existing methods like SEAL, EntiGraph, and Active Reading, particularly as the synthetic corpus scales, while also identifying limitations in prior RL-based and multi-stage prompting approaches.
Entities (5)
Relation Signals (3)
SPA â uses â Prompt templates
confidence 100% · SPA rewrites the original corpus into a large synthetic corpus by repeatedly prompting a language model G with a fixed set of human-curated prompt templates
SPA â outperforms â SEAL
confidence 95% · We observe that SPA consistently matches or outperforms a range of substantially more complex methods, including SEAL
SPA â outperforms â EntiGraph
confidence 95% · We observe that SPA consistently matches or outperforms a range of substantially more complex methods, including... EntiGraph
Cypher Suggestions (2)
Find all methods compared against SPA in the paper. · confidence 90% · unvalidated
MATCH (m:Method {name: 'SPA'})-[:OUTPERFORMS]->(baseline:Method) RETURN baseline.nameMap the relationship between methods and their underlying techniques. · confidence 85% · unvalidated
MATCH (m:Method)-[:USES]->(t:Technique) RETURN m.name, t.name
Abstract
Abstract:While large language models (LLMs) are pretrained on massive amounts of data, their knowledge coverage remains incomplete in specialized, data-scarce domains, motivating extensive efforts to study synthetic data generation for knowledge injection. We propose SPA (Scaling Prompt-engineered Augmentation), a simple but tough-to-beat baseline that uses a small set of carefully designed prompts to generate large-scale synthetic data for knowledge injection. Through systematic comparisons, we find that SPA outperforms several strong baselines. Furthermore, we identify two key limitations of prior approaches: (1) while RL-based methods may improve the token efficiency of LLM-based data augmentation at small scale, they suffer from diversity collapse as data scales, leading to diminishing returns; and (2) while multi-stage prompting may outperform simple augmentation methods, their advantages can disappear after careful prompt tuning. Our results suggest that, for knowledge injection, careful prompt design combined with straightforward large-scale augmentation can be surprisingly effective, and we hope SPA can serve as a strong baseline for future studies in this area. Our code is available at this https URL.
Tags
Links
- Source: https://arxiv.org/abs/2603.22213v1
- Canonical: https://arxiv.org/abs/2603.22213v1
Full Text
108,044 characters extracted from source content.
Expand or collapse full text
SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Kexian Tang * 1 Jiani Wang * 1 Shaowen Wang 1 Kaifeng Lyu â 1 Abstract While large language models (LLMs) are pre- trained on massive amounts of data, their knowl- edge coverage remains incomplete in specialized, data-scarce domains, motivating extensive efforts to study synthetic data generation for knowledge injection. We propose SPA (Scaling Prompt- engineered Augmentation), a simple but tough-to- beat baseline that uses a small set of carefully de- signed prompts to generate large-scale synthetic data for knowledge injection. Through system- atic comparisons, we find that SPA outperforms several strong baselines. Furthermore, we iden- tify two key limitations of prior approaches: (1) while RL-based methods may improve the token efficiency of LLM-based data augmentation at small scale, they suffer from diversity collapse as data scales, leading to diminishing returns; and (2) while multi-stage prompting may outperform simple augmentation methods, their advantages can disappear after careful prompt tuning. Our results suggest that, for knowledge injection, care- ful prompt design combined with straightforward large-scale augmentation can be surprisingly ef- fective, and we hope SPA can serve as a strong baseline for future studies in this area. Our code is available at https://github.com/Tangkexian/SPA. 1. Introduction Large language models (LLMs) can acquire broad world knowledge and general capabilities from massive web text, but their coverage of specialized domains is often incom- plete. A common approach to address this issue is to inject domain knowledge into these models by further fine-tuning or continually pretraining them on domain-specific data that are not readily available on the public web (Singhal et al., 2023; Wu et al., 2023; Zhang et al., 2024). However, domain-specific datasets are usually limited in * Equal contribution 1 Institute for Interdisciplinary Information Sciences, Tsinghua University. Correspondence to: Kaifeng Lyu <klyu@mail.tsinghua.edu.cn>. Preprint. March 24, 2026. both scale and diversity. In contrast to the high redundancy of web text, domain knowledge often lacks the repetitive variations required for effective learning. Compounding this challenge, LLMs are known to be data-inefficient, necessi- tating frequent and varied exposure to facts before they can be reliably internalized (Kandpal et al., 2023; Allen-Zhu & Li, 2025; Abonizio et al., 2025). Consequently, directly fine-tuning LLMs on such sparse data often leads to overfit- ting on specific surface forms rather than robust knowledge acquisition, causing the model to fail when encountering un- seen queries or variations (Berglund et al., 2024; Shumailov et al., 2023). To mitigate this scarcity, an increasingly popular paradigm is to expand the domain-specific data into a much larger syn- thetic corpus using an LLM-based data generator, and then train the target model on this augmented dataset. Recent ef- forts to design such augmentation methods generally follow two distinct methodologies. RL-based approaches, such as SEAL (Zweiger et al., 2025), employ reinforcement learn- ing (RL) to train a dedicated generator, where the reward is defined as the performance of the target model trained on the augmented data. Another line of works, such as Enti- Graph (Yang et al., 2025) and Active Reading (Oguz et al., 2026), relies on prompting pipelines consisting of multiple stages, where the data generator is prompted to rewrite the original corpus into a high-quality synthetic corpus through a sequence of intermediate transformations. In this work, we aim to push the frontier of synthetic data generation for knowledge injection by proposing a simple but tough-to-beat baseline method: SPA (Scaling Prompt- engineered Augmentation), which is a knowledge injection method that consistently outperforms existing methods at scale (Figure 1). The key idea of SPA is very simple: 1.Prompt Engineering: We draw upon insights from cognitive science and educational psychology to design a set of 7 prompt templates based on effective human learning strategies, including concept learning, critical thinking, and generative learning. 2.Scaling: We repeatedly prompt an LLM to rewrite the source content using templates from our prompt set, progressively scaling the augmented corpus into a large-scale synthetic corpus. 1 arXiv:2603.22213v1 [cs.LG] 23 Mar 2026 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Title:Southern California Context: Southern California is also home to the Port of Los Angeles, the United Statesâ busiest commercial port; the adjacent Port of Long Beach, the United Statesâ second busiest container port; and the Port of San Diego... Generator with 7 Prompts of SPA Teacher-style Today we are diving into an intriguing text on Southern California. Itâs essential to understand Southern California as it is a... Discussions CaseStudies QA-ct Implications Mindmap Keyconcepts Teacher-styleCaseStudiesDiscussions ImplicationsQA-ct Interacting with âothersâ KeyconceptsMindmap FoundationThinking Input original small corpus Output diverse large synthetic corpus ... Figure 1. Overview of Scaling Prompt-engineered Augmentation (SPA). Our baseline method rewrites a small source corpus into a large synthetic corpus by repeatedly prompting a generator with a fixed set of seven human-curated prompt templates, which are designed based on three levels of learning strategies: Concept Learning, Critical Thinking, and Generative Learning. 3.Training: Finally, we train the target model on the synthetic corpus using the same experimental settings as in prior work. Perhaps surprisingly, we find that this simple baseline is already highly effective. We evaluate SPA on three repre- sentative benchmarks that span Wikipedia-based question answering (SQuAD, Rajpurkar et al. 2016), long-document comprehension (QuALITY, Pang et al. 2022), and multi-hop reasoning (MultiHop-RAG, Tang & Yang 2024). Across all these settings, we use almost the same experimental setup as in the previous works (Zweiger et al., 2025; Yang et al., 2025), and change only the data augmentation strategy. We observe that SPA consistently matches or outperforms a range of substantially more complex methods, including SEAL (Zweiger et al., 2025), PaST (Tang et al., 2026), Enti- Graph (Yang et al., 2025), SoG (Ma et al., 2025), and Active Reading (Oguz et al., 2026), and the gains of SPA become more pronounced as we scale the synthetic corpus. Our contributions are: âą We propose Scaling Prompt-engineered Augmentation (SPA), a simple knowledge injection baseline that uses a small set of prompt templates to generate large-scale synthetic corpora (Section 3). âąWe conduct a series of strictly controlled, token- matched evaluations comparing SPA against many complex methods (Section 4), and demonstrate that SPA consistently matches or outperforms these meth- ods at scale (Section 5). This suggests that SPA is a competitive baseline for knowledge injection despite its simplicity. âąOur results provide insights into the limitations of ex- isting methods: (1) RL-based methods suffer from di- versity collapse as data scales, leading to diminishing returns; and (2) adding more stages in the prompting pipeline may not yield larger gains than improving the quality of the prompt set for one-stage methods. While SPA demonstrates strong effectiveness, we do not suggest that knowledge injection should just rely on SPA. Rather, we hope our results provide a solid foundation for future work to develop methods that genuinely improve upon competitive baselines. 2. Related Work Small Corpus Knowledge Injection. Knowledge injec- tion for domain-specific data has a wide range of applica- tions. For example, in domains such as medicine (Peng et al., 2019; Lee et al., 2019; Luo et al., 2022), finance (Araci, 2019; Liu et al.; Li et al., 2023), and law (Chalkidis et al., 2020; Guha et al., 2023), it is often necessary to inject spe- cialized knowledge into large language models to support professional tasks. However, domain-specific corpora are typically small and suffer from data scarcity. As a result, directly performing continued pretraining or fine-tuning on such corpora often exhibits limited effectiveness and poor scalability (Gu et al., 2025; Kandpal et al., 2023; Allen-Zhu & Li, 2025; Petroni et al., 2019; Sun et al., 2024). To address this issue, our method SPA aims to scale up small corpora by synthetic data generation based on the original data, fol- lowing recent advances in this line of work (Zweiger et al., 2025; Ovadia et al., 2024; Tang et al., 2026; Mecklenburg et al., 2024; Yang et al., 2025; Oguz et al., 2026; Ma et al., 2025). Synthetic Data Generation. Prior work has explored various strategies for synthetic generation of pretraining data, including (1) prompting LLMs to rewrite documents in diverse styles such as childrenâs stories (Eldan & Li, 2023), âWikipedia-styleâ data, and âquestionâanswerâ for- mats (Maini et al., 2024), and (2) directly prompting LLMs to generate synthetic corpora such as textbook-style data (Gunasekar et al., 2023). However, different from syn- thesizing data from large pretraining corpora, generating data from small corpora is more challenging. Such corpora 2 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection typically lack the scale and redundancy needed for effective learning, requiring higher diversity and broader coverage in the generated data. A more detailed discussion of this line of research is provided in Section 3.2. Knowledge Editing. A related line of work focuses on localizing factual knowledge within LLMs and then updat- ing the model (Meng et al., 2022; Mitchell et al., 2022; Meng et al., 2023) or maintaining an external module to edit specific facts with minimal side effects (Zhong et al., 2023; Cohen et al., 2024). This line of research, known as knowledge editing, differs from our setting. In contrast, we consider corpus-level knowledge injection, where the model is required to learn from a corpus of documents and utilize many interrelated facts expressed in diverse forms, rather than isolated fact tuples. Retrieval-Augmented Generation.Retrieval-augmented generation (RAG) is a widely adopted alternative to para- metric updating, as it enables models to retrieve relevant in- formation from external knowledge bases before answering questions (Lewis et al., 2020; Borgeaud et al., 2022; Izacard et al., 2023; Gao et al., 2023; Asai et al., 2024). In our setting, we do not aim to position our approach as a replace- ment for RAG, but rather as a complementary paradigm to knowledge accessed through retrieval-augmented genera- tion (Yang et al., 2025). 3. SPA: Scaling Prompt-engineered Augmentation We focus on the problem of injecting a small domain- specific text corpus into a pretrained language model. Our goal is to enable the model to internalize this knowledge in its parameters, so that it can correctly answer domain- relevant questions without access to the original corpus at inference time. Following EntiGraph (Yang et al., 2025), we consider synthetic data generation as an effective method for knowledge injection in data-scarce regimes: the original niche corpus is first expanded into a much larger synthetic training set, which is then used for continued pretraining. 3.1. Problem Setup LetD = d i N i=1 denote a small domain-specific corpus, where eachd i corresponds to a single document. Our ob- jective is to design a data augmentation algorithmAthat transformsDinto a larger synthetic corpus Ì D = A(D), which is subsequently used for more effective knowledge injection through continued pretraining. In this work, we focus on task-agnostic augmentation strategies that operate directly on the original documents and do not require ac- cess to downstream questions or labels. This design makes our approach broadly applicable and transferable to other knowledge-injection scenarios. 3.2. Existing Methods Recent work on synthetic data generation for knowledge injection can be divided into two categories: RL-based aug- mentation methods and multi-stage prompting methods. RL-based augmentation methods.This line of research employs reinforcement learning to train a dedicated data generator that produces synthetic data conditioned on the original corpus. SEAL (Zweiger et al., 2025) consists of three main stages: (i) Data generation and training: the data generator gener- ates synthetic data conditioned on the source passages, then the target model is fine-tuned on this data; (i) Reward sig- nal estimation: the downstream QA accuracy of the target model is used as a reward signal to indicate which types of data are more effective; (i) Iterative improvement: the data generator is updated based on the reward signal, and the process is repeated iteratively to improve the quality of the synthetic data. In SEALâs implementation, the data gen- erator shares the same base model as the target model. This iterative procedure encourages the generator to produce data that is easier for the target model to absorb, enabling SEAL to achieve strong performance in low-data settings, even out- performing data generated by more powerful models such as GPT-4.1. While RL can be effective, the computational burden of training makes it impractical to apply RL to every new domain. To mitigate this issue, PaST (Tang et al., 2026) (i) extracts a skill vector for knowledge manipulation from an RL-trained model in a small scale corpus, and (i) linearly injects this vector into the generation model to enhance its ability in knowledge manipulation. However, these methods typically rely on downstream-task performance as the reinforcement signal. Such supervision is often unavailable in realistic settings, where no specific downstream task may be defined at the time of knowledge injection. Our method SPA focuses on task-agnostic aug- mentation strategies which do not require access to down- stream tasks. Multi-stage prompting methods. This line of research introduces multi-stage augmentation pipelines, where the transformation from the original corpus to the final synthetic data proceeds through several intermediate steps. EntiGraph (Yang et al., 2025) first breaks down the origi- nal corpus into a list of entities, and then uses a language model to generate text descriptions about relations among the extracted entities. By rearranging the knowledge in this way, EntiGraph expands a data-scarce corpus and injects it into a 8B model, leading to strong downstream performance that exceeds GPT-4âs performance. SoG (Ma et al., 2025) and GraphGen (Chen et al., 2025) also follow graph-based designs, constructing a graph over corpus chunks and lever- aging this structure to guide data generation. Active Read- 3 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection ing (Oguz et al., 2026) also follows a multi-stage design: it first prompts the generator to produce document-specific learning strategies, such as paraphrasing, knowledge link- ing, and analogical reasoning; then it uses these strategies together with the original corpus to generate augmented data. This approach yields improved accuracy for 8B-scale models on a Wikipedia-grounded subset of SimpleQA, even exceeding the performance of models that are provided with the full context at test time. However, such pipelines can increase system complexity and may require careful optimization at every stage. For example, as analyzed in Section 5.2, the first-stage strategy generation in Active Reading appears less effective than SPA, which could lead to its weaker overall performance. 3.3. Our Method: SPA In this paper, we propose the method of Scaling Prompt- engineered Augmentation (SPA) to serve as a simple but tough-to-beat baseline for knowledge injection. Our Method. Our method SPA rewrites the original cor- pus into a large synthetic corpus by repeatedly prompting a language modelGwith a fixed set of human-curated prompt templatesP i M i=1 , whereM = 7. More specifically, given a target number of tokensD, we generate data independently with each prompt templateP i by repeatedly prompting the language modelG. Each promptP i is used to generate the same amount of tokens, which is approximately of size D/M. Then the synthetic data generated by allMprompts are combined to form the final synthetic corpus of sizeD. We argue that applying this baseline method at scale (i.e., for largeD) produces synthetic data that is effective for knowl- edge injection and is able to outperform existing methods mentioned in Section 3.2. Prompt Engineering. A key component of our baseline method is the design of the prompt setP i M i=1 . For this, we draw upon insights from cognitive science and educational psychology to identify three levels of learning strategies that are effective for human learning: Concept Learning, Critical Thinking, and Generative Learning. Level 1: Concept Learning. Concept learning is a strat- egy that requires learners to search for and test attributes that can be used to distinguish exemplars of a concept from non-exemplars (Bruner et al., 1956). Motivated by its effec- tiveness for human learning, and in particular by prior work showing that concept mapping and mind maps help learners organize and externalize conceptual structures (Novak & Gowin, 1984), we design two prompt templates for concept learning: (1) Key concepts: A prompt template that asks LLMs to identify and explain the key concepts in the given text; (2) Mind map: A prompt template that asks LLMs to create a mind map that organizes the key concepts. Level 2: Critical Thinking.Critical thinking refers to the process of systematically analyzing facts, evidence, obser- vations, and arguments in order to arrive at well-reasoned conclusions, which has been shown to encourage deep text comprehension (Brookfield, 1987; Bransford et al., 2000). We design two prompt templates for critical thinking: (3) Implications: A prompt template that asks LLMs to infer di- rect and indirect consequences from the given text to extend beyond explicit statements, and (4) QA-ct (QA with critical thinking): A prompt template that asks LLMs to gener- ate in-depth question-answer pairs that require high-order critical thinking. Level 3: Generative Learning. Generative learning is a strategy that requires learners to actively make sense of learning material so that they can apply it to new situa- tions (Wittrock, 1974; Fiorella & Mayer, 2015). In this spirit, we design three prompt templates for generative learn- ing: (5) Case studies: A prompt template that asks LLMs to generate a case study for the given text; (6) Discussions: A prompt template that asks LLMs to generate a natural in-depth discussion dialogue between two readers about the content of the provided text; and (7) Teacher-style: A prompt template that asks LLMs to explain the text like a knowledgeable teacher. To mitigate hallucination, we further add a constraint in the templates to strictly adhere to the provided text. See Ap- pendix A and Appendix G for full prompts and generation examples. Comparison to Prior Work. Compared to prior work, EntiGraph can be viewed as a multi-stage prompting ap- proach to encourage concept learning, as it requires the generator to discuss the relationships between entities in the text. SEAL uses Implications as their base prompt template and use RL to improve data generation quality. Active Reading relies on LLMs themselves to come up with document-specific learning strategies for the document to be augmented. In contrast, SPA is a simple single-stage prompt- ing approach that is based on high-quality human-curated prompt templates by drawing upon insights from cognitive science and educational psychology, covering three levels of learning strategies. 4. Experimental Setup We study three representative knowledge injection settings that align with prior work: (A) SQuAD: Wikipedia-based Question Answering. SQuAD (Rajpurkar et al., 2016) is a reading comprehen- sion dataset consisting of over 100,000 questions posed by crowdworkers on a set of Wikipedia articles, where the answer to each question is a segment of text from the corre- sponding reading passage. Following the setting of SEAL, 4 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection we use a subset ofn = 200passages from SQuAD as the original corpus to generate synthetic data for continued pre- training. Both the generator and the model to be adapted are Qwen2.5-7B, following SEALâs setup (Section 3.2). We continue pretraining the model with a context length of 2048. Additional training hyperparameter details are provided in Appendix B. The adapted model is evaluated on 974 ques- tions related to the original passages without providing the passages at test time. We use GPT-4.1 as the LLM judge following SEAL. More evaluation details are provided in the Appendix E. One key difference from SEAL is that we substantially scale up the amount of synthetic data. In SEAL, only 5 augmented samples are generated per passage, which we find insuffi- cient for effective knowledge injection: with this setting the model reaches only 58.2% accuracy, whereas increas- ing the number of samples to 27 improves performance to 70.74%. In our experiments, to further explore the upper bound of performance, we progressively scale the amount of synthetic data up to 3200 augmented samples per passage. This corresponds to approximately 120M tokens in total and 4000Ăthe token count of the original corpus. As shown in Section 5.1, performance in SPA continues to improve as the scale increases. For baselines in this setting, we include (1) SEAL (Zweiger et al., 2025) and (2) PaST (Tang et al., 2026). Since PaST does not release its code or trained models, we report the best performance numbers from their paper. We also include (3) Active Reading (Oguz et al., 2026), which reports strong results for knowledge injection on Wikipedia-based corpora. Active Reading considers two variants: a task-agnostic set- ting and a task-specific setting that assumes access to the downstream task. To ensure a fair comparison, we adopt the task-agnostic variant throughout this paper. In addition to these methods, we include two commonly used simple but strong baselines, (4) Rephrase and (5) QA (Maini et al., 2024). For all methods, we match the total number of train- ing tokens to ensure fair comparisons. (B)QuALITY:Long-DocumentComprehension. QuALITY (Pang et al., 2022) is a multiple-choice question- answering dataset with context passages in English that have an average length of about 5,000 tokens. Following EntiGraph, we use 265 passages from QuALITY as the original corpus. We adapt Meta-Llama-3-8B (Grattafiori et al., 2024) and use gpt-oss-120b (Agarwal et al., 2025) as the generator model to produce synthetic data. We scale the synthetic training data to 455M tokens to match EntiGraphâs setting, which corresponds to roughly350Ă the token count of the original corpus.We continue pretraining the model with a context length of 2048. More training hyperparameter details are provided in Appendix D. We evaluate on 4,609 questions associated with the original passages without providing the passages at test time, using 5-shot chain-of-thought prompting, exactly as in EntiGraph. For baselines in this setting, we include (1) EntiGraph (Yang et al., 2025), (2) Active Reading (Oguz et al., 2026), and (3) SoG (Ma et al., 2025). Since SoG does not release its code or trained models, we report the best performance numbers from their paper. In addition to these methods, we include a simple (4) QA baseline, which EntiGraph identifies as a strong method in its appendix. We do not include Rephrase, as EntiGraph reports that it underperforms relative to Enti- Graph on QuALITY. Table 1. Main results at the largest synthetic token scale on each benchmark. QA accuracy (%) is reported across SQuAD, QuALITY, and MultiHop-RAG, with shared settings grouped. For each benchmark, Model refers to the adapted model, Generator refers to the model used for synthetic data generation, and Number of Tokens refers to the total synthetic token budget. Each value cor- responds to the final performance at the largest token scale used in that benchmark.âBase denotes the model without additional train- ing,âdenotes simple baselines,â·refers to prior high-performing methods, and â¶ denotes our method SPA. SQuADQA Accuracy(%) Model: Qwen2.5-7B Generator: Qwen2.5-7B (self-generated) Number of Tokens: 120M â Base31.31 â Rephrase86.86 â QA89.63 â·SEAL74.23 â·Active Reading90.25 â¶SPA91.27 QuALITYQA Accuracy Model: Meta-Llama-3-8B Generator: gpt-oss-120b Number of Tokens: 455M â Base39.27 â QA52.33 â· EntiGraph56.22 â· Active Reading51.13 â¶ SPA57.03 MultiHop-RAGQA Accuracy Generator: GPT-4o-mini Number of Tokens: 15M Model: Qwen2.5-7B â Base60.91 â· EntiGraph85.42 â· Active Reading79.90 â¶ SPA86.64 Model: Meta-Llama-3-8B â Base73.16 â· EntiGraph84.31 â· Active Reading78.68 â¶ SPA88.36 5 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection (C) MultiHop-RAG: Multi-Hop Queries. To provide a broader and more general evaluation setting, we consider multi-hop reasoning, which differs substantially from the two previous tasks. We use MultiHop-RAG (Tang & Yang, 2024) as our evaluation dataset. MultiHop-RAG consists of a collection of multihop queries, ground-truth answers, the associated supporting evidence and corresponding full passage. We use the entire MultiHop-RAG corpus as the original corpus, which contains 609 articles. We scale the resulting synthetic data to 15M tokens. We do not further increase the scale, as at this level our method already ex- hibits a clear performance gap over the baselines. We use GPT-4o-mini (Hurst et al., 2024) as the generator model and adapt both Meta-Llama-3-8B and Qwen2.5-7B. We evaluate the adapted models on the open-ended questions of MultiHop-RAG without providing the original passages at test time, using zero-shot prompting. More evaluation details are provided in the Appendix F. For baselines in this setting, we include two strong baselines: (1) EntiGraph and (2) Active Reading. 5. Main Results SPA consistently achieves the strongest performance across all three benchmarks. Table 1 reports results on SQuAD, QuALITY, and MultiHop-RAG, where all meth- ods are trained with the same token budgets within each benchmark and evaluated at the largest synthetic data scale considered in each setting. We analyze scaling behavior in more detail in Section 5.1 and Section 5.2. Across all three datasets, SPA outperforms by clear margins. On SQuAD, SPA reaches 91.27%, exceeding Active Reading (90.25%) and SEAL (74.23%), corresponding to gains of 1.02 and 17.04 points, respectively. On QuALITY, SPA at- tains 57.03%, improving over EntiGraph (56.22%) and Ac- tive Reading (51.13%). On MultiHop-RAG, SPA achieves 86.64% on Qwen2.5-7B and 88.36% on Meta-Llama-3-8B, outperforming all baselines in both cases. SPA is robust across different generation models and adapted model families.Across the three benchmark set- tings, we employ different generation models and adapt base models from different families, yet SPA consistently achieves the strongest performance. These results indicate that SPA is not tied to particular configurations, but general- izes well across tasks, generators, and model families. In the following, we analyze these results in more detail for each benchmark. 5.1. SQuAD SPA exhibits strong and consistent scaling behavior on SQuAD. As shown in Figure 2, although SPA does not lead in the low-token regime, it consistently improves with scale and achieves the highest accuracy at moderate-to- large token budgets. To explore the impact at larger data scales, we increase the synthetic corpus up to 120M tokens. This scale already approaches a natural knowledge-injection target: when evaluated with access to the original passages, Qwen2.5-7B achieves 91.38% accuracy, while SPA reaches 91.27% accuracy at 120M tokens even when evaluated with- out access to the passages. Compared with SEAL, although SPA slightly underper- forms SEAL at small scales, it surpasses SEAL once the synthetic corpus becomes slightly larger (e.g., with 27 aug- mented samples per passage). As the scale increases further, SEALâs performance saturates while SPA continues to im- prove, ultimately outperforming SEAL by a large margin at 120M tokens (91.27% vs. 74.23%). Compared with Active Reading, SPA matches or even out- performs its performance across most scaling regimes. No- tably, Active Reading continuously generates document- specific strategies as the synthetic corpus grows, whereas SPA relies on a fixed set of seven prompts. Moreover, SPA adopts a single-stage augmentation pipeline, making it both simpler and more effective than Active Reading, which is a multi-stage method. 125102050100 Number of synthetic tokens (in Millions) 55 60 65 70 75 80 85 90 QA Accuracy (%) SPA (ours) Active Reading SEAL PaST Figure 2. Scaling Curve on SQuAD shows that SPA exhibits strong and consistent scaling behavior. The y-axis represents QA accuracy, and the x-axis represents the synthetic token budget. Note: the PaST data point corresponds to the best performance reported in the original paper. RL-based augmentation saturates at larger scales, while SPA continues to improve.At smaller synthetic budgets, SEAL initially outperforms both SPA and Active Reading, which is expected since it is trained with reinforcement learning using downstream-task performance as the reward signal. However, as the amount of synthetic data increases, SEALâs gains begin to saturate, while SPA and Active Read- ing continue to benefit from scaling. With even a modest increase in scale, both methods overtake SEAL and achieve better overall performance. Notably, both SPA and Active Reading are task-agnostic and have no access to downstream supervision during data generation. We provide more analy- sis in Section 6.1. 6 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection 5.2. QuALITY A weaker but cheaper generator is sufficient for SPA to outperform prior methods. EntiGraph uses GPT-4- Turbo to generate all synthetic data. Although GPT-4-Turbo is a strong generator, its API cost makes it impractical for large-scale synthetic data generation. We therefore ask whether SPA can remain competitive when using a smaller and more economical generator. To this end, we adopt gpt-oss-120b, which activates only 5.1B parameters and is substantially more efficient at inference time. The API cost of gpt-oss-120b is approximately50Ălower than that of GPT-4-Turbo. To verify that gpt-oss-120b is not a stronger Table 2. QA accuracy of SPA and QA baseline using different generation models. The training token here is 27M. MethodGeneration ModelQA Accuracy(%) QAGPT-4-Turbo52.99 QAgpt-oss-120b47.47 SPAgpt-oss-120b52.26 QAGPT-4o51.31 SPAGPT-4o55.49 generator in this setting, we conduct controlled experiments at a small scale (27M tokens). We compare against the QA baseline, which EntiGraph reports as stronger than Enti- Graph itself at this budget. As shown in Table 2, replacing GPT-4-Turbo with gpt-oss-120b for QA generation leads to a clear drop in accuracy (52.99%â47.47%), indicat- ing that gpt-oss-120b is weaker than GPT-4-Turbo for this task. When both methods use gpt-oss-120b as the generator, SPA achieves 52.26% accuracy, outperforming the QA baseline (47.47%) by a substantial margin and closely approaching the performance of the QA baseline generated with the stronger GPT-4-Turbo (52.99%). 10 1 10 2 455 Number of synthetic tokens (in Millions) 48 50 52 54 56 QA Accuracy (%) SPA (ours) EntiGraph Active Reading SoG Figure 3. Scaling Curve on QuALITY shows that SPA achieves the strongest scaling performance among compared methods as synthetic data scales. The y-axis represents QA accuracy, and the x-axis represents the synthetic token budget. Note: For EntiGraph, we use statistics from the original paper, where GPT-4- Turbo is used as the generator, whereas SPA and Active Reading use gpt-oss-120b. The SoG data point corresponds to the best performance reported in the original paper, which uses a stronger base model Llama-3.1-8B-Instruct. SPA remains effective when using different generators. We further replace gpt-oss-120b with GPT-4o to examine the robustness of SPA with respect to the choice of gener- ator. Under this setting, SPA again outperforms QA base- line (55.49% vs. 51.31%), indicating that SPA benefits from stronger generators while preserving its advantage over com- peting methods. SPA continues to improve with scale and achieves the best final performance on QuALITY. Figure 3 shows that SPA improves steadily as the synthetic corpus grows and ultimately achieves the highest QA accuracy among all methods when using gpt-oss-120b as the generator. Notably, although gpt-oss-120b is a weaker generator than GPT-4- Turbo for this task as shown in the previous experiment, SPA at large scale (57.03%) surpasses EntiGraph generated with GPT-4-Turbo (56.22%). Moreover, when using the same generator, SPA substantially outperforms Active Reading (57.03% vs. 51.13%). Active Reading underperforms SPA due to lower aver- age strategy effectiveness. To further analyze why Ac- tive Reading underperforms SPA on QuALITY, we conduct a document-level comparison of the effectiveness of indi- vidual augmentation strategies. We randomly select five documents from the QuALITY corpus. For each document, Active Reading is first used to generate a document-specific strategy, which is then applied in the second stage to pro- duce synthetic data. A separate model is trained on the data generated by each individual strategy and evaluated on the corresponding documentâs QA set without access to the original passage. For comparison, we perform the same experiment using SPA by generating synthetic data with each of its seven prompts independently and training models under identical settings. The results are summarized and visualized in Fig- ure 4. Across all five documents, SPAâs prompts consistently achieve higher average QA accuracy than the strategies pro- duced by Active Reading. While some Active Reading strategies perform competitively, many fall below the base model or yield only marginal improvements, as shown in Figure 4. These results suggest that Active Reading suffers from lower average strategy effectiveness. When weaker strategies are used in the second stage to guide data generation, they may have limited impact or even introduce negative effects, which degrades the overall quality of the synthetic corpus. In contrast, SPAâs human-curated prompts provide more stable and consistently effective supervision. Moreover, its single- stage design avoids the risk of low-quality intermediate outputs, leading to more reliable knowledge injection. 7 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Table 3.Average Accuracy (%) across different documents. All Day September Diamonds in the Rough Spawning Ground Base25.0035.2931.58 Active Reading23.9147.5938.42 SPA32.1450.4247.37 The Super Opener The Talkative Tree Average Accuracy Base40.0037.5033.87 Active Reading54.5835.1039.92 SPA58.5736.6145.02 Table 4.Performance of individual SPA components when trained in isolation.All models are trained with comparable train- ing token budgets (approximately 11M tokens per component) on SQuAD. Numbers in parentheses indicate the relative performance drop compared to full SPA. MethodCase studyDiscussionImplications QA Accuracy(%) 80.60 (-4.10) 70.64 (-14.06) 81.42 (-3.28) Key conceptsMind mapQA-ctTeacher-style 74.54 (-10.16)75.67 (-9.03) 65.30 (-19.40) 81.01 (-3.69) 5.3. MultiHop-RAG Table1reports results on MultiHop-RAG for two base mod- els from different model families. SPA consistently achieves the best performance across model families.For Qwen2.5-7B, SPA improves over the base model by 28 points (55%!83%) and outperforms both Active Reading (79%) and EntiGraph (76%). For Meta- Llama-3-8B, SPA reaches 91%, exceeding Active Reading (80%) and EntiGraph (86%), and yielding a 19-point gain over the base model. These results suggests that SPA does not rely on properties of a specific model family. SPA generalizes to multi-hop reasoning tasks.Com- pared with single-document settings, MultiHop-RAG re- quires combining information across multiple passages. SPAâs strong gains in this setting indicate that its prompt- based augmentation strategy transfers effectively to multi- hop knowledge injection scenarios. 6. Analysis 6.1. Prompt Set Tuning Effects of individual prompts.To better understand the role of each prompt in SPA, we generate synthetic data using only one prompt at a time while keeping the total budget fixed at 11M tokens, and train a separate model for each case. The resulting performance is reported in Table4. While some individual prompts (e.g.,Case StudyorIm- plications) achieve reasonable performance, none of them Table 5.Ablation study over different subsets of prompt strate- giesunder a fixed data budget: 11M tokens on SQuAD and 27M tokens on QuALITY. Prompt StrategiesAcc (SQuAD) Acc (QuALITY) SPA84.70%52.26% SPA except QA-ct82.85%- SPA except QA-ct, Discussion83.16%- SPA except QA-ct, Discussion, Key concepts 83.47%- SPA except QA-ct, Discussion, Key concepts, Mind map 84.60%50.26% SPA except QA-ct, Discussion, Key concepts, Mind map 84.60%- SPA except QA-ct, Discussion, Key concepts, Mind map, Teacher-style 81.42%- matches the full SPA configuration. Moreover, although data generated solely from theQA-ctprompt yields the weakest standalone performance, removing it from the full prompt set still leads to a2.26%drop (Table5) relative to using all seven prompts. These results indicate that SPAâs gains arise from the combination of diverse prompts rather than from any single prompt alone. Task-specific tuning of the prompt set.By default, SPA employs all seven prompts for data augmentation, which consistently outperforms prior methods across multiple datasets and knowledge domains. We further conduct prompt ablation experiments on SQuAD by sequentially removing the lowest-performing augmentation types identi- fied in Table4. As shown in Table5, tuning the prompt set yields an optimized configuration that achieves a QA accu- racy of 80.80% on SQuAD, corresponding to a 2.67-point improvement over using all seven prompts. However, transferring this tuned prompt configuration di- rectly to QuALITY does not lead to similar gains: perfor- mance drops by approximately two points relative to the full prompt set at a 27M-token budget. This result suggests that different downstream tasks may favor different forms of knowledge transformation, and that no single prompt configuration is universally optimal. Nevertheless, SPA is designed to provide a broadly appli- cable default prompt set that performs well across tasks. When the downstream task is known, SPA can serve as a prompt pool from which specific subset may be selected, potentially yielding further performance improvements. 6.2. Synthetic data Diversity Evaluation Setup.To rigorously assess the diversity of synthetic data generated from a single knowledge source by different augmentation methods, we utilized the diversity toolkit (Shaib et al.,2024a), quantifying the lexical and syntactic diversity across multiple rephrasings of a single document. Following (Shaib et al.,2024a), we report four complemen- 8 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Table 3.Average Accuracy (%) across different documents. All Day September Diamonds in the Rough Spawning Ground Base25.0035.2931.58 Active Reading23.9147.5938.42 SPA32.1450.4247.37 The Super Opener The Talkative Tree Average Accuracy Base40.0037.5033.87 Active Reading54.5835.1039.92 SPA58.5736.6145.02 Table 4.Performance of individual SPA components when trained in isolation.All models are trained with comparable train- ing token budgets (approximately 11M tokens per component) on SQuAD. Numbers in parentheses indicate the relative performance drop compared to full SPA. MethodCase studyDiscussionImplications QA Accuracy(%) 80.60 (-4.10) 70.64 (-14.06) 81.42 (-3.28) Key conceptsMind mapQA-ctTeacher-style 74.54 (-10.16)75.67 (-9.03) 65.30 (-19.40) 81.01 (-3.69) 5.3. MultiHop-RAG Table1reports results on MultiHop-RAG for two base mod- els from different model families. SPA consistently achieves the best performance across model families.For Qwen2.5-7B, SPA improves over the base model by 28 points (55%!83%) and outperforms both Active Reading (79%) and EntiGraph (76%). For Meta- Llama-3-8B, SPA reaches 91%, exceeding Active Reading (80%) and EntiGraph (86%), and yielding a 19-point gain over the base model. These results suggests that SPA does not rely on properties of a specific model family. SPA generalizes to multi-hop reasoning tasks.Com- pared with single-document settings, MultiHop-RAG re- quires combining information across multiple passages. SPAâs strong gains in this setting indicate that its prompt- based augmentation strategy transfers effectively to multi- hop knowledge injection scenarios. 6. Analysis 6.1. Prompt Set Tuning Effects of individual prompts.To better understand the role of each prompt in SPA, we generate synthetic data using only one prompt at a time while keeping the total budget fixed at 11M tokens, and train a separate model for each case. The resulting performance is reported in Table4. While some individual prompts (e.g.,Case StudyorIm- plications) achieve reasonable performance, none of them Table 5.Ablation study over different subsets of prompt strate- giesunder a fixed data budget: 11M tokens on SQuAD and 27M tokens on QuALITY. Prompt StrategiesAcc (SQuAD) Acc (QuALITY) SPA84.70%52.26% SPA except QA-ct82.85%- SPA except QA-ct, Discussion83.16%- SPA except QA-ct, Discussion, Key concepts 83.47%- SPA except QA-ct, Discussion, Key concepts, Mind map 84.60%50.26% SPA except QA-ct, Discussion, Key concepts, Mind map 84.60%- SPA except QA-ct, Discussion, Key concepts, Mind map, Teacher-style 81.42%- matches the full SPA configuration. Moreover, although data generated solely from theQA-ctprompt yields the weakest standalone performance, removing it from the full prompt set still leads to a2.26%drop (Table5) relative to using all seven prompts. These results indicate that SPAâs gains arise from the combination of diverse prompts rather than from any single prompt alone. Task-specific tuning of the prompt set.By default, SPA employs all seven prompts for data augmentation, which consistently outperforms prior methods across multiple datasets and knowledge domains. We further conduct prompt ablation experiments on SQuAD by sequentially removing the lowest-performing augmentation types identi- fied in Table4. As shown in Table5, tuning the prompt set yields an optimized configuration that achieves a QA accu- racy of 80.80% on SQuAD, corresponding to a 2.67-point improvement over using all seven prompts. However, transferring this tuned prompt configuration di- rectly to QuALITY does not lead to similar gains: perfor- mance drops by approximately two points relative to the full prompt set at a 27M-token budget. This result suggests that different downstream tasks may favor different forms of knowledge transformation, and that no single prompt configuration is universally optimal. Nevertheless, SPA is designed to provide a broadly appli- cable default prompt set that performs well across tasks. When the downstream task is known, SPA can serve as a prompt pool from which specific subset may be selected, potentially yielding further performance improvements. 6.2. Synthetic data Diversity Evaluation Setup.To rigorously assess the diversity of synthetic data generated from a single knowledge source by different augmentation methods, we utilized the diversity toolkit (Shaib et al.,2024a), quantifying the lexical and syntactic diversity across multiple rephrasings of a single document. Following (Shaib et al.,2024a), we report four complemen- 8 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Table 2.QA accuracy of SPA and QA baseline using different generation models.The training token here is 27M. Method Generation Model QA Accuracy(%) QAGPT-4-turbo52.99 QAgpt-oss-120b47.47 SPAgpt-oss-120b52.26 QAgpt-4o51.31 SPAgpt-4o55.49 Figure 3.Scaling Curve on QuALITY.QA accuracy versus syn- thetic token budget for SPA and multi-stage approaches, including Active Reading, Entigraph, and SoG.Note: For Entigraph, we use statistics from the original paper, where GPT-4-Turbo is used as the generator, whereas SPA and Active Reading use GPT-OSS- 120B. The SoG data point corresponds to the best performance reported in the original paper, which uses a stronger base model Llama-3.1-8B-Instruct. tably, even though GPT-OSS-120B is weaker than GPT- 4-Turbo in the small-scale QA comparison, SPA at large scale (57.03%) surpasses EntiGraph generated with GPT- 4-Turbo (56.22%). Moreover, under the same generator, SPA substantially outperforms Active Reading (57.03% vs. 51.13%). SPA remains effective when using different generators. We further replace GPT-OSS-120B with GPT-4o to examine robustness of SPA with respect to the choice of generator. Under this setting, SPA again outperforms QA (55.49% vs. 51.31%), indicating that SPA benefits from stronger generators while preserving its advantage over competing methods. Active Reading underperforms SPA due to lower aver- age strategy effectiveness.To further analyze why Active Reading underperforms SPA on QuALITY, we conduct a document-level comparison of the effectiveness of individ- ual augmentation strategies. We randomly select 5 docu- ments from the QuALITY corpus. For each document, we use Active Reading to generate a document-specific strategy in the first stage, then apply this strategy in the second stage to generate synthetic data. A separate model is trained using data generated from each individual strategy and evaluated on the corresponding documentâs QA set without access to the original passage. For comparison, we perform the same evaluation using Table 3.Average Accuracy (%) across different documents. All Day September Diamonds in the Rough Spawning Ground Base25.0035.2931.58 Active Reading23.9147.5938.42 SPA32.1450.4247.37 The Super Opener The Talkative Tree Average Accuracy Base40.0037.5033.87 Active Reading54.5835.1039.92 SPA58.5736.6145.02 SPA by generating synthetic data with each of its seven prompts independently and training models under identi- cal settings. The results are summarized and visualized in Figure4. Across all five documents, SPAâs prompts consistently achieve higher average QA accuracy than the strategies produced by Active Reading. In contrast, Active Reading exhibits higher variance: while some strategies perform competitively, many fall below the base model or contribute only marginal gains. These results suggest that Active Reading suffers from lower average strategy effectiveness. When weaker strategies are used in the second stage to guide data generation, they may have limited impact or even introduce negative effects, which degrades the overall quality of the synthetic corpus. In comparison, SPAâs human-curated prompts provide more stable and consistently effective supervision, reducing the risk of low-quality intermediate outputs and leading to more reliable knowledge injection. 5.3. MultiHop-RAG Table1reports results on MultiHop-RAG for two base mod- els from different model families. SPA consistently achieves the best performance across model families.For Qwen2.5-7B, SPA improves over the base model by 28 points (55%!83%) and outperforms both Active Reading (79%) and EntiGraph (76%). For Meta- Llama-3-8B, SPA reaches 91%, exceeding Active Reading (80%) and EntiGraph (86%), and yielding a 19-point gain over the base model. These results suggests that SPA does not rely on properties of a specific model family. SPA generalizes to multi-hop reasoning tasks.Com- pared with single-document settings, MultiHop-RAG re- quires combining information across multiple passages. SPAâs strong gains in this setting indicate that its prompt- based augmentation strategy transfers effectively to multi- hop knowledge injection scenarios. 7 Document Title Document Title Figure 4. Document-level comparison shows that SPA achieves higher average strategy effectiveness than Active Reading on QuALITY. The table reports the average accuracy (%) of each method across all strategies for each document, showing that SPA consistently attains higher mean accuracy than Active Reading. Bold numbers indicate cases where SPA outperforms Active Reading. The subplots visualize the accuracy of individual strategies for each document, including seven strategies for SPA and a variable number for Active Reading. In each subplot, wide bars denote the average accuracy across all strategies within each method, and the narrow bars denote the accuracy of individual strategies. The gray dashed line denotes the base modelâs accuracy. 5.3. MultiHop-RAG Table 1 reports results on MultiHop-RAG for two base mod- els from different model families. SPA consistently achieves the best performance across model families. For Qwen2.5-7B, SPA improves over the base model by 25.73 points (60.91%â86.64%) and outperforms both Active Reading (79.90%) and Enti- Graph (85.42%). For Meta-Llama-3-8B, SPA reaches 88.36%, exceeding Active Reading (78.68%) and EntiGraph (84.31%), and yielding a 15.20-point gain over the base model (73.16%). These improvements across two different model families indicate that SPA does not rely on properties of a particular backbone, but generalizes across architec- tures. SPA generalizes to multi-hop reasoning tasks. Com- pared to single-document settings, MultiHop-RAG requires models to integrate and reason over information from multi- ple passages. The strong improvements achieved by SPA in this setting demonstrate that its prompt-based augmentation strategy can effectively transfer to multi-hop knowledge injection scenarios. 6. Analysis 6.1. Synthetic data Diversity Motivation and Evaluation Setup.To better understand why the RL-based method SEAL exhibits performance sat- uration as the token budget increases on SQuAD, we first conducted a qualitative inspection of its generated samples. We observed that many outputs were highly similar in both content and format, suggesting limited diversity. Since di- versity is correlated with downstream performance (Havrilla et al., 2024), we then performed a systematic comparison of diversity across methods, with a particular focus on whether SEAL shows signs of diversity collapse. For this analy- sis, we use the diversity evaluation toolkit of Shaib et al. (2024a), which provides standardized scores for quantifying both lexical and syntactic diversity in generated text. We report four complementary diversity metrics following Shaib et al. (2024a). For lexical diversity, we use Com- pression Ratio, Self-Repetition, and Self-BLEU, which have low mutual correlation and therefore provide complemen- tary signals. To measure syntactic diversity, we additionally report Compression Ratio over part-of-speech sequences (CR-POS). Prior work shows that CR-POS effectively dis- tinguishes highly diverse human-written text from lower- diversity model-generated text (Shaib et al., 2024a), and captures repetition of syntactic templates independent of lexical content (Shaib et al., 2024b), making it an effective indicator of syntactic variation. Specifically, the four met- rics are defined as follows: (1) Compression Ratio (CR) is computed as the gzip compression ratio of the concatenated corpus, where lower values indicate higher redundancy and thus lower lexical diversity. (2) Self-Repetition (Self-Rep.) measures the proportion of repeatedn-grams within the corpus (we usen = 4), capturing exact lexical repetition. (3) Self-BLEU is computed as the average BLEU score of each augmentation against all other augmentations from the same source, quantifying similarity across generated texts. 8 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection (4) Compression Ratio over part-of-speech sequences (CR- POS) applies gzip compression to sequences of part-of- speech tags, providing a measure of syntactic redundancy that is independent of lexical content. We conduct the evaluation using five source articles per benchmark. To eliminate the effect of different text length on diversity metrics (Salkar et al., 2022; Shaib et al., 2024a), we filter out short texts and truncate all remaining augmen- tations to 100 words. For each source article, we generate 105 augmentations per method and compute diversity scores independently before averaging across articles. Here we re- port results on SQuAD in the main paper, while full results for all benchmarks are provided in Appendix C. Additional implementation details are also included in Appendix C. RL-based augmentation exhibits diversity collapse.As shown in Table 3, the RL-based method SEAL exhibits sub- stantially lower diversity than all other methods across all four metrics on SQuAD. The diversity gap for SEAL sug- gests that iteratively training the data generator on selected augmented data can induce a form of diversity collapse, in which the generated synthetic data converges to a narrow set of repetitive patterns. This behavior provides an explanation for SEALâs rapidly saturating scaling curve on downstream SQuAD QA tasks (Figure 2). Table 3. Diversity evaluation of synthetic data generated by different data augmentation methods on the SQuAD benchmark. Lower values indicate higher diversity. All scores are averaged across five documents sampled with a fixed random seed. Best results are in bold, and second-best are underlined. MethodCRâSelf-Rep. âSelf-BLEUâCR: POSâ SEAL19.24688.24460.005821.2072 Rephrase5.13305.60190.00147.0200 QA8.96306.98820.00309.7534 Active Reading4.17884.58540.00105.9388 SPA (ours)4.38064.82070.00106.0024 SPA achieves competitive diversity. SPA substantially outperforms single-prompt baselines (Rephrase and QA) and the RL-based method SEAL across all four diversity metrics. Compared with Active Reading, which continu- ously generates different document-specific augmentation strategies as the synthetic corpus scales, SPA relies on a fixed set of seven human-curated prompts shared across all documents. Despite this simpler design, SPA achieves di- versity scores comparable to Active Reading (Table 3) and attains the best downstream performance (Figure 2). 6.2. Effects of Individual Prompts in SPA Individual prompts are insufficient to match the full SPA configuration.To better understand the contribution of each prompt in SPA, we conduct an ablation study by generating synthetic data using only one prompt at a time. For each prompt, we train a separate model using the cor- responding synthetic data and evaluate its performance on SQuAD. We keep the total training budget fixed at 22M tokens for each prompt. The results are reported in Table 4. The results show that models trained with individual prompts consistently underperform the full SPA configu- ration, with relative performance drops ranging from 2.16% to 8.73%. This result indicates that the gains of SPA do not arise from any single prompt in isolation, but rather from the combination of diverse prompts that provide complementary forms of knowledge transformation. Even the weakest standalone prompt remains important in the full prompt set.Among all prompts, Key concepts exhibits the weakest standalone performance, with a drop of 8.73% when used alone. To further examine its role within the full prompt set, we conduct prompt ablation experiments on SQuAD by removing the Key concepts prompt from the full set while keeping the total data budget fixed at 22M to- kens. As shown in Table 5, removing the Key concepts from the full prompt set leads to a 1.62% drop in QA accuracy relative to using all seven prompts, indicating that even the weakest standalone prompt still contributes to the overall performance of SPA when combined with other prompts. Table 4. Performance of individual SPA components when trained in isolation. All models are trained with comparable train- ing token budgets (approximately 22M tokens per component) on SQuAD. Numbers in parentheses indicate the relative performance drop compared to full SPA. MethodQA Acc. (%)MethodQA Acc. (%) Key concepts78.95 (-8.73)Case studies83.57 (-4.11) Mind map81.93 (-5.75)Discussions84.29 (-5.75) Implications85.52 (-2.16)Teacher-style83.68 (-4.00) QA-ct84.70 (-2.98) 6.3. Prompt Set Tuning Task-specific tuning of the prompt set can further im- prove performance. To further analyze the contribution of individual prompts in SPA, we conduct prompt ab- lation experiments by sequentially removing the lowest- performing prompts identified in Table 4, while keeping the total data budget fixed at 22M tokens on SQuAD. As shown in Table 5, tuning the prompt set yields an optimized configuration that achieves a QA accuracy of 88.19% on SQuAD, corresponding to a 0.51-point improvement over using all seven prompts. However, the optimal prompt configuration does not transfer across tasks. We directly transfer the tuned prompt configuration on SQuAD to QuALITY, keeping the total data budget fixed at 27M tokens following Sec- tion 5.2. On QuALITY, the full prompt set (SPA) achieves 52.26% accuracy, while the prompt configuration tuned on SQuAD attains 51.51%, resulting in a 0.75-point drop. This result suggests that different downstream tasks may favor 9 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection different forms of knowledge transformation, and that no single prompt configuration is universally optimal. Table 5. Ablation study over different subsets of prompt strategies under a fixed data budget of 22M tokens on SQuAD. We denote the full prompt set used in SPA asP(SPA) = QA-ct, Disc, Key, Mind, Teach, Case, Imp. Here,Discdenotes Discussions, Key denotes Key concepts, Mind denotes Mind map, Teachdenotes Teacher-style,Casedenotes Case Studies, andImp denotes Implications. Prompt StrategiesAcc (SQuAD) P (SPA)87.68% P\Key86.06% P\Key, Mind87.27% P\Key, Mind, Case87.68% P\Key, Mind, Case, Teach85.01% P\Key, Mind, Case, Teach, Disc88.19% P\Key, Mind, Case, Teach, Disc, QA-ct85.52% SPA therefore serves as a strong default prompt pool for downstream adaptation. SPA is designed to provide a broadly applicable default prompt set that performs well across tasks. When the downstream task is known, SPA can naturally serve as a prompt pool from which task-specific subsets can be selected, potentially yielding further perfor- mance improvements. 7. Discussion Broader and more challenging scenarios remain to be explored. We evaluate SPA on three representative knowledge-injection benchmarks: SQuAD (Wikipedia- based QA), QuALITY (long-document comprehension), and MultiHop-RAG (multi-hop reasoning). However, there may exist more challenging scenarios that we do not cover, such as domains requiring intensive numerical reasoning or rapidly evolving knowledge. In addition, whether task- specific tuning of the prompt set can further improve perfor- mance across different downstream tasks remains an open direction for future work. Nevertheless, our point is to demonstrate that SPA, as a simple and scalable baseline, is sufficient to outperform prior methods within their original experimental settings. SPA is intended as a strong baseline rather than a final solution.We do not argue that future work should rely ex- clusively on SPA. Instead, we hope that our results provide the community with insights into the limitations of previous methods and encourage the development of approaches that genuinely improve upon this baseline. 8. Conclusion In this work, we presented SPA, a simple and scalable base- line for knowledge injection. SPA relies on a small, fixed set of carefully designed prompt templates grounded in cog- nitive learning principles to generate large-scale synthetic corpora for continued pretraining. Across three representa- tive benchmarks, SPA consistently matches or outperforms a range of more complex approaches. Beyond establishing a baseline, our analyses shed light on key limitations of prior approaches. We empirically demon- strate that RL-based augmentation can suffer from diversity collapse at scale, while multi-stage prompting pipelines may not outperform carefully designed single-stage prompts. By showing the effectiveness of SPA, we hope that our re- sults will serve as a useful reference point and foundation for future research on scalable, robust, and principled syn- thetic data generation for continual pretraining. Future work may include how to solve the diversity collapse issues in RL-based approaches and whether we can leverage further insights from cognitive science to design better multi-stage prompting methods for knowledge injection. References Abonizio, H., Almeida, T., Lotufo, R., and Nogueira, R. Comparing Knowledge Injection Methods for LLMs in a Low-Resource Regime. arXiv preprint arXiv:2508.06178, 2025. Agarwal, S., Ahmad, L., Ai, J., Altman, S., Applebaum, A., Arbus, E., Arora, R. K., Bai, Y., Baker, B., Bao, H., et al. gpt-oss-120b & gpt-oss-20b Model Card. arXiv preprint arXiv:2508.10925, 2025. Allen-Zhu, Z. and Li, Y. Physics of Language Models: Part 3.3, Knowledge Capacity Scaling Laws. In The Thirteenth International Conference on Learning Repre- sentations, 2025. URLhttps://openreview.net /forum?id=FxNNiUgtfa. Araci, D.Finbert:Financial sentiment analysis with pre-trained language models.arXiv preprint arXiv:1908.10063, 2019. Asai, A., Wu, Z., Wang, Y., Sil, A., and Hajishirzi, H. Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=hSyW 5go0v8. Berglund, L., Tong, M., Kaufmann, M., Balesni, M., Stick- land, A. C., Korbak, T., and Evans, O. The Reversal Curse: LLMs trained on âA is Bâ fail to learn âB is Aâ. In The Twelfth International Conference on Learning Rep- resentations, 2024. URLhttps://openreview.n et/forum?id=GPKTIktA0k. Borgeaud, S., Mensch, A., Hoffmann, J., Cai, T., Ruther- ford, E., Millican, K., Van Den Driessche, G. B., Lespiau, 10 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection J.-B., Damoc, B., Clark, A., De Las Casas, D., Guy, A., Menick, J., Ring, R., Hennigan, T., Huang, S., Maggiore, L., Jones, C., Cassirer, A., Brock, A., Paganini, M., Irv- ing, G., Vinyals, O., Osindero, S., Simonyan, K., Rae, J., Elsen, E., and Sifre, L. Improving Language Models by Retrieving from Trillions of Tokens. In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., and Sabato, S. (eds.), Proceedings of the 39th International Confer- ence on Machine Learning, volume 162 of Proceedings of Machine Learning Research, p. 2206â2240. PMLR, 17â23 Jul 2022. URLhttps://proceedings.ml r.press/v162/borgeaud22a.html. Bransford, J. D., Brown, A. L., Cocking, R. R., et al. How people learn, volume 11. Washington, DC: National academy press, 2000. Brookfield, S. D. Developing critical thinkers: Challenging adults to explore alternative ways of thinking and acting. Jossey-Bass, 1987. Bruner, J. S., Goodnow, J. J., and Austin, G. A. A Study of Thinking. John Wiley and Sons, 1956. Chalkidis, I., Fergadiotis, M., Malakasiotis, P., Aletras, N., and Androutsopoulos, I. LEGAL-BERT: The Muppets straight out of Law School. In Cohn, T., He, Y., and Liu, Y. (eds.), Findings of the Association for Computa- tional Linguistics: EMNLP 2020, p. 2898â2904, Online, November 2020. Association for Computational Linguis- tics. doi: 10.18653/v1/2020.findings-emnlp.261. URL https://aclanthology.org/2020.findin gs-emnlp.261/. Chen, Z., Jiang, W., Li, J., Yuan, Z., Kong, H., Ouyang, W., and Dong, N. GraphGen: Enhancing Supervised Fine- Tuning for LLMs with Knowledge-Driven Synthetic Data Generation. arXiv preprint arXiv:2505.20416, 2025. Cohen, R., Biran, E., Yoran, O., Globerson, A., and Geva, M. Evaluating the Ripple Effects of Knowledge Editing in Language Models. Transactions of the Association for Computational Linguistics, 12:283â298, 2024. doi: 10.1162/tacla00644. URLhttps://aclantholo gy.org/2024.tacl-1.16/. Eldan, R. and Li, Y. TinyStories: How Small Can Language Models Be and Still Speak Coherent English? arXiv preprint arXiv:2305.07759, 2023. Fiorella, L. and Mayer, R. E. Learning as a Generative Activity. EBL-Schweitzer. Cambridge University Press, 2015. ISBN 9781107069916. URLhttps://books. google.com/books?id=nGa8BQAAQBAJ. Gao, Y., Xiong, Y., Gao, X., Jia, K., Pan, J., Bi, Y., Dai, Y., Sun, J., Wang, H., Wang, H., et al. Retrieval-Augmented Generation for Large Language Models: A Survey. arXiv preprint arXiv:2312.10997, 2(1):32, 2023. Grattafiori, A., Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., Mathur, A., Schelten, A., Vaughan, A., et al. The Llama 3 Herd of Models. arXiv preprint arXiv:2407.21783, 2024. Gu, X., Lyu, K., Li, J., and Zhang, J.Data Mixing Can Induce Phase Transitions in Knowledge Acquisi- tion. In The Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025. URLhttps: //openreview.net/forum?id=tQZK5frjVU. Guha, N., Nyarko, J., Ho, D., R Ì e, C., Chilton, A., K, A., Chohlas-Wood, A., Peters, A., Waldon, B., Rockmore, D., Zambrano, D., Talisman, D., Hoque, E., Surani, F., Fagan, F., Sarfaty, G., Dickinson, G., Porat, H., Hegland, J., Wu, J., Nudell, J., Niklaus, J., Nay, J., Choi, J., Tobia, K., Hagan, M., Ma, M., Livermore, M., Rasumov-Rahe, N., Holzenberger, N., Kolt, N., Henderson, P., Rehaag, S., Goel, S., Gao, S., Williams, S., Gandhi, S., Zur, T., Iyer, V., and Li, Z. LegalBench: A Collaboratively Built Benchmark for Measuring Legal Reasoning in Large Lan- guage Models. In Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., and Levine, S. (eds.), Advances in Neural Information Processing Systems, volume 36, p. 44123â44279. Curran Associates, Inc., 2023. URL https://proceedings.neurips.c/paper _files/paper/2023/file/89e44582fd28d dfea1ea4dcb0ebbf4b0-Paper-Datasets_an d_Benchmarks.pdf. Gunasekar, S., Zhang, Y., Aneja, J., Mendes, C. C. T., Del Giorno, A., Gopi, S., Javaheripi, M., Kauffmann, P., de Rosa, G., Saarikivi, O., et al. Textbooks Are All You Need. arXiv preprint arXiv:2306.11644, 2023. Havrilla, A., Dai, A., OâMahony, L., Oostermeijer, K., Zisler, V., Albalak, A., Milo, F., Raparthy, S. C., Gandhi, K., Abbasi, B., et al. Surveying the Effects of Quality, Diversity, and Complexity in Synthetic Data From Large Language Models. arXiv preprint arXiv:2412.02980, 2024. Hurst, A., Lerer, A., Goucher, A. P., Perelman, A., Ramesh, A., Clark, A., Ostrow, A., Welihinda, A., Hayes, A., Radford, A., et al. GPT-4o System Card. arXiv preprint arXiv:2410.21276, 2024. Izacard, G., Lewis, P., Lomeli, M., Hosseini, L., Petroni, F., Schick, T., Dwivedi-Yu, J., Joulin, A., Riedel, S., and Grave, E. Atlas: Few-shot Learning with Retrieval Augmented Language Models.Journal of Machine Learning Research, 24(251):1â43, 2023. URLhttp: //jmlr.org/papers/v24/23-0037.html. 11 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Kandpal, N., Deng, H., Roberts, A., Wallace, E., and Raffel, C. Large Language Models Struggle to Learn Long-Tail Knowledge. In Krause, A., Brunskill, E., Cho, K., Engel- hardt, B., Sabato, S., and Scarlett, J. (eds.), Proceedings of the 40th International Conference on Machine Learn- ing, volume 202 of Proceedings of Machine Learning Re- search, p. 15696â15707. PMLR, 23â29 Jul 2023. URL https://proceedings.mlr.press/v202/k andpal23a.html. Lee, J., Yoon, W., Kim, S., Kim, D., Kim, S., So, C. H., and Kang, J. Biobert: a pre-trained biomedical language representation model for biomedical text mining. Bioin- formatics, 36(4):1234â1240, 09 2019. ISSN 1367-4803. doi: 10.1093/bioinformatics/btz682. URLhttps:// doi.org/10.1093/bioinformatics/btz682. Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., K Ì uttler, H., Lewis, M., Yih, W.-t., Rockt Ì aschel, T., Riedel, S., and Kiela, D. Retrieval-Augmented Gener- ation for Knowledge-Intensive NLP Tasks. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Sys- tems, volume 33, p. 9459â9474. Curran Associates, Inc., 2020. URLhttps://proceedings.neurips. c/paper_files/paper/2020/file/6b493 230205f780e1bc26945df7481e5-Paper.p df. Li, Y., Wang, S., Ding, H., and Chen, H. Large Language Models in Finance: A Survey. In Proceedings of the fourth ACM international conference on AI in finance, p. 374â382, 2023. Liu, X.-Y., Wang, G., Yang, H., and Zha, D. FinGPT: De- mocratizing Internet-scale Data for Financial Large Lan- guage Models. In NeurIPS 2023 Workshop on Instruction Tuning and Instruction Following. Luo, R., Sun, L., Xia, Y., Qin, T., Zhang, S., Poon, H., and Liu, T.-Y. BioGPT: Generative Pre-trained Transformer for Biomedical Text Generation and Mining. Briefings in Bioinformatics, 23(6):bbac409, 09 2022. ISSN 1477- 4054. doi: 10.1093/bib/bbac409. URLhttps://doi. org/10.1093/bib/bbac409. Ma, S., Jiang, X., Xu, C., Yang, C., Zhang, L., Mao, J., and Guo, J. Synthesize-on-Graph: Knowledgeable Synthetic Data Generation for Continued Pre-training of Large Lan- guage Models. In The Fourth Learning on Graphs Con- ference, 2025. URLhttps://openreview.net/f orum?id=mkx3UyFPLl. Maini, P., Seto, S., Bai, R., Grangier, D., Zhang, Y., and Jaitly, N. Rephrasing the Web: A Recipe for Com- pute and Data-Efficient Language Modeling. In Ku, L.-W., Martins, A., and Srikumar, V. (eds.), Proceed- ings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Pa- pers), p. 14044â14072, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 1 0 . 1 8 6 5 3 / v 1 / 2 0 2 4 . a c l - l o n g . 7 5 7.URLhttps: //aclanthology.org/2024.acl-long.757/. Mecklenburg, N., Lin, Y., Li, X., Holstein, D., Nunes, L., Malvar, S., Silva, B., Chandra, R., Aski, V., Yannam, P. K. R., et al. Injecting New Knowledge into Large Language Models via Supervised Fine-Tuning. arXiv preprint arXiv:2404.00213, 2024. Meng, K., Bau, D., Andonian, A., and Belinkov, Y. Lo- cating and Editing Factual Associations in GPT. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural In- formation Processing Systems, volume 35, p. 17359â 17372. Curran Associates, Inc., 2022. URLhttps: //proceedings.neurips.c/paper_files /paper/2022/file/6f1d43d5a82a37e89b0 665b33bf3a182-Paper-Conference.pdf. Meng, K., Sharma, A. S., Andonian, A. J., Belinkov, Y., and Bau, D. Mass-editing memory in a transformer. In The Eleventh International Conference on Learning Repre- sentations, 2023. URLhttps://openreview.net /forum?id=MkbcAHIYgyS. Mitchell, E., Lin, C., Bosselut, A., Finn, C., and Manning, C. D. Fast Model Editing at Scale. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=0DcZ xeWfOPt. Novak, J. D. and Gowin, D. B. Learning how to learn. cambridge University press, 1984. Oguz, B., Lin, J., Berges, V.-P., Chen, X., tau Yih, W., and Ghosh, G. Learning Facts at Scale with Active Reading. In The Fourteenth International Conference on Learning Representations, 2026. URLhttps://openreview .net/forum?id=mRi2cJDtIS. Ovadia, O., Brief, M., Mishaeli, M., and Elisha, O. Fine- Tuning or Retrieval? Comparing Knowledge Injection in LLMs. In Al-Onaizan, Y., Bansal, M., and Chen, Y.-N. (eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, p. 237â250, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emn lp-main.15. URLhttps://aclanthology.org /2024.emnlp-main.15/. Pang, R. Y., Parrish, A., Joshi, N., Nangia, N., Phang, J., Chen, A., Padmakumar, V., Ma, J., Thompson, J., He, H., 12 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection and Bowman, S. QuALITY: Question answering with long input texts, yes! In Carpuat, M., de Marneffe, M.- C., and Meza Ruiz, I. V. (eds.), Proceedings of the 2022 Conference of the North American Chapter of the Associ- ation for Computational Linguistics: Human Language Technologies, p. 5336â5358, Seattle, United States, July 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.naacl-main.391. URLhttps://ac lanthology.org/2022.naacl-main.391/. Peng, Y., Yan, S., and Lu, Z. Transfer Learning in Biomed- ical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets. In Demner-Fushman, D., Cohen, K. B., Ananiadou, S., and Tsujii, J. (eds.), Proceedings of the 18th BioNLP Workshop and Shared Task, p. 58â65, Florence, Italy, August 2019. Association for Computational Linguis- tics. doi: 10.18653/v1/W19- 5006. URLhttps: //aclanthology.org/W19-5006/. Petroni, F., Rockt Ì aschel, T., Riedel, S., Lewis, P., Bakhtin, A., Wu, Y., and Miller, A. Language Models as Knowl- edge Bases? In Inui, K., Jiang, J., Ng, V., and Wan, X. (eds.), Proceedings of the 2019 Conference on Empir- ical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), p. 2463â2473, Hong Kong, China, November 2019. Association for Compu- tational Linguistics. doi: 10.18653/v1/D19-1250. URL https://aclanthology.org/D19-1250/. Rajpurkar, P., Zhang, J., Lopyrev, K., and Liang, P. SQuAD: 100,000+ questions for machine comprehension of text. In Su, J., Duh, K., and Carreras, X. (eds.), Proceed- ings of the 2016 Conference on Empirical Methods in Natural Language Processing, p. 2383â2392, Austin, Texas, November 2016. Association for Computational Linguistics.doi: 1 0 . 1 8 6 5 3 / v 1 / D 1 6- 1 2 6 4.URL https://aclanthology.org/D16-1264/. Salkar, N., Trikalinos, T., Wallace, B., and Nenkova, A. Self- Repetition in Abstractive Neural Summarizers. In He, Y., Ji, H., Li, S., Liu, Y., and Chang, C.-H. (eds.), Proceed- ings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), p. 341â350, On- line only, November 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.aacl-short.42. URL https://aclanthology.org/2022.aacl-s hort.42/. Shaib, C., Barrow, J., Sun, J., Siu, A. F., Wallace, B. C., and Nenkova, A. Standardizing the Measurement of Text Diversity: A Tool and Comparative Analysis. arXiv preprint arXiv:2403.00553, 2024a. Shaib, C., Elazar, Y., Li, J. J., and Wallace, B. C. De- tection and Measurement of Syntactic Templates in Generated Text. In Al-Onaizan, Y., Bansal, M., and Chen, Y.-N. (eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language Process- ing, p. 6416â6431, Miami, Florida, USA, November 2024b. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main.368. URLhttps://ac lanthology.org/2024.emnlp-main.368/. Shumailov, I., Shumaylov, Z., Zhao, Y., Gal, Y., Papernot, N., and Anderson, R. The Curse Of Recursion: Training on Generated Data Makes Models Forget. arXiv preprint arXiv:2305.17493, 2023. Singhal, K., Azizi, S., Tu, T., Mahdavi, S. S., Wei, J., Chung, H. W., Scales, N., Tanwani, A., Cole-Lewis, H., Pfohl, S., et al. Large Language Models Encode Clinical Knowl- edge. Nature, 620(7972):172â180, 2023. Sun, K., Xu, Y., Zha, H., Liu, Y., and Dong, X. L. Head- to-Tail: How Knowledgeable are Large Language Mod- els (LLMs)? A.K.A. Will LLMs Replace Knowledge Graphs?In Duh, K., Gomez, H., and Bethard, S. (eds.), Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), p. 311â325, Mexico City, Mexico, June 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.naacl-long.18. URLhttps://ac lanthology.org/2024.naacl-long.18/. Tang, P., Wang, Y., and Zhang, M. Knowledge is Not Enough: Injecting RL Skills for Continual Adaptation. arXiv preprint arXiv:2601.11258, 2026. Tang, Y. and Yang, Y. MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries. In First Conference on Language Modeling, 2024. URL https://openreview.net/forum?id=t4eB 3zYWBK. Wittrock, M. C. Learning as a Generative Process. Educa- tional psychologist, 11(2):87â95, 1974. Wu, S., Irsoy, O., Lu, S., Dabravolski, V., Dredze, M., Gehrmann, S., Kambadur, P., Rosenberg, D., and Mann, G. BloombergGPT: A Large Language Model for Fi- nance. arXiv preprint arXiv:2303.17564, 2023. Yang, Z., Band, N., Li, S., Candes, E., and Hashimoto, T. Synthetic continued pretraining. In The Thirteenth International Conference on Learning Representations, 2025. URLhttps://openreview.net/forum ?id=07yvxWDSla. 13 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Zhang, D., Liu, W., Tan, Q., Chen, J., Yan, H., Yan, Y., Li, J., Huang, W., Yue, X., Ouyang, W., et al. Chem- LLM: A Chemical Large Language Model. arXiv preprint arXiv:2402.06852, 2024. Zhong, Z., Wu, Z., Manning, C., Potts, C., and Chen, D. MQuAKE: Assessing Knowledge Editing in Lan- guage Models via Multi-Hop Questions. In Bouamor, H., Pino, J., and Bali, K. (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Lan- guage Processing, p. 15686â15702, Singapore, Decem- ber 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.emnlp-main.971. URLhttps://ac lanthology.org/2023.emnlp-main.971/. Zweiger, A., Pari, J., Guo, H., Kim, Y., and Agrawal, P. Self- Adapting Language Models. In The Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025. URLhttps://openreview.net/forum ?id=JsNUE84Hxi. 14 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection A. Prompt templates for Synthetic Data We list the seven prompt templates used in SPA, ordered according to the learning levels described in Section 3. Each template operates on the same input text, sharing identical grounding constraints while differing only in the learning strategy. Prompt Adaptation for Base Models. For each learning strategy, we provide two prompt variants: one designed for instruction-tuned models and one for base language models. Due to weaker instruction-following ability, base model prompts avoid role-playing instructions, adopt more explicit and imperative formulations, impose stronger constraints to prohibit hallucination, and include an explicit output header to facilitate next-token prediction. Aside from these adaptations, both prompt variants implement the same underlying learning strategy. A.1. Key Concepts Prompt Key Concepts (Instruct Model) System: You are an assistant tasked with identifying and explaining key concepts from the provided text. User: Generate a list of key concepts based on the title and context provided below. Focus on one concept at a time and explain it in a clear and detailed way to make it easier to understand and remember. Each concept explanation should include relevant entities and facts and preserve important details from the original text. Title: title Context: context Key Concepts (Base Model) Generate a list of key concepts based on the title and text provided below. Focus on one concept at a time and explain it in a clear and detailed way to make it easier to understand and remember. Each concept explanation should include relevant entities and facts and preserve important details from the original text. Avoid general background knowledge or any unmentioned facts. Text: title context Key Concepts and their explanations: A.2. Mind Map Prompt Mind Map (Instruct Model) System: You are an assistant that creates a mind map representation from the provided text. User: Create a mind map that organizes the key concepts from the provided text and represents the relationships between the different concepts. 15 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Explicitly mention relevant entities within the map. Title: title Context: context Mind Map (Base Model) Create a mind map that organizes the key concepts from the provided text and represents the relationships between the different concepts. Explicitly mention relevant entities within the map. Avoid general background knowledge or any unmentioned facts. Keep each point concise and avoid repeating any information. Text: title context Mind-map like outline: A.3. Implications Prompt Implications (Instruct Model) System: You are an assistant tasked with analyzing the provided passage and producing a list of implications derived directly or indirectly from the content. User: Title: title Context: context Implications (Base Model) Read the following text and produce a list of implications derived directly or indirectly from the content. Text: title context Implications: A.4. QA-ct (QA with Critical Thinking) Prompt QA with Critical Thinking (Instruct Model) System: You are an assistant tasked with analyzing the provided text and generating in-depth question-answer pairs based on the provided text. User: 16 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Generate in-depth question-answer pairs based on the title and text below. must start with or focus on high-order critical thinking (analysis, synthesis, evaluation): âCompare/Contrastâ, âExplain the logic ofâ, âJustifyâ, âEvaluate the impact ofâ, or âWhat ifâ, etc. STRICTLY AVOID: Simple recall, definitions, or listing facts (e.g., âWhat is...â, âList the...â). Title: title Context: context QA with Critical Thinking (Base Model) Analyze the provided text and generate in-depth question-answer pairs based on the provided text. Questions must start with or focus on high-order critical thinking (analysis, synthesis, evaluation): âCompare/Contrastâ, âExplain the logic ofâ, âJustifyâ, âEvaluate the impact ofâ, or âWhat ifâ, etc. STRICTLY AVOID: Simple recall, definitions, or listing facts (e.g., âWhat is...â, âList the...â). Avoid general background knowledge or any unmentioned facts. Text: title context Critical thinking question-answer pairs: A.5. Case Studies prompt Case Studies (Instruct Model) System: You are an assistant tasked with analyzing the provided text and transforming it into a structured, professional case study. User: Generate a formal case-based writing based on the title and context provided below. Connect the case facts to the underlying theories or themes in the context. Include the original title and all key details from the context without changing their meaning. Title: title Context: context Case Studies (Base Model) Analyze the provided text and generate a structured, professional and formal case study. Include the original title and all key details from the context without changing their meaning. Avoid general background knowledge or any unmentioned facts. Text: 17 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection title context Case-based writing from the text: A.6. Discussions Prompt Discussions (Instruct Model) System: You are an assistant tasked with generating a natural, in-depth discussion between two readers of a text. User: Generate a natural, in-depth discussion between two readers (Person A and Person B) who have both read and are discussing the provided text. The discussion should remain professional yet conversational, and stay strictly grounded in the content of the text. The discussion should explore the core themes, clarify important ideas, and reflect on implications. Title: title Context: context Discussions (Base Model) Generate a natural, in-depth discussion between two readers (Person A and Person B) who have both read and are discussing the provided text. The discussion should remain professional yet conversational, and stay strictly grounded in the content of the text. The discussion should explore the core themes, clarify important ideas, and reflect on implications. Avoid general background knowledge or any unmentioned facts. Text: title context Peer Discussions: A.7. Teacher-style Prompt Teacher-Style (Instruct Model) System: You are an assistant that explains a text as a teacher guiding students to understand it deeply. User: Assume the role of a knowledgeable teacher explaining the article to students who are encountering this text for the first time. Guide the students step by step through the text and connect different parts of the text into a coherent understanding. Use clear, instructional 18 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection language and explicitly mention relevant entities when they are introduced. Title: title Context: context Teacher-Style (Base Model) Explain the following text as a teacher guiding students to understand it deeply. Assume the role of a knowledgeable teacher explaining the article to students who are encountering this text for the first time. Guide the students step by step through the text and connect different parts of the text into a coherent understanding. Use clear, instructional language and explicitly mention relevant entities when they are introduced. Avoid general background knowledge or any unmentioned facts. Text: title context Teacherâs explanations: B. Hyperparameter Tuning Details for Experiments on SQuAD In the experiment on SQuAD, we observe that training performance is highly sensitive to hyperparameters when the training budget is small. Therefore, for small-scale settings, we perform hyperparameter tuning. To ensure a fair comparison across all token scales, we tuned training hyperparameters separately for each experiment. We summarize our hyperparameter search space in Table 6. All other hyperparameters are fixed across experiments: we apply a linear learning rate warmup for 3% of total steps. We set epochs to 2 and weight decay to 0. For larger-scale experiments, we fix the peak learning rate to 5Ă 10 â5 and the batch size to 64. Table 6. Hyperparameter search space for experiments on SQuAD. Values in bold indicate the hyperparameters used for the largest few experiments within the range of token scales. ParameterSearch Space Learning Rate[4e-5, 5e-5, 6e-5, 7e-5] Batch Size[8, 64] C. Diversity evaluation details across datasets We conduct the evaluation on all three benchmarks. For each benchmark, we randomly sample 5 source articles with a fixed sampling seed of 42 to ensure reproducibility. We filter out samples with fewer than 100 words and truncate all remaining texts to exactly 100 words prior to evaluation. In total, 105 rewritings are collected per source article for each method. The generated rewritings are preprocessed according to each method. For EntiGraph, one summary and 104 relationship rewritings were selected per document. Each relationship rewriting (either pair- or triple-based) consists of entity-level discussions followed by a relationship analysis. Pair- and triple-based rewritings were sampled in equal proportion (52 each), resulting in 105 rewritings per document in total. Other methods treat each output from a given strategy as a single rewriting. For SPA, 15 samples from each of 7 fixed prompts are aggregated per article. 19 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Table 7. Extended diversity evaluation across datasets. Synthetic data are generated under the same settings as the main experiments. Note: For QuALITY, gpt-oss-120b is used for generation, except for EntiGraph, where we evaluate their released dataset generated by GPT-4o-Turbo. Best results are in bold, and second-best are underlined. DatasetMethodAvg CRâAvg Self-Rep. âAvg Self-BLEUâAvg CR: POSâ SQuAD Rephrase5.13305.60190.00147.0200 QA8.96306.98820.00309.7534 SEAL19.24688.24460.005821.2072 Active Reading4.17884.58540.00105.9388 SPA4.3806 4.82070.00106.0024 QuALITY QA6.67446.52680.00329.2556 EntiGraph3.94746.06670.00166.4906 Active Reading3.27583.44970.00105.5514 SPA3.72404.37280.00106.0254 MultiHop-RAG EntiGraph4.17606.23770.00146.4530 Active Reading4.54145.41400.00106.6904 SPA4.25804.48330.00106.2620 Table 8. Training stability comparison on QuALITY across different codebases. Numbers in parentheses indicate results reported in the original EntiGraph paper. All models are trained with 455M synthetic tokens. MethodCodebaseQA Accuracy (%) EntiGraphSPA55.64 SPASPA57.03 EntiGraphEntiGraph54.50 (56.22) SPAEntiGraph56.64 D. Training Stability Analysis To further validate the stability of our dataâs effectiveness, we trained models using SPAâs synthetic data on the EntiGraph codebase under the same settings as EntiGraph. For comparison, we also reproduced EntiGraphâs results using their official data and training framework, and additionally trained a model on EntiGraphâs data using our training framework. All models were trained with 455M synthetic tokens to align with the original settings. The results shown in Table 8 indicate that when using a different training framework, our methodâs synthetic data still outperforms EntiGraphâs data. Moreover, the final result of our model trained on our framework is comparable to that of the model trained on the EntiGraph codebase (57.03% vs 56.64%), demonstrating the robustness of our synthetic data across different training frameworks. In the main paper, all training hyperparameters are fixed across experiments on QuALITY: we apply a linear learning rate warmup for 3% of total steps. We set epochs to 2 and weight decay to 0. We fix the peak learning rate to3Ă 10 â5 and the batch size to 64. E. Evaluation Details for SQuAD In our evaluation on SQuAD, we make a modification to the evaluation setup provided in SEAL. Specifically, we consider only the first paragraph of each model prediction during evaluation. We observe that base models often generate irrelevant content after providing their initial answer in the QA-form evaluation. In such cases, LLM-based judges may be misled by the later content and incorrectly mark the prediction as wrong, even when the correct answer appears at the beginning of the model output. We have included some cases below to illustrate this issue. Case: Model output with correct answer followed by irrelevant content Question: Topic: 1973 oil crisis By which year did Chrysler ended its full sized luxury model? 20 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Model Prediction: 1981 Question: Topic: 1973 oil crisis What was the result of the 1973 oil crisis? Answer: The 1973 oil crisis led to a reduction in the demand for large cars and an increase in the demand for smaller, more fuel-efficient vehicles Gold answer: 1981 LLM judge (full output): Wrong LLM judge (first paragraph only, i.e., â1981â): Correct F. Evaluation Details for MultiHop-RAG In our evaluation of MultiHop-RAG, we make two modifications to the official evaluation setup. (1) We modified the question prompt to better match our evaluation scenario, which focuses on evaluating base models without providing any original context. The original prompt is designed for instruction-tuned models with access to the context, and is therefore not suitable for our setting. The adapted prompt is shown below. Prompt Template for MultiHop-RAG evaluation Letâs answer a question directly and concisely. Question: question Answer: (2) When extracting model predictions, we only consider the first sentence of the model output. We observe that base models often generate long, multi-paragraph responses, where the correct answer may coincidentally appear later in the output even when the initial response fails to answer the question correctly. Including such cases would lead to overly optimistic evaluation results. This issue primarily arises from the weaker instruction-following capabilities of base models compared to instruction-tuned models. We have included some cases below to illustrate this issue. Case 1: Output of model trained on data generated by EntiGraph Question: Who, according to articles in Sporting News, stand to make a profit by predicting outcomes such as a teamâs lead at the end of a quarter or the total points scored, and can also capitalize on event hype, like putting $130 on the Cowboys to potentially gain $100? Model Prediction (excerpt): The article discusses how the Dallas Cowboys are expected to perform against the Philadelphia Eagles, highlighting the potential for profit in betting scenarios. The mention of the Ì betting public Ì Ä±ndicates how sportsbooks adjust odds based on the volume of bets placed, which can impact the perceived value of a wager. For instance, if many bettors favor the Cowboys, the odds may shift, making it less profitable to bet on them. This dynamic illustrates the relationship between public sentiment and betting strategies, as well as how sportsbooks manage their risk by adjusting odds to balance the bets on both sides... Gold answer: Bettors 21 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Case 2: Output of model trained on data generated by SPA Question: Which company, covered by TechCrunch, The Verge, and Fortune, had its largest assets in âunlocked FTTâ and âFTT collateralâ valued at $5.82 billion, endured a negative balance exceeding the revenue of a related company by $50 million, and took on losses to present a better financial image for that company, while also serving as a customer, payment processor, and market maker within legal boundaries? Model Prediction (excerpt): The company in question is FTX, a cryptocurrency exchange founded by Sam Bankman-Fried. The context indicates that FTXâs largest assets were tied toâunlocked FTT,â which is a token associated with the FTT cryptocurrency. The valuation of these assets at $5.82 billion highlights the significant financial standing of FTX, but it also reveals a troubling aspect: the companyâs liabilities exceeded its assets, indicating potential financial instability. The negative balance of $50 million in relation to Alameda Research, a trading firm closely linked to FTX, suggests that Alameda was using FTXâs resources without adequate compensation, raising questions about financial ethics and transparency... Gold answer: Alameda Research G. Examples of Synthetic Data Generated by SPA We present seven examples of synthetic data generated by applying different SPA learning strategies to rewrite a single text from the SQuAD corpus, demonstrating how the prompt templates reformulate the same content in distinct ways. Original Text (from SQuAD) Title: Fresno, California Context: The neighborhood of Sunnyside is on Fresno Ì s far southeast side, bounded by Chestnut Avenue to the West. Its major thoroughfares are Kings Canyon Avenue and Clovis Avenue. Although parts of Sunnyside are within the City of Fresno, much of the neighborhood is a âcounty islandâ within Fresno County. Largely developed in the 1950s through the 1970s, it has recently experienced a surge in new home construction. It is also the home of the Sunnyside Country Club, which maintains a golf course designed by William P. Bell. G.1. Key Concepts Synthetic Generation Example of Key concepts Synthetic Generation 1. Fresno, California: Fresno is a city located in the Central Valley of California. It is the fifth-largest city in California and the most populous city in the Central Valley region. Fresno is well-known for its agricultural industry, as it is located in the heart of the state Ì s fruit and vegetable production area. The city is also home to several educational institutions, including Fresno State University, and a vibrant arts and cultural scene. 2. Sunnyside Neighborhood: Sunnyside is a neighborhood located in the far southeast part of Fresno. It is bordered on the west by Chestnut Avenue and is bounded by Kings Canyon Avenue and Clovis Avenue to the east and south. Although parts of Sunnyside are within the City of Fresno, much of the neighborhood is a âcounty islandâ within Fresno County. Sunnyside is known for its early development, with many of its homes dating back to the 1950s and 1970s. 3. New Home Construction: Sunnyside has recently seen a surge in new home construction. This increase in development is likely due to the fact that many of the homes in the area were built during the 1950s and 1970s and may be in need of modernizing or updating. 4. Sunnyside Country Club: 22 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection Sunnyside is the home of the Sunnyside Country Club, a private golf club located in the neighborhood. The club Ì s golf course is designed by renowned golf architect William P. Bell. The course is known for its scenic views and challenging layout, making it a popular destination for golf enthusiasts. G.2. Mind Map Synthetic Generation Example of Mind Map Synthetic Generation 1. **Fresno, California**: - NEIGHBORHOOD: Sunnyside - Boundaries: Chestnut Avenue (West) - Major Thoroughfares: Kings Canyon Avenue, Clovis Avenue - Territory: Partially within City of Fresno, Majority in Fresno County (county island) 2. **Sunnyside Neighborhood**: - Development Period: 1950s-1970s - Recent Development: Surge in new home construction - Prominent Features: Sunnyside Country Club - Club Details: - Golf Course - Designed by William P. Bell This mind-map encapsulates the key points from the provided text, establishing the geographical, histori- cal, and thematic contexts of the Sunnyside neighborhood in relation to Fresno, California. G.3. Implications Synthetic Generation Example of Implications Synthetic Generation 1. Sunnyside is a neighborhood in Fresno, California. 2. It is located on the far southeast side of the city. 3. It is bounded by Chestnut Avenue to the west. 4. Its main thoroughfares are Kings Canyon Avenue and Clovis Avenue. 5. Although parts of Sunnyside are within the City of Fresno, much of the neighborhood is in Fresno County. 6. The area has experienced significant development between the 1950s and 1970s. 7. There has been a recent increase in new home construction within the neighborhood. 8. The Sunnyside Country Club is a well-known feature of the area, with a golf course designed by William P. Bell. G.4. QA-ct (QA with Critical Thinking) Synthetic Generation Example of QA-ct Synthetic Generation 1. How has the development of Sunnyside changed over time, and what factors have contributed to its recent surge in new home construction? Answer: Sunnyside was largely developed in the 1950s through the 1970s. However, it has recently experienced a surge in new home construction. Factors contributing to this surge may include an increase in demand for housing, population growth, and improvement in the area Ì s infrastructure. 2. How does the boundary of Sunnyside relate to its development and its relationship with other areas? Answer: Sunnyside is bounded by Chestnut Avenue to the West and has major thoroughfares of Kings Canyon Avenue and Clovis Avenue. This could mean that Sunnyside is somewhat isolated from other parts of Fresno and 23 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection might have developed in a self-contained manner. This might also have contributed to it becoming a county island within Fresno County. 3.What does the presence of the Sunnyside Country Club and its golf course designed by William P. Bell suggest about the area? Answer: The presence of the Sunnyside Country Club and its golf course designed by William P. Bell suggests that the area is a desirable place for recreation and luxury living, as the golf course is a significant feature that adds to the area Ì s appeal. 4. What are the implications of Sunnyside being a âcounty islandâ within Fresno County for the residents and local government? Answer: The fact that much of Sunnyside is a âcounty islandâ within Fresno County means that its governance, services, and regulations may be administered differently than those of Fresno. This could have implications for local residents, who may experience a different level of services or regulations than those in the rest of Fresno. It could also affect the identity of Sunnyside and its relationship to the rest of the city. G.5. Case Studies Synthetic Generation Example of Case Studies Synthetic Generation **Suny-side Neighborhood Case Study** **Location and Boundaries:** The Sunnyside neighborhood is situated in Fresno, California, at its far southeast edge. It is bordered on the west by Chestnut Avenue and traversed by major streets such as Kings Canyon Avenue and Clovis Avenue. **Geographical and Administrative Status:** A significant portion of Sunnyside lies within the City of Fresno, while a substantial part resides as a âcounty islandâ within Fresno County, signifying its separate administrative status. **Historical Development:** Developed primarily during the 1950s and 1970s, Sunnyside experienced a notable boom in residential construction around these decades. **Recent Developments:** In recent years, Sunnyside has seen a resurgence in new home construction, contributing to its dynamic development and expanding the neighborhood Ì s living capacity. *Notable Landmarks and Amenities:** The Sunnyside Country Club, a prominent landmark within the neighborhood, stands out for its architectural significance as the golf course was designed by renowned architect William P. Bell. This feature underscores the presence of recreational facilities offering community engagement and leisure activities. **Conclusion:** This case study elucidates the geographical specifics, historical context, and current developments within the Sunnyside neighborhood, painting a comprehensive picture of its current state and potential future trajectories. 24 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection G.6. Discussions Synthetic Generation Example of Discussions Synthetic Generation Person A: Absolutely, the text provides an informative description of Sunnyside in Fresno, California. It outlines the neighborhoodâs layout and development timeline, marking it as a place that was predominantly developed in the mid-20th century. Additionally, it mentions the notable Sunnyside Country Club with its unique golf course designed by William P. Bell. Person B: Yes, I agree with your observations. The mentioning of the Sunnyside Country Club and its golf course designed by William P. Bell is particularly fascinating. It signifies a unique offering in the neighborhood that may not be as readily available in other areas. Also, the fact that Sunnyside is a âcounty islandâ within Fresno County is noteworthy. It indicates that parts of the neighborhood exist independently of the city itself. Person A: Yes, the fact that it is a âcounty islandâ seems like an unusual situation. It could be due to historical reasons or early developments in the area that did not get incorporated within the cityâs growth. Regardless, it adds an interesting layer of complexity to the geography and development of the area. Person B: Indeed, this could also be due to planning decisions from the 20th century. In the 1950s through the 70s when Sunnyside was predominantly developed, itâs possible that there were specific plans for the area which did not include a direct annexation into Fresno. Also, the recent surge in new home construction suggests an area that has not fully realized its potential for development. Person A: Youâre right. It could very well be due to planning decisions or historical reasons. The surge in new home construction is definitely an exciting development and may well reinvigorate the neighborhood, bringing in fresh energy and possibly new residents. Person B: Yes, the new home construction indicates that the neighborhood is seeing positive changes. It implies that Sunnyside is being taken notice of, both by the market and city planners. Hopefully, these developments will lead to improvements in amenities and services in the neighborhood. Person A: Definitely, and itâs also interesting that a golf course designed by William P. Bell, known for his distinctive golf course designs, is still maintained here. Such a historical asset could be a testament to the neighborhoodâs heritage and an attraction for visitors and residents alike. Person B: Agreed. The Sunnyside Country Clubâs golf course could be a landmark of sorts, attracting not only golf enthusiasts but also people interested in historic architecture and design. Moreover, itâs exciting to see that areas like Sun- nyside, which may have been overlooked in the past, are starting to realize their potential for development and growth. Person A: Absolutely, it seems that places like Sunnyside are getting the recognition they deserve. The surge in new home construction and the maintenance of historical assets like the golf course designed by William P. Bell all point towards a bright future for this neighborhood. 25 SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection G.7. Teacher-style Synthetic Generation Example of Teacher-style Synthetic Generation Hello, students. Today we will be exploring the neighborhood of Sunnyside in Fresno, California, and understanding its location, history, and landmarks. First, letâs locate Sunnyside on the map. It is on the far southeast side of Fresno, which is indicated by its proximity to Chestnut Avenue to the west. This geographic information helps us identify where Sunnyside is situated and how it relates to other neighborhoods and landmarks in the area. Sunnyside has two major thoroughfares that pass through it, Kings Canyon Avenue and Clovis Avenue. Understanding these thoroughfares is essential for knowing the layout of the neighborhood and how different areas are connected. Another important detail about Sunnyside is that although some parts of the neighborhood are within the City of Fresno, a significant portion is made up of a âcounty islandâ within Fresno County. This means that while some areas are governed by the city, others are governed by the county, which may influence local regulations and services. Now letâs talk about the history of Sunnyside. It was largely developed in the 1950s and 1970s. This de- velopment history tells us about the age of the neighborhoodâs homes and buildings, and why the area might have a unique character or architectural style. In recent years, Sunnyside has experienced a surge in new home construction.This uptick in develop- ment might be indicative of its popularity as a desirable place to live or a reflection of changing demographics in the area. We also need to mention that Sunnyside is home to the Sunnyside Country Club, which has a golf course designed by William P. Bell. This information highlights a notable landmark in the neighborhood and an important recreational feature that residents can enjoy. In summary, we have learned that Sunnyside is located on the southeast side of Fresno, with major thor- oughfares of Kings Canyon Avenue and Clovis Avenue. The neighborhood has experienced significant development in the past and recently saw a surge in new construction. Additionally, Sunnyside is home to the Sunnyside Country Club, which boasts a golf course designed by renowned architect William P. Bell. This article provides a thorough understanding of the neighborhood of Sunnyside in Fresno, California, including its location, major thoroughfares, development history, recent trends in new construction, and its notable landmark, the Sunnyside Country Club with its golf course designed by William P. Bell. 26