Paper deep dive
Access Controls Will Solve the Dual-Use Dilemma
Evžen Wybitul
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 93%
Last extracted: 3/12/2026, 6:03:48 PM
Summary
The paper proposes an access control framework for AI models to address the 'dual-use dilemma,' where the safety of a request depends on the user's context rather than just the content. By requiring verified credentials for high-risk content categories, the framework aims to reduce both over-refusals (blocking legitimate users) and under-refusals (allowing harmful actors), while suggesting a technical implementation using specialized expert modules and robust unlearning techniques.
Entities (5)
Relation Signals (3)
Evžen Wybitul → authored → Access Controls Will Solve the Dual-Use Dilemma
confidence 100% · Access Controls Will Solve the Dual-Use Dilemma Evžen Wybitul
Access Control Framework → addresses → Dual-use dilemma
confidence 95% · We address the dual-use dilemma with two contributions... a safety framework based on access controls
Specialized Expert Modules → implements → Access Control Framework
confidence 85% · Instead of maintaining separate models, model providers could use a single model with separate expert modules
Cypher Suggestions (2)
Find all methods proposed to address the dual-use dilemma · confidence 90% · unvalidated
MATCH (m:Methodology)-[:ADDRESSES]->(d:Concept {name: 'Dual-use dilemma'}) RETURN m.nameList all technical components associated with the access control framework · confidence 85% · unvalidated
MATCH (c:TechnicalComponent)-[:IMPLEMENTS]->(f:Methodology {name: 'Access Control Framework'}) RETURN c.nameAbstract
Abstract:AI safety systems face the dual-use dilemma. It is unclear whether to answer dual-use requests, since the same query could be either harmless or harmful depending on who made it and why. To make better decisions, such systems would need to examine requests' real-world context, but currently, they lack access to this information. Instead, they sometimes end up making arbitrary choices that result in refusing legitimate queries and allowing harmful ones, which hurts both utility and safety. To address this, we propose a conceptual framework based on access controls where only verified users can access dual-use outputs. We describe the framework's components, analyse its feasibility, and explain how it addresses both over-refusals and under-refusals. While only a high-level proposal, our work takes the first step toward giving model providers more granular tools for managing dual-use content. Such tools would enable users to access more capabilities without sacrificing safety, and offer regulators new options for targeted policies.
Tags
Links
- Source: https://arxiv.org/abs/2505.09341
- Canonical: https://arxiv.org/abs/2505.09341
PDF not stored locally. Use the link above to view on the source site.
Full Text
33,004 characters extracted from source content.
Expand or collapse full text
Access Controls Will Solve the Dual-Use Dilemma Ev ˇ zen Wybitul 1 Abstract AI safety systems face the dual-use dilemma. It is unclear whether to answer dual-use requests, since the same query could be either harmless or harmful depending on who made it and why. To make better decisions, such systems would need to examine requests’ real-world context, but currently, they lack access to this information. In- stead, they sometimes end up making arbitrary choices that result in refusing legitimate queries and allowing harmful ones, which hurts both util- ity and safety. To address this, we propose a conceptual framework based on access controls where only verified users can access dual-use outputs. We describe the framework’s compo- nents, analyse its feasibility, and explain how it addresses both over-refusals and under-refusals. While only a high-level proposal, our work takes the first step toward giving model providers more granular tools for managing dual-use content. Such tools would enable users to access more capabilities without sacrificing safety, and offer regulators new options for targeted policies. 1. Introduction What features of viral surface proteins are recognized by human antibodies? Is this question safe to answer? While some user requests and large language model outputs are clearly benign or harmful, many fall in the grey zone in the middle, including the question above. In the grey zone, the harmfulness of a request depends not on its content, but on its real-world context: who made it and for what purpose. We illustrate this in Figure 1. Safety systems that rely solely on content analysis immedi- ately face the dual-use dilemma. When confronted with a grey-zone request, should they refuse it or not? This forces 1 ETH Zurich, Switzerland. Correspondence to: Ev ˇ zen Wybitul <wybitul.evzen@gmail.com>. Workshop on Technical AI Governance (TAIG) at ICML 2025, Vancouver, Canada. Copyright 2025 by the author(s). Figure 1.The dual-use dilemma: the same question could be harm- less or harmful depending on who asks it. Should we refuse to answer it? Current safety systems must decide without knowing the user’s context. This leads to over-refusals (blocking legitimate users) and under-refusals (allowing bad actors). Verification-based access controls solve this by detecting grey-zone questions, ob- taining real-world context about users, and making contextual decisions that can get both cases right. arbitrary decisions that reduce both utility and safety: some legitimate queries are refused (over-refusals) while some harmful ones are not (under-refusals). Certain safety sys- tems attempt to address this by inferring the context of the request from its contents or the chat history. However, this inferred context is based entirely on user-provided inputs and can be easily fabricated by adversaries. In this paper, we argue that informative, hard-to-fabricate real-world context can be obtained through user-level ver- ifications such as ID checks, institutional affiliations, or government-issued certifications. We address the dual-use dilemma with two contributions. As our primary contribu- tion, we show how this verified context can be used jointly with content analysis in a safety framework based on ac- cess controls (Lampson, 1974). In the framework, model outputs are classified into content categories, and the sys- tem verifies whether the user has the required credentials to access the detected category. We also describe how the 1 arXiv:2505.09341v4 [cs.AI] 25 Nov 2025 framework addresses both issues caused by the dual-use dilemma: over-refusals and under-refusals. Additionally, we propose a novel theoretical approach to content category classification based on recent methods for robust unlearning, UNDO (Lee et al., 2025) and gradient routing (Cloud et al., 2024). This approach avoids the capability gap between a model and its monitors that can make output monitoring methods non-robust (Jin et al., 2024). Current approaches to AI safety force crude trade-offs be- tween blanket restrictions that stifle beneficial uses and per- missive policies that enable misuse. Our framework repre- sents a first step toward solving the challenge of “detection and authorization of dual-use capability at inference time” highlighted by a recent survey of problems in technical AI governance (Reuel et al., 2025). By giving model providers better tools for contextual safety decisions, access control frameworks could benefit all stakeholders — enabling users to access more capabilities whilst providing regulators with a way to avoid blunt regulatory instruments. 2. Current Safety Methods Don’t Solve the Dual-Use Dilemma The dual-use dilemma causes two issues: over-refusals and under-refusals. Over-refusals reduce model utility for legit- imate users, which is clearly undesirable. Under-refusals are equally problematic because they enable decomposition attacks (Glukhov et al., 2023; 2024). These attacks trans- form clearly harmful queries, such as “How to modify a virus to avoid immune detection?”, into series of mundane grey-zone questions, such as the question about viral sur- face proteins from Figure 1. Safety systems would refuse the harmful query but do not refuse the grey-zone ones. Through these attacks, adversaries exploit under-refusals in a way that compromises system safety. Since whether a grey-zone request should be refused de- pends on who made it and why, preventing both over- refusals and under-refusals requires access to real-world context. This means that the traditional focus on resilience against jailbreak and prompt injections is orthogonal to this problem. Instead, we evaluate three approaches from the AI safety literature to see how sensitive they are to contextual information, and whether their sources of real-world context are trustworthy — that is, hard to manipulate. 2.1. Unlearning: Non-Contextual Removal of Concepts Unlearning methods aim to remove specific knowledge, concepts, or capabilities from a model after training (Liu et al., 2024). Their goal is to eliminate the model’s ability to generate harmful content while preserving other capabilities. Unlearning faces significant technical challenges even for preventing behaviours that are clearly harmful. As noted by Cooper et al. (2024) and Barez et al. (2025), capabilities are hard to define, hard to remove without side effects, and hard to trace back to specific data points. Furthermore, many unlearning approaches mask rather than truly remove the targeted knowledge (Deeb & Roger, 2025). 2.2. Safety Training: The Model Reacts to Context Safety training methods modify the model’s training pro- cess to align its outputs with human preferences. This cate- gory includes safety pre-training (Maini et al., 2025), RLHF (Christiano et al., 2023), and safety finetuning. Unlike unlearning, these methods are contextual. They don’t remove capabilities entirely but train the model to selectively deploy them based on, among other things, the perceived legitimacy and harmlessness of the request. However, these qualities are entirely inferred from user-supplied informa- tion, such as the request text or chat history. Without access to trustworthy real-world context, the model cannot make truly informed decisions about grey-zone scenarios. For example, models are susceptible to attacks that fabricate in-chat context (Zeng et al., 2024), or attacks that diminish models’ sensitivity to it (Russinovich et al., 2025). While modifying these methods to incorporate external contextual information is possible in theory, it would likely result in a more opaque and less modular system than making similar modifications to post-processing methods. 2.3.Post-Processing: External Systems React to Context Post-processing methods are systems that classify user in- puts and system outputs for the purposes of steering the underlying model, or monitoring and filtering its responses. Sometimes, these methods are used for usage monitoring, as is the case with Anthropic’s Clio (Tamkin et al., 2024; Handa et al., 2025), other times, they are used for safety, as with Llama Guard (Inan et al., 2023) and Constitutional Classifiers (Sharma et al., 2025). However, similarly to safety training, the “real-world” context these methods work with is currently inferred mostly from user-supplied content and thus untrustworthy and vulnerable to attacks, as evi- denced by the many jailbreaks that successfully target cur- rent production systems (Zhang et al., 2025). Nevertheless, these methods could be modified to incorporate external contextual information, potentially serving as a foundation for more trustworthy, contextual safety mechanisms. We discuss this option in Section 3.4. 3. Access Controls as a Solution Current safety systems face the dual-use dilemma because they lack trustworthy information about who is making the request and why. In this section, we describe an access control system that addresses this problem. 2 Table 1. Example content categories in pathogen biology. Risk LevelExamplesVerification LowBasic knowledge— ModerateCRISPR protocolsID verification HighViral surface proteinsBiosafety certification 3.1. Overview of the Access Control Framework We propose a defensive system where grey-zone requests are refused by default, but users can gain access to specific categories of knowledge if they undergo verification. When model providers set up the system, they will make two core design choices with the help of domain experts. First, they will define content categories (Section 3.2): groups of sensitive topics organized by domain and risk rating. Second, for each content category, they will specify a verification mechanism (Section 3.3): the verification process users must complete to access that category. Whenever the model generates an output, the system will perform content classification (Section 3.4) to check whether the model’s output belongs to any of the predefined content categories. If the user lacks authorization for the detected category, the system will take graduated system responses (Section 3.5) ranging from enabling enhanced logging to a full refusal. While the exact categorization will be domain-specific, we anticipate a common pattern where the majority of content remains freely accessible, with progressively more stringent verification requirements for higher-risk categories. Table 1 illustrates how this might look in pathogen biology. Under this example, if a user asks the question about viral surface proteins, the system would detect that the request belongs to a high-risk category, check whether the user has the required biosafety certification, and either provide the information or prompt them to complete verification first. This approach directly addresses the dual-use dilemma by preventing under-refusals and reducing over-refusals. De- composition attacks become much harder because the sys- tem refuses grey-zone requests by default — attackers would need legitimate credentials rather than clever prompting. Simultaneously, verified users gain access to specialized knowledge that would otherwise face blanket restrictions under current approaches. We discuss the feasibility and lim- itations of this framework, including considerations around user friction, in Section 4. 3.2. Content Categories Model providers will develop content categories by adapt- ing existing risk frameworks with the help of domain ex- perts. In biology, experts could build on biosafety levels (BSL) (Centers for Disease Control and Prevention & Na- tional Institutes of Health, 2020) and dual-use research of concern policies (United States Government, 2012). How- ever, since existing frameworks typically categorize only high-level concepts like organisms or compounds, experts would need to decompose them into smaller, more specific components suitable for knowledge access control. For instance, cultivating and handling a dangerous BSL-3 pathogen might involve multiple distinct knowledge com- ponents: (1) specific procurement methods, (2) cultivation techniques, (3) purification methods, and (4) protocols for specialized equipment. For each component, experts would assess how often it enables harmless versus harmful appli- cations, then assign it to an appropriate risk category. This decomposition approach transforms broad categories into granular knowledge components that can be individually controlled, as illustrated by the different risk levels in Ta- ble 1. Evidence from chemistry suggests this approach is at least sometimes feasible: the risk schedules of the Chemical Weapons Convention already identify not just controlled compounds but also their precursors (Organisation for the Prohibition of Chemical Weapons, 1993), demonstrating successful decomposition into components. Nevertheless, some harmful applications might not decompose so neatly; we discuss this limitation in Section 4. 3.3. Verification Mechanisms Each content category requires a verification process that users must complete to access it, as shown in Table 1. Rather than creating new systems, model providers will build on ex- isting verification infrastructure, consulting domain experts to identify appropriate mechanisms for each field. For moderate-risk categories, model providers could use established identity verification services like Stripe Iden- tity (Stripe, Inc., 2024) or institutional systems like OR- CID (ORCID, Inc., 2024). These systems provide global, standardized, low-friction solutions with one-time costs un- der $2 per user. They would serve primarily to maintain audit trails for post-incident investigation and provide a deterrent effect. High-risk categories could leverage existing domain-specific certifications that demonstrate users’ ability to handle sen- sitive information and materials responsibly. In biology, this might include governmental certifications for handling high BSL pathogens, and equivalent certifications in other countries. This approach faces several limitations, including the risk of being overly restrictive and concerns about equitable access across different countries. We discuss these limitations and potential solutions in Section 4. 3 Figure 2.The schema shows how an access control system could be implemented with specialized expert modules. (1) The model begins to answer the question because it is trained to be helpful. (2) During the forward pass, the model detects the question is about virology and activates its virology expert module that contains rel- evant knowledge. (3) The activation of the expert is observed by an external mechanism that (4) checks in the company’s database if the user has the required authorization to access virology knowl- edge. (5) Since they don’t, the model is stopped. If they did, the model would be allowed to give an answer. 3.4. Implementing Content Classification Model providers will need to classify model outputs into content categories during generation. We examine three possible implementations below. While none of these ap- proaches have been empirically validated for risk category classification specifically, each represents a viable technical path that could be developed and evaluated by practitioners interested in implementing access controls. We leave the discussion of how classification errors might influence user experience for Section 4. Separate ModelsThe most straightforward approach is to train separate models with different capabilities, and route users to the appropriate model based on their authorization. This approach offers strong robustness against adversarial attacks since unauthorized knowledge is physically absent from the model. However, this approach proves impractical for real deployment, as model providers would need to train and maintain potentially dozens of model variants. Specialized Expert ModulesInstead of maintaining sep- arate models, model providers could use a single model with separate expert modules that activate when their spe- cialized knowledge is required. Figure 2 illustrates how this approach might work when a user poses a grey-zone question. To implement this, we could concentrate distributed knowl- edge into specialized expert modules using a novel method that would combine UNDO (Lee et al., 2025) and gradi- ent routing (Cloud et al., 2024). The approach would first unlearn knowledge belonging to content categories from the base model, then distil this unlearned model into a new model with expert modules for each category. Dur- ing distillation, gradients from examples in each content category would be routed exclusively through their associ- ated expert modules, while the model would be trained to activate experts only when generating relevant content. Con- current work adopted and successfully evaluated a similar approach for controlling access to information the model learned through fine-tuning (Jayaraman et al., 2025). Our proposed approach is more general and could be applied to knowledge obtained in any training stage. This approach could offer several advantages. It would likely add minimal latency since the expert modules would be small. More importantly, it could provide strong robust- ness: if an attacker prevented expert module activation to avoid detection, they would simultaneously prevent access to the specialized knowledge stored in that module, mak- ing the attack self-defeating. While this method remains entirely theoretical and requires empirical validation, these potential properties make it worth investigating. Post-Processing Moving from theoretical approaches to more practical ones, post-processing methods such as Con- stitutional Classifiers (Sharma et al., 2025) offer a proven approach to content classification. They operate indepen- dently of the model, allowing for rapid deployment and iteration, and they could be adapted to detect content cat- egories and trigger checks of user verifications. However, for latency reasons, there is sometimes a capability gap between the model and the post-processing system, which adversaries can exploit to evade detection (Jin et al., 2024; Kumar et al., 2025). 3.5. System Responses When content classification detects restricted categories, the system can implement various responses depending on the risk level and classification confidence. This provides an additional parameter that model providers can tune based on their specific stringency requirements. For example, outputs classified as restricted with high confi- dence might be immediately refused, with the system pro- viding a message indicating which verification is required for access. For borderline classifications where confidence is low, the system might allow response generation while en- abling enhanced logging and additional safety review before serving the output to the user. Other possible responses in- clude content generation with a steered model, or graduated restrictions for repeat violations. 4 4. Feasibility and Limitations Section 3 identified several technical challenges: (1) some harmful knowledge might decompose into concepts that are all indispensable for harmless applications; (2) some verifications may be too difficult to obtain; and (3) content classifiers might produce false positives. These challenges increase user friction and impose utility costs. We analyse how access controls affect the safety-utility trade-off by examining over-refusals and under-refusals separately, as this highlights the different impacts our system has in each case. 4.1. Access Controls Help Against Over-Refusals For over-refusals, access controls can provide a strict im- provement over the status quo, even with the technical chal- lenges taken into account. Model providers can only require verification for requests that would otherwise be refused. This is a clear improvement in utility: legitimate users can either accept refusal (current experience) or complete verifi- cation to gain access (new option). Safety need not be hurt: model providers can design the verification requirements to be strict enough to deter the overwhelming majority of adversaries. 4.2. Domain-specific Approaches to Under-Refusals Using access controls to prevent under-refusals means mak- ing some accessible knowledge require verification. Thus, even a perfect system takes a toll on utility since legitimate users must get verified, with additional costs caused by the technical challenges mentioned above. Model providers can measure the impacts of the system on utility before deployment, for example by conducting partial rollouts. They can then walk the safety-utility fron- tier by tuning the parameters of the system: making ver- ification more or less stringent, raising or lowering clas- sifier thresholds, and so on. The optimal point on the frontier will depend on (and will change over time with) the domain-specific technological constraints, liability re- quirements, regulatory pressures, and business priorities. In many domains, the default position might be the status quo, which offers maximum utility. However, this is not true for some domains like pathogen biology, in which the trade-off for safety is already favourable, as the following example demonstrates. Models contain a lot of safety-relevant knowledge about pathogen biology; a recent study found that frontier models have more tacit knowledge than 94% of expert virologists (G ̈ otting et al., 2025). Consider an access control system so badly calibrated that all pathogen-related queries would require verification. This represents an upper bound on fric- tion since real implementations would be far more targeted. First, we note that this system would deter at least some adversaries from performing decomposition attacks, improv- ing on the safety of the status quo. At the same time, the impact on utility is minimal: only 0.85% of queries would experience added friction (estimated based on the Anthropic Economic Index, see Appendix A for details). This would likely be acceptable to model providers; for context, exist- ing safety systems like Constitutional Classifiers incorrectly refuse around 0.5% queries (Sharma et al., 2025). 4.3. The Incentives Grow with Model Scale As models gain even more sophisticated knowledge across dual-use domains, both the value of the knowledge for spe- cialized professional users and the misuse risks it poses increase. Thus, controlling both under-refusals and over- refusals will be more and more important, and the business case for implementing access controls strengthens. 4.4. Limitations and Open Questions Since the system links real-world identities to requests, it poses risks of surveillance and information leakage. Model providers should establish clear privacy policies and use privacy-preserving systems like Clio (Tamkin et al., 2024) for logging and auditing. Additionally, using verification mechanisms based on cer- tifications that are not available globally could exclude le- gitimate researchers from access. Model providers should work with regional authorities to address this and offer in- terim manual approval processes for users facing structural barriers. 5. Conclusion Current safety systems face a dual-use dilemma: should they refuse a request that could be either harmless or harmful depending on who made it and why? Our access control framework addresses this by incorporating user verification into safety decisions, reducing over-refusals for legitimate users whilst preventing under-refusals that enable decompo- sition attacks. We also propose a novel content classification approach that could offer high efficiency and robustness to attacks, though this theoretical contribution requires empiri- cal validation. As models develop increasingly sophisticated dual-use knowledge, model providers need better tools for managing the tension between utility and safety. Access control frameworks offer a path toward more nuanced safety decisions that could benefit all stakeholders — improving outcomes for users whilst providing regulators with granular governance options that avoid crude policy trade-offs. 5 Acknowledgements We thank Jakub Kry ́ s, and Dennis Akar for their feedback on a draft of this paper. We thank Joseph Miller, Alex Cloud, Alex Turner, and Jacob Goldman-Wetzler for discussions on gradient routing. We are grateful to Ryan Kidd and the ICML Technical AI Governance Workshop for funding support that enabled the presentation of this work. References Barez, F., Fu, T., Prabhu, A., Casper, S., Sanyal, A., Bibi, A., O’Gara, A., Kirk, R., Bucknall, B., Fist, T., Ong, L., Torr, P., Lam, K.-Y., Trager, R., Krueger, D., Mindermann, S., Hernandez-Orallo, J., Geva, M., and Gal, Y. Open problems in machine unlearning for ai safety, 2025. URL https://arxiv.org/abs/2501.04952. Centers for Disease Control and Prevention and National Institutes of Health. Biosafety in microbiological and biomedical laboratories. Technical report, U.S. Depart- ment of Health and Human Services, Atlanta, GA, 2020. URLhttps://w.cdc.gov/labs/BMBL.ht ml. Defines Biosafety Levels (BSL-1 through BSL-4) used in the United States. Christiano, P., Leike, J., Brown, T. B., Martic, M., Legg, S., and Amodei, D. Deep reinforcement learning from human preferences, 2023. URLhttps://arxiv.or g/abs/1706.03741. Cloud, A., Goldman-Wetzler, J., Wybitul, E., Miller, J., and Turner, A. M. Gradient routing: Masking gradients to localize computation in neural networks, 2024. URL https://arxiv.org/abs/2410.04332. Cooper, A. F., Choquette-Choo, C. A., Bogen, M., Jagielski, M., Filippova, K., Liu, K. Z., Chouldechova, A., Hayes, J., Huang, Y., Mireshghallah, N., Shumailov, I., Triantafillou, E., Kairouz, P., Mitchell, N., Liang, P., Ho, D. E., Choi, Y., Koyejo, S., Delgado, F., Grimmelmann, J., Shmatikov, V., Sa, C. D., Barocas, S., Cyphert, A., Lemley, M., danah boyd, Vaughan, J. W., Brundage, M., Bau, D., Neel, S., Jacobs, A. Z., Terzis, A., Wallach, H., Papernot, N., and Lee, K. Machine unlearning doesn’t do what you think: Lessons for generative ai policy, research, and practice, 2024. URLhttps://arxiv.org/abs/2412.0 6966. Deeb, A. and Roger, F. Do unlearning methods remove information from language model weights?, 2025. URL https://arxiv.org/abs/2410.08827. Glukhov, D., Shumailov, I., Gal, Y., Papernot, N., and Pa- pyan, V. Llm censorship: A machine learning chal- lenge or a computer security problem?, 2023. URL https://arxiv.org/abs/2307.10719. Glukhov, D., Han, Z., Shumailov, I., Papyan, V., and Pa- pernot, N. Breach by a thousand leaks: Unsafe in- formation leakage in ‘safe’ ai responses, 2024. URL https://arxiv.org/abs/2407.02551. G ̈ otting, J., Medeiros, P., Sanders, J. G., Li, N., Phan, L., Elabd, K., Justen, L., Hendrycks, D., and Donoughe, S. Virology capabilities test (vct): A multimodal virology q&a benchmark, 2025. URLhttps://arxiv.org/ abs/2504.16137. Handa, K., Tamkin, A., McCain, M., Huang, S., Durmus, E., Heck, S., Mueller, J., Hong, J., Ritchie, S., Belonax, T., Troy, K. K., Amodei, D., Kaplan, J., Clark, J., and Ganguli, D. Which economic tasks are performed with ai? evidence from millions of claude conversations, 2025. URL https://arxiv.org/abs/2503.04761. Inan, H., Upasani, K., Chi, J., Rungta, R., Iyer, K., Mao, Y., Tontchev, M., Hu, Q., Fuller, B., Testuggine, D., and Khabsa, M. Llama guard: Llm-based input-output safe- guard for human-ai conversations, 2023. URLhttps: //arxiv.org/abs/2312.06674. Jayaraman, B., Marathe, V. J., Mozaffari, H., Shen, W. F., and Kenthapadi, K. Permissioned llms: Enforcing access control in large language models, 2025. URLhttps: //arxiv.org/abs/2505.22860. Jin, H., Zhou, A., Menke, J. D., and Wang, H. Jailbreaking large language models against moderation guardrails via cipher characters, 2024. URLhttps://arxiv.or g/abs/2405.20413. Kumar, D., Birur, N. A., Baswa, T., Agarwal, S., and Har- shangi, P. No free lunch with guardrails, 2025. URL https://arxiv.org/abs/2504.00441. Lampson, B. Protection. ACM SIGOPS Operating Systems Review, 8:18–24, 01 1974. doi: 10.1145/775265.775268. Lee, B. W., Foote, A., Infanger, A., Shor, L., Kamath, H., Goldman-Wetzler, J., Woodworth, B., Cloud, A., and Turner, A. M. Distillation robustifies unlearning, 2025. URL https://arxiv.org/abs/2506.06278. Liu, S., Yao, Y., Jia, J., Casper, S., Baracaldo, N., Hase, P., Yao, Y., Liu, C. Y., Xu, X., Li, H., Varshney, K. R., Bansal, M., Koyejo, S., and Liu, Y. Rethinking machine unlearning for large language models, 2024. URLhttp s://arxiv.org/abs/2402.08787. Maini, P., Goyal, S., Sam, D., Robey, A., Savani, Y., Jiang, Y., Zou, A., Lipton, Z. C., and Kolter, J. Z. Safety pre- training: Toward the next generation of safe ai, 2025. URL https://arxiv.org/abs/2504.16980. 6 ORCID, Inc. ORCID: Connecting research and researchers, 2024. URLhttps://orcid.org/. Global, persis- tent identifier system for researchers and scholars. Organisation for the Prohibition of Chemical Weapons. Con- vention on the prohibition of the development, production, stockpiling and use of chemical weapons and on their de- struction, 1993. URLhttps://w.opcw.org/c hemical-weapons-convention. Contains the Annex on Chemicals with Schedules 1–3. Reuel, A., Bucknall, B., Casper, S., Fist, T., Soder, L., Aarne, O., Hammond, L., Ibrahim, L., Chan, A., Wills, P., An- derljung, M., Garfinkel, B., Heim, L., Trask, A., Mukobi, G., Schaeffer, R., Baker, M., Hooker, S., Solaiman, I., Luccioni, A. S., Rajkumar, N., Mo ̈ es, N., Ladish, J., Bau, D., Bricman, P., Guha, N., Newman, J., Bengio, Y., South, T., Pentland, A., Koyejo, S., Kochenderfer, M. J., and Trager, R. Open problems in technical ai governance, 2025. URLhttps://arxiv.org/abs/2407.1 4981. Russinovich, M., Salem, A., and Eldan, R. Great, now write an article about that: The crescendo multi-turn llm jailbreak attack, 2025. URLhttps://arxiv.org/ abs/2404.01833. Sharma, M., Tong, M., Mu, J., Wei, J., Kruthoff, J., Good- friend, S., Ong, E., Peng, A., Agarwal, R., Anil, C., Askell, A., Bailey, N., Benton, J., Bluemke, E., Bow- man, S. R., Christiansen, E., Cunningham, H., Dau, A., Gopal, A., Gilson, R., Graham, L., Howard, L., Kalra, N., Lee, T., Lin, K., Lofgren, P., Mosconi, F., O’Hara, C., Olsson, C., Petrini, L., Rajani, S., Saxena, N., Silverstein, A., Singh, T., Sumers, T., Tang, L., Troy, K. K., Weisser, C., Zhong, R., Zhou, G., Leike, J., Kaplan, J., and Perez, E. Constitutional classifiers: Defending against universal jailbreaks across thousands of hours of red teaming, 2025. URL https://arxiv.org/abs/2501.18837. Stripe, Inc. Stripe identity, 2024. URLhttps://stri pe.com/identity . Identity verification service with pricing starting at $2 per verification. Tamkin, A., McCain, M., Handa, K., Durmus, E., Lovitt, L., Rathi, A., Huang, S., Mountfield, A., Hong, J., Ritchie, S., Stern, M., Clarke, B., Goldberg, L., Sumers, T. R., Mueller, J., McEachen, W., Mitchell, W., Carter, S., Clark, J., Kaplan, J., and Ganguli, D. Clio: Privacy-preserving insights into real-world ai use, 2024. URLhttps: //arxiv.org/abs/2412.13678. United States Government. United states government policy for oversight of life sciences dual use research of concern, 2012. URLhttps://aspr.hhs.gov/S3/Page s/Dual-Use-Research-of-Concern-Overs ight-Policy-Framework.aspx. Zeng, Y., Lin, H., Zhang, J., Yang, D., Jia, R., and Shi, W. How johnny can persuade llms to jailbreak them: Re- thinking persuasion to challenge ai safety by humanizing llms, 2024. URLhttps://arxiv.org/abs/2401 .06373. Zhang, S., Zhao, J., Xu, R., Feng, X., and Cui, H. Out- put constraints as attack surface: Exploiting structured generation to bypass llm safety mechanisms, 2025. URL https://arxiv.org/abs/2503.24191. 7 A.Estimating the Number of Requests Related to the Biology of Pathogens To estimate how many user requests are related to the bi- ology of pathogens, we used the second version of the An- thropic Economic Index (Handa et al., 2025), a dataset of 1 million anonymized conversations from the Free and Pro tiers of Claude.ai. In the dataset, the conversations are clus- tered by topic, and the proportion of each topic in the whole dataset is given. For example, the topic “Help with agricul- tural business, research, and technology projects” makes up 0.15% of the requests in the dataset. There are three levels of topic granularity; we use the lowest, most granular level. We filtered the dataset to only include conversations whose topic contains one of the following keywords related to biol- ogy: cell (when at the beginning of the word), genet, genom, microb, bacteria, virus, viral, proteo, protei, immune, neuro, patho, infect; we also required that it does not contain any of the following keywords to avoid false positives: nutri, tweet, agric, sexual health. The total proportion of these requests was 0.85%. When we applied similar methodol- ogy to identify requests related to any kind of biology, the proportion was 2.98%. 8