Paper deep dive
Cognitive Amplification vs Cognitive Delegation in Human-AI Systems: A Metric Framework
Eduardo Di Santi
Abstract
Abstract:Artificial intelligence is increasingly embedded in human decision-making, where it can either enhance human reasoning or induce excessive cognitive dependence. This paper introduces a conceptual and mathematical framework for distinguishing cognitive amplification, in which AI improves hybrid human-AI performance while preserving human expertise, from cognitive delegation, in which reasoning is progressively outsourced to AI systems. To characterize these regimes, we define a set of operational metrics: the Cognitive Amplification Index (CAI*), the Dependency Ratio (D), the Human Reliance Index (HRI), and the Human Cognitive Drift Rate (HCDR). Together, these quantities provide a low-dimensional metric space for evaluating not only whether human-AI systems achieve genuine synergistic performance, but also whether such performance is cognitively sustainable for the human component over time. The framework highlights a central design tension in human-AI systems: maximizing short-term hybrid capability does not necessarily preserve long-term human cognitive competence. We therefore argue that human-AI systems should be designed under a cognitive sustainability constraint, such that gains in hybrid performance do not come at the cost of degradation in human expertise.
Tags
Links
- Source: https://arxiv.org/abs/2603.18677v1
- Canonical: https://arxiv.org/abs/2603.18677v1
Intelligence
Status: not_run | Model: - | Prompt: - | Confidence: 0%
Entities (0)
Relation Signals (0)
No relation signals yet.
Cypher Suggestions (0)
No Cypher suggestions yet.
Full Text
29,499 characters extracted from source content.
Expand or collapse full text
Cognitive Amplification vs Cognitive Delegation in Human–AI Systems: A Metric Framework Eduardo Di Santi Abstract Artificial intelligence is increasingly embedded in human decision-making, where it can either enhance human reasoning or induce excessive cognitive dependence. This paper introduces a conceptual and mathematical framework for distinguishing cognitive amplification, in which AI improves hybrid human–AI performance while preserving human expertise, from cognitive delegation, in which reasoning is progressively outsourced to AI systems. To characterize these regimes, we define a set of operational metrics: the Cognitive Amplification Index (CAI∗CAI^*), the Dependency Ratio (D), the Human Reliance Index (HRIHRI), and the Human Cognitive Drift Rate (HCDRHCDR). Together, these quantities provide a low-dimensional metric space for evaluating not only whether human–AI systems achieve genuine synergistic performance, but also whether such performance is cognitively sustainable for the human component over time. The framework highlights a central design tension in human–AI systems: maximizing short-term hybrid capability does not necessarily preserve long-term human cognitive competence. We therefore argue that human–AI systems should be designed under a cognitive sustainability constraint, such that gains in hybrid performance do not come at the cost of degradation in human expertise. 1 Introduction Artificial intelligence is rapidly becoming a central component of human decision-making across domains such as medicine, engineering, finance, and scientific research. In principle, AI systems can act as powerful amplifiers of human cognition, expanding the range of problems that individuals and teams can solve. However, the increasing reliance on AI introduces a structural risk: humans may gradually delegate cognitive processes to automated systems, resulting in the erosion of analytical skills, domain understanding, and critical reasoning. This paper proposes a simple but operational framework to analyze this phenomenon. We distinguish two regimes: • Cognitive Amplification: AI increases the effective intelligence of human–AI systems. • Cognitive Delegation: humans offload reasoning to AI systems, reducing their own cognitive capacity over time. To formalize this distinction, we introduce a set of measurable quantities that describe the behavior of hybrid human–AI systems. 2 Related Work The distinction between human cognition and artificial computation is a cornerstone of modern philosophy of mind and cognitive neuroscience. Central to this debate is the question of whether machines can replicate genuine understanding or if they merely simulate it. The Chinese Room argument [19] famously posits that syntactic manipulation of symbols, no matter how sophisticated, does not constitute semantic understanding. This foundational critique suggests that human–AI systems are not merely a union of two identical types of "intelligence," but rather a hybrid of different ontological processes. The conceptual motivation for this paper is deeply rooted in the debates introduced by Byrne in his analysis of computation and consciousness [5]. By exploring the limits of functionalism and the Turing Test, Byrne highlights that the "mental" cannot be easily reduced to mere output, raising critical questions about what happens when humans begin to treat AI as a functional substitute for their own reasoning [3]. If the mind is "transparent" in its self-knowledge [4], then the delegation of cognitive steps to an opaque AI model represents a fundamental shift in how humans engage with their own belief-forming processes. Furthermore, the biological grounding of human cognition suggests that intelligence is not a "dry" computational process. Solms argues that consciousness and the source of intentionality are rooted in affective, homeostatic mechanisms—the "hidden spring" of the sentient brain [23]. This neuropsychoanalytic perspective implies that human cognition is intrinsically tied to subjective experience and the Free Energy Principle [22], which prioritizes the reduction of uncertainty through active engagement with the environment. When humans delegate reasoning to AI, they risk bypassing the very "active loops" that characterize biological intelligence, potentially leading to the cognitive drift modeled in this work. This risk is partially addressed by the extended mind hypothesis, which suggests that cognitive processes can and do extend into external artifacts [7]. However, while tools can expand our reach [20], empirical research on cognitive offloading shows that the tendency to minimize internal effort often leads to long-term changes in memory and reasoning strategies [18]. Recent studies in human–AI collaboration confirm that high-performance metrics in hybrid systems can mask a dangerous overreliance [25], where "automation bias" [14] and the inheritance of AI-driven errors [27] result in a net loss of human analytical competence. Recent work on human–AI teaming emphasizes that the effectiveness of hybrid systems depends not only on AI capability but also on the quality of human–AI interaction and complementarity [1, 13, 15]. This paper proposes a conceptual and operational framework for analyzing human–AI collaboration. We distinguish two regimes: cognitive amplification, in which AI increases the effective problem-solving capacity of human–AI systems while preserving human expertise, and cognitive delegation, in which reasoning tasks are progressively outsourced to AI systems, leading to potential erosion of human cognitive capability. To formalize this distinction, we introduce a set of measurable quantities that characterize the behavior of hybrid human–AI systems, including the Cognitive Amplification Index, the Dependency Ratio, and the Human Cognitive Drift Rate. Together, these metrics define a space of collaboration regimes that allows us to evaluate not only the performance of human–AI systems, but also their long-term cognitive sustainability. 3 System Model Consider a hybrid system composed of a human and an AI agent. S=H+AS=H+A (1) where • H represents human intelligence • A represents artificial intelligence We define the effective problem-solving capacity of the system as Q(S)=Q(H,A)Q(S)=Q(H,A) (2) where Q denotes the effective capability to solve tasks within a domain. Human intelligence alone corresponds to QH=Q(H)Q_H=Q(H) (3) while AI capability alone corresponds to QA=Q(A)Q_A=Q(A) (4) The hybrid system capability is therefore QHA=Q(S)Q_HA=Q(S) (5) 4 Cognitive Amplification In the ideal scenario, AI acts as a cognitive amplifier. The hybrid system exhibits synergy between human reasoning and machine exploration. This can be modeled as QHA=QH+QA+αQHQAQ_HA=Q_H+Q_A+α Q_HQ_A (6) where α represents the strength of human–AI interaction. If α>0α>0, the system exhibits super-linear performance gains relative to the sum of the isolated components.[26, 25] This expression should be interpreted as an idealized conceptual model of human–AI synergy, rather than as a directly observable empirical law. In practical applications, task-level performance measures such as accuracy, recall, F1 score, or time-to-solution are typically bounded, and empirical evaluation is therefore better conducted using relative performance metrics defined with respect to the best standalone agent.[26] 5 Cognitive Amplification Index The original model QHA=QH+QA+αQHQAQ_HA=Q_H+Q_A+α Q_HQ_A describes an idealized interaction between human and artificial intelligence. For empirical evaluation, however, it is more informative to measure how much the hybrid system improves over the best individual component. We therefore define the Cognitive Amplification Index as CAI∗=QHA−max(QH,QA)max(QH,QA).CAI^*= Q_HA- (Q_H,Q_A) (Q_H,Q_A). (7) This index directly quantifies the relative performance gain of the hybrid system compared to the best standalone agent. CAI∗CAI^* Interpretation >0>0 Cognitive amplification over best component =0=0 Hybrid matches best component (no net gain) <0<0 Cognitive degradation (integration harms performance) Values above zero indicate that human–AI collaboration produces a genuine synergistic effect, rather than merely reproducing the performance of either humans or AI alone.[26, 25] 6 Dependency Ratio We next quantify how strongly hybrid performance depends on the AI component. Given a task-specific performance measure Q, we define the Dependency Ratio as D=QAQHA.D= Q_AQ_HA. (8) This ratio compares standalone AI performance to hybrid system performance and therefore serves as an operational measure of relative AI reliance within the hybrid configuration. High values of D indicate that hybrid performance is close to the AI baseline, suggesting that the marginal contribution of the human component is limited and that the system may be operating in an AI-dominated regime. This interpretation is consistent with empirical work on overreliance and automation bias in human–AI teams.[14, 25] For interpretive convenience, we also define a Human Reliance Index (HRI) as the complement of D: HRI=1−D=QHA−QAQHA.HRI=1-D= Q_HA-Q_AQ_HA. (9) Unlike D, the HRI does not provide independent information; rather, it expresses the same relation from the perspective of the human contribution to hybrid performance. Range D Range HRIHRI Qualitative regime <0.5<0.5 >0.5>0.5 Human-dominant cognition 0.50.5–0.80.8 0.20.2–0.50.5 Balanced collaboration >0.8>0.8 <0.2<0.2 AI-dominated cognition, risk of delegation These thresholds are not universal constants, but operational regions that help distinguish human-dominant, balanced, and AI-dominated modes of collaboration in empirical studies of human–AI interaction.[26, 25] 7 Human Cognitive Drift Recent empirical studies suggest that AI assistance can improve immediate task performance while reducing unassisted performance in subsequent evaluations [8]. To model long-term effects on human expertise, we consider the evolution of human-only performance over time. Let QH(t)Q_H(t) denote the problem-solving capability of the human when operating without AI assistance, evaluated through periodic “AI-off” assessment blocks in the same task domain. We define the Human Cognitive Drift rate as HCDR=QH(t2)−QH(t1)t2−t1.HCDR= Q_H(t_2)-Q_H(t_1)t_2-t_1. (10) Two regimes can occur: • Amplification regime HCDR≥0HCDR≥ 0 (11) Human cognition is preserved or improved over time, even in the presence of AI assistance. This regime is more likely when AI tools scaffold human reasoning rather than replace it directly.[6, 25] • Delegation regime HCDR<0HCDR<0 (12) Human cognition deteriorates over time due to excessive reliance on AI and reduced metacognitive monitoring. This phenomenon is closely related to automation bias, in which human operators tend to over-trust automated decision aids and reduce independent verification [21]. Empirical work in educational and decision-making settings shows that unrestricted access to AI can improve immediate performance while degrading unassisted performance when the AI is removed, consistent with negative human cognitive drift. Empirical research on cognitive offloading and human reliance on AI suggests that AI can improve immediate assisted performance while still undermining the user’s capacity to independently detect errors, critique outputs, or solve comparable tasks without assistance.[6, 24, 27] In practical settings, QH(t)Q_H(t) and QHAQ_HA can be approximated through task performance metrics such as problem-solving accuracy, decision quality, or task completion efficiency within a given operational domain.[6] Figure 1 summarizes the two regimes of human–AI interaction considered in this work: cognitive amplification, in which AI enhances human reasoning while preserving human cognitive capacity, and cognitive delegation, in which excessive reliance on AI leads to progressive erosion of human expertise. Cognitive AmplificationHumanIntelligenceQHQ_HArtificialIntelligenceQAQ_AActive cognitive loopcritique, validation,hypothesis refinementEnhanced hybrid systemQHAQ_HACAI∗>0, CAI^*>0,HCDR≥0 HCDR≥ 0Synergistic human–AI cognitionCognitive DelegationHumanIntelligenceQHQ_HArtificialIntelligenceQAQ_APassive cognitive loopacceptance, outsourcing,reduced verificationAI-dominated hybrid systemQH↓Q_H →1, D→ 1,HCDR<0 HCDR<0Loss of human cognitive autonomy Figure 1: Two regimes of human–AI interaction. Left: cognitive amplification emerges when AI supports an active human cognitive loop. Right: cognitive delegation arises when reasoning is progressively outsourced to AI, increasing dependency and reducing human cognitive engagement over time. 8 Regimes of Human–AI Collaboration The metrics introduced above can be interpreted as defining a state space for human–AI systems. In particular, the pair (D,CAI∗)(D,CAI^*) provides a useful low-dimensional representation of the interaction regime, where CAI∗CAI^* captures the performance gain of the hybrid system over the best standalone component, and D measures the degree of AI dominance within the collaboration. Figure 1 shows … Within this space, qualitatively distinct regimes can be identified. Human Cognitive Drift (HCDRHCDR) then determines whether a regime is stable from the standpoint of long-term human expertise or whether it tends toward progressive cognitive erosion. Figure 2 shows … Dependency Ratio D(AI dominance)Cognitive Amplification IndexCAI∗CAI^*0 Cognitive Amplification Balanced collaboration CAI∗>0CAI^*>0, moderate D HCDR≥0HCDR≥ 0 Efficient Delegation High performance AI-dominated Human-Dominant Regime Limited AI contribution Low dependency Automation Trap High dependency Cognitive degradation risk HCDR<0HCDR<0 Figure 2: Conceptual phase diagram of human–AI collaboration regimes. The horizontal axis represents the Dependency Ratio D, measuring the degree of AI dominance in the hybrid system. The vertical axis represents the Cognitive Amplification Index CAI∗CAI^*, indicating whether the human–AI system achieves genuine synergy over the best standalone component. Human Cognitive Drift (HCDRHCDR) determines whether a regime preserves or degrades long-term human expertise. These regimes illustrate that maximizing hybrid performance QHAQ_HA does not necessarily guarantee cognitively sustainable human–AI collaboration. 9 Design Implications The metric framework introduced above suggests concrete targets for the design of human–AI systems. In safety-critical domains, the goal is not only to maximize short-term hybrid performance, but to maintain cognitive amplification (CAI∗>0CAI^*>0), avoid excessive AI dominance (moderate D and non-trivial HRIHRI), and keep human cognitive drift non-negative (HCDR≥0HCDR≥ 0). This leads to three high-level design objectives: • sustain positive amplification at the system level (CAI∗>0CAI^*>0) rather than merely matching the best standalone component; • avoid AI-dominated regimes in which D→1D→ 1 and HRI→0HRI→ 0, which are associated with overreliance and automation bias; • preserve or improve human-only performance over time (HCDR≥0HCDR≥ 0), preventing cognitive atrophy when AI assistance is removed.[6, 10] These objectives can be translated into design principles for interfaces and workflows. Rather than providing final answers that humans merely accept, AI systems should scaffold human reasoning and keep users in an active cognitive loop. Possible design principles include: • Forcing explanation of reasoning steps. Interfaces can require users to articulate their own hypothesis or rationale before revealing AI suggestions or detailed explanations. This reduces the risk that users simply adopt AI outputs without independent analysis and helps maintain HCDR≥0HCDR≥ 0.[6, 12] • Exposing uncertainty and model confidence. Communicating calibrated uncertainty, confidence intervals, or alternative candidate solutions helps users calibrate trust and reduces blind acceptance of AI recommendations. This is consistent with recent work on human–AI collaborative uncertainty quantification.[17, 9] • Requiring human hypothesis generation. Interaction protocols can be structured so that the human must generate an initial diagnosis or plan, and the AI acts as a critic or augmenter rather than as a primary oracle. Such designs increase the interpretive visibility of the human contribution and make it less likely that D approaches one.[10, 16] • Supporting exploration rather than final answers. Systems can present evidence, counterfactual scenarios, or alternative solution paths that invite exploration, instead of single authoritative outputs. This shifts the role of AI from answer provider to exploration companion, encouraging deeper engagement and reducing the risk of long-term skill degradation.[10, 11] • Embedding AI-off evaluation blocks. Periodic tasks in which users must operate without AI support enable direct measurement of QH(t)Q_H(t) and thus HCDRHCDR. These assessments can be integrated into training, simulation, or certification pipelines to detect early signs of cognitive delegation.[24, 6] From an engineering perspective, these principles suggest that human–AI systems should be instrumented not only with standard performance metrics, but also with telemetry that allows continuous estimation of CAI∗CAI^*, D, and HCDRHCDR. Such instrumentation would enable organizations to detect when their human–AI workflows drift from cognitive amplification toward cognitive delegation, and to redesign interfaces and training protocols accordingly.[16, 2] 10 Illustrative Measurement of Q The framework introduced in this paper does not require a universal measure of intelligence. Instead, Q represents the effective problem-solving capability of a system within a specific task domain. In practical settings, Q can be approximated using task-level performance metrics such as decision accuracy, solution quality, task completion time, or error rates in controlled problem-solving environments. To illustrate this idea, consider a simplified engineering diagnostic task. An engineer must identify the root cause of a system anomaly from a set of sensor signals. Suppose the following performance levels are observed: System Diagnostic Accuracy Human alone (QHQ_H) 0.70 AI alone (QAQ_A) 0.80 Human + AI (QHAQ_HA) 0.92 The Cognitive Amplification Index is then estimated as CAI∗=QHA−max(QH,QA)max(QH,QA).CAI^*= Q_HA- (Q_H,Q_A) (Q_H,Q_A). (13) Substituting the observed values: CAI∗=0.92−0.800.80≈0.15.CAI^*= 0.92-0.800.80≈ 0.15. (14) The Dependency Ratio becomes D=QAQHA=0.800.92≈0.87.D= Q_AQ_HA= 0.800.92≈ 0.87. (15) Its complementary Human Reliance Index is therefore HRI=1−D≈0.13.HRI=1-D≈ 0.13. (16) In this case the hybrid system exhibits a clear gain over the best individual component (CAI∗>0CAI^*>0). However, the high value of D indicates that hybrid performance remains strongly anchored to the AI baseline, suggesting an AI-dominated configuration despite the positive value of CAI∗CAI^*. This pattern is consistent with empirical findings on overreliance, where human–AI teams can achieve high immediate performance while risking long-term degradation of human expertise.[25, 27] 10.1 Illustrative Examples of the Framework Pharmaceutical Safety Monitoring. Consider a pharmacovigilance task in which safety analysts must identify potential adverse drug reactions from large collections of clinical reports. Let Q represent the ability to correctly identify safety signals. Performance can be measured through standard metrics such as recall, precision, or F1 score. Suppose the following results are observed in a controlled evaluation: System Signal Detection Recall Human experts (QHQ_H) 0.72 AI model (QAQ_A) 0.78 Human + AI collaboration (QHAQ_HA) 0.91 The resulting Cognitive Amplification Index is CAI∗=0.91−0.780.78≈0.17.CAI^*= 0.91-0.780.78≈ 0.17. The Dependency Ratio becomes D=0.780.91≈0.86.D= 0.780.91≈ 0.86. The complementary Human Reliance Index is HRI=1−D≈0.14.HRI=1-D≈ 0.14. In this scenario, the hybrid system clearly outperforms both humans and the AI model alone (CAI∗>0CAI^*>0), indicating cognitive amplification at the system level. However, the high value of D suggests that hybrid performance remains close to the AI baseline, placing the team in an AI-dominated regime. If analysts primarily accept model suggestions without actively interrogating them, this configuration may drift toward cognitive delegation, with a risk of long-term degradation of QH(t)Q_H(t).[24, 27] Industrial Anomaly Diagnosis. Consider an engineering monitoring system in which operators must diagnose anomalies in a complex industrial process using sensor data. Let Q represent diagnostic accuracy under time constraints. Performance can be measured as the proportion of correctly identified failure modes during simulated operational scenarios. Assume the following experimental results: System Diagnostic Accuracy Human operator (QHQ_H) 0.68 AI diagnostic system (QAQ_A) 0.81 Human + AI collaboration (QHAQ_HA) 0.95 The Cognitive Amplification Index becomes CAI∗=0.95−0.810.81≈0.17.CAI^*= 0.95-0.810.81≈ 0.17. The corresponding Dependency Ratio is D=0.810.95≈0.85.D= 0.810.95≈ 0.85. The complementary Human Reliance Index is HRI=1−D≈0.15.HRI=1-D≈ 0.15. If the hybrid interface is designed so that operators must generate, critique, and refine diagnostic hypotheses—for example, by requiring explanation of reasoning steps and uncertainty inspection—the system may sustain a regime of cognitive amplification with non-negative human cognitive drift. By contrast, if operators routinely outsource diagnosis to the AI, empirical findings on automation bias and cognitive offloading suggest that QH(t)Q_H(t) may deteriorate, pushing the system into a delegation regime despite high short-term performance.[14, 6, 24] 11 Discussion The framework proposed here does not attempt to measure intelligence in an absolute sense. Instead, it provides a relative metric system to evaluate whether AI systems amplify or replace human cognition. As AI systems become more capable, this distinction becomes critical for maintaining human intellectual autonomy. Safety-critical industries such as railway systems, aviation, pharmaceutical manufacturing, and medical technology depend on highly trained human expertise for decision making under uncertainty. In these domains, artificial intelligence should not replace human judgment but rather amplify the cognitive capabilities of expert operators. This challenge is closely related to well-documented phenomena such as automation bias, in which operators may over-trust automated decision aids and reduce independent verification [14, 25]. This observation highlights a fundamental tension in the design of human–AI systems: maximizing short-term hybrid performance does not necessarily preserve long-term human expertise. A system that maximizes QHAQ_HA while driving HCDR<0HCDR<0 may achieve short-term efficiency at the cost of a progressive loss of human expertise. In this sense, the design of human–AI systems should be guided not only by immediate hybrid performance, but also by the long-term cognitive sustainability of human expertise. The framework proposed in this paper provides a way to evaluate whether human–AI systems operate in a regime of cognitive amplification or cognitive delegation, and whether such regimes are sustainable over time. This distinction is particularly relevant in safety-critical environments, where the degradation of human expertise due to excessive reliance on automation may introduce systemic risks [25, 14]. The present framework therefore suggests that human–AI systems should be optimized under a sustainability constraint: improvements in hybrid performance should not come at the expense of negative human cognitive drift. Formally, this principle can be expressed as maxQHAsubject toHCDR≥0, Q_HA to HCDR≥ 0, which states that gains in hybrid system capability should not be achieved by degrading the long-term cognitive competence of the human component. 12 Conclusion Artificial intelligence presents a dual possibility: it can amplify human intelligence or gradually replace it. The metric framework proposed in this paper provides a way to analyze and monitor this transition. Future work should focus on empirical evaluation of these metrics in real-world human–AI interaction settings. In safety-critical domains, the goal of artificial intelligence should not be the replacement of human cognition but the amplification of expert judgment. Systems that increase automation while degrading human expertise may ultimately reduce, rather than increase, the intelligence of the overall system. References [1] G. Bansal, B. Nushi, E. Kamar, E. Horvitz, and D. Weld (2021) Does the whole exceed its parts? the effect of ai explanations on complementary team performance. Proceedings of the AAAI Conference on Artificial Intelligence. Cited by: §2. [2] A. Borg and S. Wachter (2025) Automation bias in the EU AI act: on the legal implications of human oversight. European Journal of Risk Regulation 16 (1), p. 1–24. Cited by: §9. [3] A. Byrne and J. Kim (2012) Philosophy of mind. Westview Press. Cited by: §2. [4] A. Byrne (2018) Transparency and self-knowledge. Oxford University Press. Cited by: §2. [5] A. Byrne (2019) Minds and machines. Note: MITx Online / MIT Open Learning Library External Links: Link Cited by: §2. [6] G. Chirayath and D. Gerlich (2025) Cognitive offloading or cognitive overload? how AI alters the mental architecture of coping. Frontiers in Psychology 16. Cited by: §10.1, §7, §7, §7, 3rd item, 1st item, 5th item. [7] A. Clark and D. J. Chalmers (1998) The extended mind. Analysis 58 (1), p. 7–19. Cited by: §2. [8] F. Dell’Acqua, E. McFowland, and E. Mollick (2023) Navigating the jagged technological frontier: field experimental evidence of the effects of ai on knowledge worker productivity and quality. Harvard Business School Working Paper. Cited by: §7. [9] A. Dix and colleagues (2020) Uncertainty, explainability, transparency, and bias in AI. Note: Northumbria University Cited by: 2nd item. [10] D. Gerlich (2025-02) Designing AI for human expertise: preventing cognitive shortcuts. UXmatters. Cited by: 3rd item, 3rd item, 4th item. [11] D. Gerlich (2026) AI’s cognitive implications: the decline of our thinking skills?. IE Insights. Cited by: 4th item. [12] J. W. Gichoya et al. (2023) AI pitfalls and what not to do: mitigating bias in AI. npj Digital Medicine 6, p. 136. Cited by: 1st item. [13] M. Gombolay, R. Jensen, and J. Shah (2023) Human–ai teaming: foundations and challenges. ACM Computing Surveys. Cited by: §2. [14] J. Green et al. (2024) Bending the automation bias curve: a study of human- and AI-based decision making. International Studies Quarterly 68 (2). Cited by: §10.1, §11, §11, §2, §6. [15] E. Kamar (2016) Complementing ai systems with human intelligence. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16), p. 4070–4073. Cited by: §2. [16] T. W. Malone (2021) Designing the intelligent organization. MIT Sloan Management Review 62 (3). Cited by: 3rd item, §9. [17] S. Noorani et al. (2025) Human–AI collaborative uncertainty quantification. arXiv preprint. External Links: 2510.23476 Cited by: 2nd item. [18] E. F. Risko and S. J. Gilbert (2016) Cognitive offloading. Trends in Cognitive Sciences 20 (9), p. 676–688. Cited by: §2. [19] J. R. Searle (1980) Minds, brains, and programs. Behavioral and Brain Sciences 3 (3), p. 417–424. Cited by: §2. [20] H. A. Simon (1996) The sciences of the artificial. 3 edition, MIT Press, Cambridge, MA. Cited by: §2. [21] L. Skitka, K. Mosier, and M. Burdick (1999) Does automation bias decision-making?. International Journal of Human-Computer Studies. Cited by: §7. [22] M. Solms and K. Friston (2019) The hard problem of consciousness and the free energy principle. Frontiers in Psychology 9, p. 2714. Cited by: §2. [23] M. Solms (2021) The hidden spring: a journey to the source of consciousness. W. W. Norton, New York. Cited by: §2. [24] A. Sultanova, M. Evans, and J. Park (2025) The impact of artificial intelligence tools on human cognitive abilities. Innovation: Technology, Governance, Globalization 6. Cited by: §10.1, §10.1, §7, 5th item. [25] K. Vaccaro, J. Waldo, et al. (2022-06) Overreliance on AI: literature review. Technical report Microsoft Aether Working Group. Cited by: §10, §11, §11, §2, §4, §5, §6, §6, §7. [26] M. Vaccaro, A. Almaatouq, and T. W. Malone (2024) When combinations of humans and AI are useful: a systematic review and meta-analysis. Nature Human Behaviour 8, p. 2293–2303. Cited by: §4, §4, §5, §6. [27] L. Vicente and H. Matute (2023) Humans inherit artificial intelligence biases. Scientific Reports 13, p. 15737. Cited by: §10.1, §10, §2, §7.