← Back to papers

Paper deep dive

Interpretative Interfaces: Designing for AI-Mediated Reading Practices and the Knowledge Commons

Gabrielle Benabdallah

Year: 2026Venue: arXiv preprintArea: cs.HCType: PreprintEmbeddings: 27

Abstract

Abstract:Explainable AI (XAI) interfaces seek to make large language models more transparent, yet explanation alone does not produce understanding. Explaining a system's behavior is not the same as being able to engage with it, to probe and interpret its operations through direct manipulation. This distinction matters for scientific disciplines in particular: scientists who increasingly rely on LLMs for reading, citing, and producing literature reviews have little means of directly engaging with how these models process and transform the texts they generate. In this ongoing design research project, I argue for a shift from explainability to interpretative engagement. This shift moves away from accounts of system behavior to instead enable users to manipulate a model's intermediate representations. Drawing on textual scholarship, computational poetics, and the history of reading and writing technologies, including practices such as marginalia, glosses, indices, and annotation systems, I propose interpretative interfaces as interactive environments in which non-expert users can intervene in the representational space of a language model. More specifically, such interfaces will allow users to select a token and follow its trajectory through the model's intermediate layers. This way, they can observe how its semantic position shifts as context is processed, and possibly annotate the transformations they find useful or meaningful. The same way readers can create their own maps within a book through annotations and bookmarks, interpretative interfaces will allow users to inscribe their reading of a model's internal representations. The goal of this project is to reframe AI interpretability as an interaction design project rather than a purely technical one, and to open a path toward AI-mediated reading that supports interpretative engagement and critical stewardship of scientific knowledge.

Tags

ai-safety (imported, 100%)cshc (suggested, 92%)preprint (suggested, 88%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Intelligence

Status: not_run | Model: - | Prompt: - | Confidence: 0%

Entities (0)

No extracted entities yet.

Relation Signals (0)

No relation signals yet.

Cypher Suggestions (0)

No Cypher suggestions yet.

Full Text

26,811 characters extracted from source content.

Expand or collapse full text

Interpretative Interfaces: Designing for AI-Mediated Reading Practices and the Knowledge Commons Gabrielle Benabdallah gabben@uw.edu University of Washington Seattle, Washington, USA Abstract Explainable AI (XAI) interfaces seek to make large language models more transparent, yet explanation alone does not produce under- standing. Explaining a system’s behavior is not the same as being able to engage with it, to probe and interpret its operations through direct manipulation. This distinction matters for scientific disci- plines in particular: scientists who increasingly rely on LLMs for reading, citing, and producing literature reviews have little means of directly engaging with how these models process and transform the texts they generate. In this ongoing design research project, I argue for a shift from explainability to interpretative engagement. This shift moves away from accounts of system behavior to instead enable users to manipulate a model’s intermediate representations. Drawing on textual scholarship, computational poetics, and the history of reading and writing technologies, including practices such as marginalia, glosses, indices, and annotation systems, I pro- pose interpretative interfaces as interactive environments in which non-expert users can intervene in the representational space of a language model. More specifically, such interfaces will allow users to select a token and follow its trajectory through the model’s in- termediate layers. This way, they can observe how its semantic position shifts as context is processed, and possibly annotate the transformations they find useful or meaningful. The same way read- ers can create their own maps within a book through annotations and bookmarks, interpretative interfaces will allow users to inscribe their reading of a model’s internal representations. The goal of this project is to reframe AI interpretability as an interaction design project rather than a purely technical one, and to open a path to- ward AI-mediated reading that supports interpretative engagement and critical stewardship of scientific knowledge. CCS Concepts • Human-centered computing→Systems and tools for inter- action design. Keywords Explainable AI, Human-Centered Explainable AI, Interface Design, Interpretation, Reading, Knowledge Production Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org. CHI ’26, Barcelona, Spain © 2026 Copyright held by the owner/author(s). Publication rights licensed to ACM. ACM ISBN 978-x-x-x-x/Y/M https://doi.org/10.1145/n.n ACM Reference Format: Gabrielle Benabdallah. 2026. Interpretative Interfaces: Designing for AI- Mediated Reading Practices and the Knowledge Commons. In Proceedings of the CHI 2026 Workshop: Ethics at the Front-End: Responsible User-Facing Design for AI Systems (CHI ’26), April 13–17, 2026, Barcelona, Spain. ACM, New York, NY, USA, 4 pages. https://doi.org/10.1145/n.n 1 Introduction 1.1 The Limits of Explanations AI systems, particularly LLMs, are increasingly mediating the ac- tivities of knowledge production, such as reading, writing, citation, literature reviews, publishing, and the discovery of scholarly and scientific literature. LLMs are changing how researchers encounter text and, to some extent, how they assess their relevance and au- thority [14,17,26]. In the face of these shifts, there is a growing body of research concerned with developing explainable AI inter- faces for end-users. Domains such as human-centered explainable AI (HCXAI), for instance, complement the goals of AI interpretabil- ity by moving beyond the objectives of algorithmic transparency alone and instead foreground the sociotechnical questions of who can understand and make sense of AI systems, and how [4]. Espe- cially in the context of reading and writing, which inform attention, interpretation, and other cognitive processes [19,27], ethical con- cerns emerge not only in model training but at the interface level, which structure how systems are understood and used. By being the sites where the modalities of interaction, and therefore of inter- pretation, meaning-making, and agency are negotiated, interfaces articulate the possibilities of participation in the activities of knowl- edge production. Interface design choices fundamentally shape how information is received, what counts as legitimate scholarly work, as well as what reading and writing practices are valued and sustained. As AI tools become part of the infrastructures of academic work, mediating individual tasks and the social practices of citation and peer review, the interfaces through which scholars engage with these systems have to be considered as interventions in the knowl- edge commons. Designing interpretative interfaces, then, is about sustaining the interpretative practices that have historically been supported by certain reading and writing practices that invite read- ers and users into the information architecture of the system. Explainability is also only one aspect of the project of opening up AI systems to a wider range of interests and concerns. Many explainable AI interfaces focus on developing interactive modali- ties that explain the behavior or operations of LLMs and classifier systems. These modalities include graphs [6], dashboards [2], chat- based interactions [12], and natural language explanations [25], among others. These explanations, while necessary, also present arXiv:2603.15863v1 [cs.HC] 16 Mar 2026 CHI ’26, April 13–17, 2026, Barcelona, SpainBenabdallah the paradoxical issue of adding an additional layer of abstraction on already complex and opaque systems. Additionally, explanations tend toward closure, resolution, and often assume authority; as such, they do not cultivate the inquisitive and critical attitude that characterize interpretation. 1.2 Interpretation as Material Engagement Interpretation is defined as the act of explaining the meaning of something, but modern hermeneutics, the tradition dealing with interpretation, complicates what this process of arriving at an ex- planation is. In reader-response theory, Wolfgang Iser argued that meaning is not inherent to the text but a “dynamic happening” produced through the reader’s active engagement with the text [8, p. 22]. This is in part because the text, like all information systems, has gaps and indeterminacies that the reader resolves internally through creative and situated reconfigurations, “working things out” rather than receiving pre-formed explanations [8, p.276]. Read- ing, therefore, is a form of manipulation: the reader intervenes in the material of the text, filling gaps, rearranging expectations, and reconfiguring meaning against the “horizon” of their own histories and concerns. Understanding does not happen as a transmission but as an act of construction, one that is enabled through hands-on engagement. This insight extends beyond literary texts and applies to all in- formation systems. Manipulation, in the sense of direct, material intervention in a system, is a more powerful vector of understand- ing than explanation. Explanations tend toward closure by resolving ambiguity, and assume an authoritative position that comes from the system (or its authors) rather than the users. Interpretation by contrast sustains inquiry. In fact, interpretation thrives on resis- tance. As the literary scholar Yves Citton writes, “[a] century of interpretive practice has taught us to turn sites of resistance to meaning into strategic points for the emergence of “truth.” Nothing is more tedious than a textual analysis that merely confirms how clear and well-ordered the narrative is ... . Inter- pretation truly begins with the grain of sand that jams the machine that produces obviousness, ... with the detection of an incoherence, an unexpected detail, a “deviation”.” [1, p. 128] (my translation). Ambiguous, polysemic, and even “glitchy” outputs are produc- tive sites of engagement; opportunities to question and re-imagine systems rather than merely optimize them. Understanding emerges precisely where the system resists easy explanation. Explanation- centric interfaces in AI place epistemic authority at the system level: the system explains itself, and the user receives it as the voice of authority. This has the paradoxical effect of reducing agency and accountability; many explainable AI interfaces tend to substitute explanation for the ability to probe and materially engage with a system. By doing so, these interfaces foreclose the possibility of in- quiry that they could instead support. Additionally, they normalize (or unproblematize) the epistemic value of transparency, and subtly relocate authority at the system level. Interpretative interfaces, by contrast, give users the means to manipulate language models not just at the interface level but at the operational level by giving access to the intermediate layers. They allow intervention in the in- formation architecture of a system, and enable users to leave traces of their interpretative work, thereby constructing understanding on their own terms. 2 The Book: A Historical Model for Interpretative Interface Design The history of the book is a strong example of an information system that became gradually easier to manipulate (and interpret). Through the development of both technical innovations and social practices over centuries, the book evolved to allow more readers to intervene in it. This history can be instructive for the design of AI interfaces. For centuries, codices were clunky and heavy artifacts that relied on specialized information architectures and conventions, such as the absence of space between words or the use of Merovingian and Visigothic scripts, which were often illegible to most but a few domain experts (monks and scribes). The literal and technical hefti- ness of early manuscripts, as well as their interface design, meant that only a few could produce and use books [16]. Books were, effectively, ‘expert systems’ [24]. Over the centuries, a series of technical developments made the book progressively more usable. Such developments include the normalization of the caroline mi- nuscule as a calligraphic standard, the refinement of paper making techniques so that it became smoother, new binding and folding techniques which shrunk the size (and reduced weight) of codices, and eventually the printing press, which, as Elizabeth Eisenstein ar- gued, radically altered the conditions under which texts circulated and endured [5]. But usability alone does not account for how the book “opened up” to interpretation. What proved equally decisive was the de- velopment of spaces for intervention: margins, interlinear gaps, and the social practices they enabled, such as glosses, margina- lia, commonplace books, and later, independent letterpress shops and zines. These graphic and social infrastructures, now largely taken for granted, offered the material conditions for interpretative practice. The margin is, in a sense, Iser’s “gap” made physical: a space within the system where the reader can intervene, annotate, contest, and construct meaning. The book became what Ivan Illich called a “convivial” tool [7] because it invited participation at the level of its own information architecture and, through doing so, extended the technology of the book to a wider range of interests and concerns. The takeaway for the design of XAI interfaces is that accessibil- ity and interpretability are not best supported by explanation or simplification but by enabling what is simultaneously a material and symbolic form of participation, which is interpretation. 3 Towards Interpretative Interfaces for LLMs LLMs, like books, have material affordances that can be leveraged to support interpretation, and without asking users to master technical explanations. Just as the margins and interlinear gaps of the page functioned as sites for annotation and interpretation, LLMs contain representational spaces that shape how meaning is internally pro- duced: the intermediate layers of transformer-based models. While not immediately accessible to end-users, techniques from the field of Interpretative Interfaces: Designing for AI-Mediated Reading Practices and the Knowledge CommonsCHI ’26, April 13–17, 2026, Barcelona, Spain AI interpretability have made it possible to inspect and manipulate these layers [13,15,21–23]. While these techniques have predomi- nantly been used in technical scenarios to better understand how models work (and, by extension, make them better), I suggest that these intermediate representations can also be approached as part of the interaction itself. I propose interpretative interfaces: interface designs that allow users to interact with and annotate the intermediate representations of LLMs. Instead of confining interaction to the input-output level, interpretative interfaces enable direct manipulation of a model’s internal operations. The interaction is structured around token- centric reading: rather than treating the model as a black box to be explained through additional layers of abstraction (through graphs, plots, and dashboards), users select individual input tokens and trace their transformations through the model’s layers, observing how semantic positions shift as context is processed. This reframes the interaction from model-centric interpretability (what did the system do?) to token-centric reading (how was this particular unit transformed, and how did it influence “meaning”). The goal is not to exhaustively expose a model’s operations, which would be over- whelming, but to create partial windows (like margins in a manu- script) through which users can engage with, annotate, and begin interpreting how meaning is constructed within the system. This approach also raises questions about how intermediate rep- resentations are themselves visualized. As media theorist Johanna Drucker has argued, dominant modes of graphic display tend to be treated as transparent representations of the phenomena they point to, when in fact they are “as unsuited to modeling interpre- tation as a thermometer is for measuring the warmth of a human emotion or the strength of an embrace” [3, p.5]. Many phenom- ena, including language, exceed the bounds of representation–the map is not the territory [11]. Representations, whether expressed through maps, visualizations, or interfaces, are always already in- terpretations. Rather than assuming transparency, interfaces that honor this fact need to call attention to their own materiality, their own particular way of representing the world, so that represen- tation does not become something to be wary of, but something to work with. Interfaces optimized solely for clarity and efficiency may actively suppress the kinds of interpretive inquiry through which users could develop critical understanding of AI-mediated knowledge production. Interpretative interfaces should instead fore- ground the constructedness of their own representations, offering an alternative to the dominant chat-based paradigm, whose anthro- pomorphic framing risks obscuring the material and contingent nature of model operations. 4 Designing Interpretative Interfaces Following a two-month ethnographic study of AI-mediated reading practices in scientific disciplines, this project is currently in an exploratory design phase. Our approach is organized around three interrelated design spaces: action (what users can do), representation (how intermediate layers are visualized), and interaction (how users physically manipulate the interface). 4.1 Action: Follow the Token The core action of an interpretative interface is to follow a token. A user selects a token, from an input prompt or a model’s output, and traces its embedding trajectory through the model’s intermediate layers, observing how its position in semantic space shifts as context is processed. These transformations are visualized through dimen- sionality reduction techniques (such as PCA or UMAP), and can be projected into a 2D visualization where users can see, compare, and annotate the trajectories of multiple tokens simultaneously. The result is a form of token-centric reading, where users directly engage with how meaning is produced at the smallest unit (token), layer by layer. This interaction is designed to enable what we call a “glossed model,” drawing on the tradition of medieval manuscript glosses, which were annotations, commentaries, and translations added between lines or in margins to interpret and educate. Similarly, interpretative interfaces allow users to annotate specific token- layer combinations, creating a cumulative record of the traces of their interpretative work with the model’s operations. 4.2 Representation: A Swatchbook of Intermediate Layers Since layer representation for non-experts is an important yet un- derexplored design space, we are developing a “swatchbook” of representation modes. Representation modes include panels, strata, organic shapes, constellations, grids, rays, mind maps, and translu- cent overlays, among others. Each mode supports a different type of sense-making of how layers “behave.” The swatchbook is itself a design research contribution that offers a design vocabulary for in- termediate layers of transformer-based models. By exploring these different representation modes, the swatchbook emphasizes that there is no single “correct” way to see a model’s intermediate op- erations, but instead that the choice of representation shapes to a significant degree what can be noticed, compared, and therefore understood and questioned. 4.3 Interaction: Manipulation as a Modality Interpretative interfaces emphasize physical, tangible interaction metaphors, such as dragging layers with “hand-cursors,” grabbing tokens from a “token bag,” scrubbing through layers as one might scrub through a video timeline, capturing “stills” as they go. These interactions aim to foreground manipulation as a modality of under- standing, therefore complementing more traditional XAI initiatives. 4.4 Implementation and Next Steps The current prototype uses TransformerLens [18] to inspect a small transformer model (GPT-2, 124M parameters, 12 layers), with a Python backend serving embedding data to a browser-based UI. We are currently conducting the design work on two parallel tracks: the first is to work out the interaction flows and representation modes in Figma; the second is to develop the technical backend to extract and reduce embedding trajectories so that they can be visualized. These tracks will ultimately converge in a functional prototype that connects model data to the designed interactions. Following this exploratory phase, we will host participatory de- sign workshops at the University of Washington, bringing together CHI ’26, April 13–17, 2026, Barcelona, SpainBenabdallah scientists, educators, artists, and researchers to test the early proto- types and iterate on their design. We wish to ground discussions around the possibility of supporting critical and interpretative en- gagement with AI tools via the prototypes themselves. They will be used as “objects to think with” [20], so as to scaffold conversations about what LLM interfaces might be. An important goal of this project is to open up the sociotechnical imaginaries [9] of how people interact with language models. 5 Conclusion This paper has argued that interpretation, not explanation, is the primary means through which people develop understanding of complex systems, and that interpretation requires the material and conceptual means to manipulate. In the context of AI for science, this is crucial as scientists increasingly rely on AI-augmented tools to find, read, and cite literature, and therefore build understandings of their field and research space. Interpretative interfaces contribute to the space of explainable AI (XAI) not through explanation but through epistemic actions [10], specifically the manipulation and annotation of intermediate layers. By structuring interaction around token-centric reading (selecting tokens, tracing their trajectories through layers, annotating the transformations that lead to semantic relevance) they offer users a way to engage with LLMs that is fundamentally different from receiving an explanation. As LLMs increasingly mediate the practices of knowledge pro- duction, such as reading, writing, citation, literature review, and publishing, the interfaces through which scholars and scientists engage with these systems become part of the knowledge infrastruc- ture itself. We cannot risk closing the possibilities of critical inquiry and material intervention that would foster the kind of interpreta- tive engagement that is crucial for healthy knowledge ecologies. Designing for interpretation sustains the conditions for critical in- quiry and the kind of material participation that has historically sustained the knowledge commons. Open scholarship depends not only on open access and open data but on open tools, whose ar- chitecture invites intervention and manipulation. Interpretative interfaces are a step toward that goal. References [1] Yves Citton. 2017. Lire, interpréter, actualiser. Editions Amsterdam, Paris. [2]Oege Dijk. 2019. explainerdashboard — explainerdashboard 0.2 documentation. https://explainerdashboard.readthedocs.io/en/latest/ [3] Johanna Drucker. 2020. Visualization and Interpretation. MIT Press. [4]Upol Ehsan, Elizabeth A Watkins, Philipp Wintersberger, Carina Manger, Sunnie S. Y. Kim, Niels Van Berkel, Andreas Riener, and Mark O Riedl. 2024. Human- Centered Explainable AI (HCXAI): Reloading Explainability in the Era of Large Language Models (LLMs). In Extended Abstracts of the CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA) (CHI EA ’24). Association for Computing Machinery, New York, NY, USA, Article 477, 6 pages. doi:10.1145/ 3613905.3636311 [5]Elizabeth L Eisenstein. 2012. The Printing Revolution in Early Modern Europe. Cambridge University Press. [6] PAIR Google. 2025. What-If Tool. https://pair-code.github.io/what-if-tool/ [7] Ivan Illich. 1985. Tools for Conviviality. Marion Boyars, London. [8] Wolfgang Iser. 1978. The Act of Reading : a Theory of Aesthetic response. Johns Hopkins Univ. Press, Baltimore. [9]Sheila Jasanoff and Sang-Hyun Kim. 2015. Dreamscapes of Modernity: Sociotech- nical Imaginaries and the Fabrication of Power. The University of Chicago Press, Chicago ; London. [10]David Kirsh and Paul Maglio. 1994. On Distinguishing Epistemic From Pragmatic Action. Cognitive Science 18, 4 (1994), 513–49. doi:10.1207/s15516709cog1804_1 [11]Alfred Korzybski. 1933. Science and Sanity : an Introduction to non-Aristotelian Systems and General Semantics. The International Non-Aristotelian Library Publishing Company, Lancaster, Pa. [12] Attila Kovari. 2025. Explainable AI chatbots towards XAI ChatGPT: A review. Heliyon 11, 2 (2025), e42077. doi:10.1016/j.heliyon.2025.e42077 [13]Bianka Kowalska and Halina Kwa’snicka. 2025. Unboxing the Black Box: Mecha- nistic Interpretability for Algorithmic Understanding of Neural Networks. ArXiv abs/2511.19265 (2025). https://api.semanticscholar.org/CorpusId:283244330 [14] Keigo Kusumegi, Xinyu Yang, Paul Ginsparg, Mathijs de Vaan, Toby Stuart, and Yian Yin. 2025. Scientific production in the era of large language models. Science 390 6779 (2025), 1240–1243. https://api.semanticscholar.org/CorpusId:284015869 [15]Pantelis Linardatos, Vasilis Papastefanopoulos, and S. Kotsiantis. 2020. Explain- able AI: A Review of Machine Learning Interpretability Methods. Entropy 23 (2020). https://api.semanticscholar.org/CorpusId:229722844 [16]Pierre Lévy. 1993. Les technologies de l’intelligence: l’avenir de la pensée à l’ère informatique. Seuil. [17]Sebastian Musslick, Laura K. Bartlett, Suyog H. Chandramouli, Marina Dubova, Fernand Gobet, Thomas L. Griffiths, Jessica Hullman, Ross D. King, J. Kutz, Christopher G. Lucas, Suhas Mahesh, F. Pestilli, Sabina J. Sloman, and William R Holmes. 2024. Automating the Practice of Science - Opportunities, Challenges, and Implications. ArXiv abs/2409.05890 (2024). https://api.semanticscholar.org/ CorpusId:272550863 [18]Neel Nanda. 2023. TransformerLens Documentation. https://transformerlensorg. github.io/TransformerLens/ [19]Walter J Ong. 2012. Orality and Literacy: the Technologizing of the Word. Rout- ledge. [20]Seymour A Papert. 1993. Mindstorms : Children, Computers, and Powerful Ideas. Basic Books. [21]Tilman Raukur, A. Ho, Stephen Casper, and Dylan Hadfield-Menell. 2022. Toward Transparent AI: A Survey on Interpreting the Inner Structures of Deep Neural Networks. 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML) (2022), 464–483. https://api.semanticscholar.org/CorpusId:251104722 [22]Oscar Skean, Md Rifat Arefin, Yann LeCun, and Ravid Shwartz-Ziv. 2024. Does Representation Matter? Exploring Intermediate Layers in Large Language Models. arXiv:2412.09563 [cs.LG] https://arxiv.org/abs/2412.09563 [23]Oscar Skean, Md Rifat Arefin, Dan Zhao, Niket Patel, Jalal Naghiyev, Yann LeCun, and Ravid Shwartz-Ziv. 2025. Layer by Layer: Uncovering Hidden Representations in Language Models. arXiv:2502.02013 [cs.LG] https://arxiv.org/abs/2502.02013 [24]Lucy Suchman. 2007. Human-machine Reconfigurations : Plans and Situated Actions. Cambridge University Press, Cambridge. [25] Qianli Wang, Tatiana Anikina, Nils Feldhus, Josef Genabith, Leonhard Hennig, and Sebastian Möller. 2024. LLMCheckup: Conversational Examination of Large Language Models via Interpretability Tools and Self-Explanations. In Proceedings of the Third Workshop on Bridging Human–Computer Interaction and Natural Lan- guage Processing, Su Lin Blodgett, Amanda Cercas Curry, Sunipa Dev, Michael Madaio, Ani Nenkova, Diyi Yang, and Ziang Xiao (Eds.). Association for Computa- tional Linguistics, Mexico City, Mexico, 89–104. doi:10.18653/v1/2024.hcinlp-1.9 [26]Sharon Whitfield and Melissa A. Hofmann. 2023. Elicit: AI literature review research assistant. Public Services Quarterly 19 (2023), 201 – 207. https://doi.org/ 10.1080/15228959.2023.2224125 [27] Maryanne Wolf. 2008. Proust and the Squid : the Story and Science of the Reading Brain. HarperCollins.