[{"title":"Generation of Explanatory Content and Requirements for Social XAI","doi":"10.1007/978-981-96-5290-7_15","date_updated":"2026-03-23T08:49:08Z","publisher":"Springer Nature Singapore","author":[{"first_name":"Kary","full_name":"Främling, Kary","last_name":"Främling"},{"full_name":"Thommes, Kirsten","id":"72497","last_name":"Thommes","first_name":"Kirsten"},{"first_name":"Britta","last_name":"Wrede","full_name":"Wrede, Britta"}],"date_created":"2026-03-23T08:48:38Z","year":"2026","place":"Singapore","citation":{"chicago":"Främling, Kary, Kirsten Thommes, and Britta Wrede. “Generation of Explanatory Content and Requirements for Social XAI.” In <i>Social Explainable AI</i>. Singapore: Springer Nature Singapore, 2026. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_15\">https://doi.org/10.1007/978-981-96-5290-7_15</a>.","ieee":"K. Främling, K. Thommes, and B. Wrede, “Generation of Explanatory Content and Requirements for Social XAI,” in <i>Social Explainable AI</i>, Singapore: Springer Nature Singapore, 2026.","ama":"Främling K, Thommes K, Wrede B. Generation of Explanatory Content and Requirements for Social XAI. In: <i>Social Explainable AI</i>. Springer Nature Singapore; 2026. doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_15\">10.1007/978-981-96-5290-7_15</a>","apa":"Främling, K., Thommes, K., &#38; Wrede, B. (2026). Generation of Explanatory Content and Requirements for Social XAI. In <i>Social Explainable AI</i>. Springer Nature Singapore. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_15\">https://doi.org/10.1007/978-981-96-5290-7_15</a>","bibtex":"@inbook{Främling_Thommes_Wrede_2026, place={Singapore}, title={Generation of Explanatory Content and Requirements for Social XAI}, DOI={<a href=\"https://doi.org/10.1007/978-981-96-5290-7_15\">10.1007/978-981-96-5290-7_15</a>}, booktitle={Social Explainable AI}, publisher={Springer Nature Singapore}, author={Främling, Kary and Thommes, Kirsten and Wrede, Britta}, year={2026} }","short":"K. Främling, K. Thommes, B. Wrede, in: Social Explainable AI, Springer Nature Singapore, Singapore, 2026.","mla":"Främling, Kary, et al. “Generation of Explanatory Content and Requirements for Social XAI.” <i>Social Explainable AI</i>, Springer Nature Singapore, 2026, doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_15\">10.1007/978-981-96-5290-7_15</a>."},"publication_identifier":{"isbn":["9789819652891","9789819652907"]},"publication_status":"published","language":[{"iso":"eng"}],"_id":"65090","project":[{"name":"TRR 318 - Subproject A3","_id":"113"},{"_id":"125","name":"TRR 318 - Subproject C2"}],"department":[{"_id":"178"},{"_id":"184"}],"user_id":"72497","abstract":[{"text":"<jats:title>Abstract</jats:title>\r\n                  <jats:p>If XAI are to become social XAI, XAI methods must have capabilities enabling them to ‘extract’ information about the underlying AI model and to generate explanatory content based on that information. In a dialog between explainer and explainee, the explanans presented in every explanation move have to relate to each other understandably and coherently in order to remain trustworthy. This signifies that the generated explanantia have to be consistent—independently of what question is answered by each explanans, in what modality, in what vocabulary, and at what level of abstraction. Moreover, it is advantageous to be able to provide a rich palette of different kinds of explanantia in order to be able to have a fluent dialog in which the explanantia can be generated and adapted to the context, the explainee, feedback, reactions during the interaction with the explainee, and so forth. This chapter attempts to identify relevant questions that an explainee might ask during an explanatory dialog, and it assesses to what extent different XAI methods are capable of addressing these questions in a coherent way. The Contextual Importance and Utility (CIU) method is used to illustrate how an XAI method can generate explanantia for most of the identified questions. CIU also provides a flexibility in how explanatory content is generated that makes it possible to create a meaningful dialog with the explainee.</jats:p>","lang":"eng"}],"status":"public","publication":"Social Explainable AI","type":"book_chapter"},{"year":"2026","place":"Singapore","citation":{"chicago":"Thommes, Kirsten. “Measuring the Outcome of SXAI.” In <i>Social Explainable AI</i>. Singapore: Springer Nature Singapore, 2026. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_28\">https://doi.org/10.1007/978-981-96-5290-7_28</a>.","ieee":"K. Thommes, “Measuring the Outcome of sXAI,” in <i>Social Explainable AI</i>, Singapore: Springer Nature Singapore, 2026.","ama":"Thommes K. Measuring the Outcome of sXAI. In: <i>Social Explainable AI</i>. Springer Nature Singapore; 2026. doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_28\">10.1007/978-981-96-5290-7_28</a>","apa":"Thommes, K. (2026). Measuring the Outcome of sXAI. In <i>Social Explainable AI</i>. Springer Nature Singapore. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_28\">https://doi.org/10.1007/978-981-96-5290-7_28</a>","bibtex":"@inbook{Thommes_2026, place={Singapore}, title={Measuring the Outcome of sXAI}, DOI={<a href=\"https://doi.org/10.1007/978-981-96-5290-7_28\">10.1007/978-981-96-5290-7_28</a>}, booktitle={Social Explainable AI}, publisher={Springer Nature Singapore}, author={Thommes, Kirsten}, year={2026} }","short":"K. Thommes, in: Social Explainable AI, Springer Nature Singapore, Singapore, 2026.","mla":"Thommes, Kirsten. “Measuring the Outcome of SXAI.” <i>Social Explainable AI</i>, Springer Nature Singapore, 2026, doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_28\">10.1007/978-981-96-5290-7_28</a>."},"publication_identifier":{"isbn":["9789819652891","9789819652907"]},"publication_status":"published","title":"Measuring the Outcome of sXAI","doi":"10.1007/978-981-96-5290-7_28","publisher":"Springer Nature Singapore","date_updated":"2026-03-23T08:45:33Z","date_created":"2026-03-23T08:45:05Z","author":[{"first_name":"Kirsten","last_name":"Thommes","id":"72497","full_name":"Thommes, Kirsten"}],"abstract":[{"lang":"eng","text":"<jats:title>Abstract</jats:title>\r\n                  <jats:p>Quantitatively evaluating the benefits of eXplainable Artificial Intelligence (XAI) and social XAI for humans is not a trivial pursuit. Therefore, we categorize the potential measures in terms of subjective and objective outcomes and short- and long-term outcomes of interactive social XAI. When reviewing the current state of the art, we observed some measurement problems in the literature: (a) Researchers do not clearly state whether they want to measure the inner state of users, users’ behavioral response, or the overall AI-human collaborative performance. (b) Moreover, most measures implicitly assume that all humans either do not react or improve in attitudes or performance. Psychological reactance (feeling or doing the opposite) is usually not captured. (c) Many researchers invent their own scale when measuring psychological constructs, thereby jeopardizing the validity of their measures and slowing down progress in the field, because general evidence and subsequent learning can be achieved only by collecting many compatible pieces of evidence. (d) Most studies look into short-term outcomes and neglect that experiences in social interactions with XAI may evolve and have long-term outcomes not only for the individual but also for groups or society at large.</jats:p>"}],"status":"public","publication":"Social Explainable AI","type":"book_chapter","language":[{"iso":"eng"}],"_id":"65088","project":[{"name":"TRR 318 - Subproject A3","_id":"113"},{"_id":"125","name":"TRR 318 - Subproject C2"}],"department":[{"_id":"178"},{"_id":"184"}],"user_id":"72497"},{"publication":"Social Explainable AI","type":"book_chapter","abstract":[{"lang":"eng","text":"<jats:title>Abstract</jats:title>\r\n                  <jats:p>Explainable AI (XAI) aims to make the decisions and behavior of an AI understandable to the people interacting with it and to those affected by its outcomes. To make XAI social, real-world XAI systems need to simulate not only the ways in which human explainers behave within explanatory dialogs but also the ways in which such dialogs can successfully achieve the intended understanding on the explainee’s side. This, in turn, requires an operationalization of the three core aspects of social XAI: multimodality, incrementality, and patterns. This chapter lays the ground for this goal by defining a basic operational model of social interactions that can be refined and extended to account for the specificities of any explanatory real-world setting. This serves as a basis for summarizing and discussing existing ideas from explainability research and related areas in order to operationalize each core aspect. Selected examples and case studies illustrate how to concretely realize such an operationalization, thereby serving as a starting point for future research on social interaction with XAI.</jats:p>"}],"status":"public","_id":"65086","project":[{"name":"TRR 318 - Subproject A3","_id":"113"},{"_id":"125","name":"TRR 318 - Subproject C2"}],"department":[{"_id":"178"},{"_id":"184"}],"user_id":"72497","language":[{"iso":"eng"}],"publication_identifier":{"isbn":["9789819652891","9789819652907"]},"publication_status":"published","year":"2026","place":"Singapore","citation":{"short":"H. Wachsmuth, K. Thommes, M. Alshomary, in: Social Explainable AI, Springer Nature Singapore, Singapore, 2026.","mla":"Wachsmuth, Henning, et al. “Operationalizing Social Interaction.” <i>Social Explainable AI</i>, Springer Nature Singapore, 2026, doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_27\">10.1007/978-981-96-5290-7_27</a>.","bibtex":"@inbook{Wachsmuth_Thommes_Alshomary_2026, place={Singapore}, title={Operationalizing Social Interaction}, DOI={<a href=\"https://doi.org/10.1007/978-981-96-5290-7_27\">10.1007/978-981-96-5290-7_27</a>}, booktitle={Social Explainable AI}, publisher={Springer Nature Singapore}, author={Wachsmuth, Henning and Thommes, Kirsten and Alshomary, Milad}, year={2026} }","apa":"Wachsmuth, H., Thommes, K., &#38; Alshomary, M. (2026). Operationalizing Social Interaction. In <i>Social Explainable AI</i>. Springer Nature Singapore. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_27\">https://doi.org/10.1007/978-981-96-5290-7_27</a>","chicago":"Wachsmuth, Henning, Kirsten Thommes, and Milad Alshomary. “Operationalizing Social Interaction.” In <i>Social Explainable AI</i>. Singapore: Springer Nature Singapore, 2026. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_27\">https://doi.org/10.1007/978-981-96-5290-7_27</a>.","ieee":"H. Wachsmuth, K. Thommes, and M. Alshomary, “Operationalizing Social Interaction,” in <i>Social Explainable AI</i>, Singapore: Springer Nature Singapore, 2026.","ama":"Wachsmuth H, Thommes K, Alshomary M. Operationalizing Social Interaction. In: <i>Social Explainable AI</i>. Springer Nature Singapore; 2026. doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_27\">10.1007/978-981-96-5290-7_27</a>"},"publisher":"Springer Nature Singapore","date_updated":"2026-03-23T08:43:25Z","author":[{"last_name":"Wachsmuth","full_name":"Wachsmuth, Henning","first_name":"Henning"},{"first_name":"Kirsten","id":"72497","full_name":"Thommes, Kirsten","last_name":"Thommes"},{"last_name":"Alshomary","full_name":"Alshomary, Milad","first_name":"Milad"}],"date_created":"2026-03-23T08:42:37Z","title":"Operationalizing Social Interaction","doi":"10.1007/978-981-96-5290-7_27"},{"type":"book_chapter","publication":"Social Explainable AI","abstract":[{"text":"<jats:title>Abstract</jats:title>\r\n                  <jats:p>This chapter examines key challenges and potential improvements in the areas of user interaction and dynamic explanations. It highlights the need for XAI systems to address context factors beyond their predefined scope, it points to the potential need to cocreate new concepts that are adapted to particular explainees, and it provides a clear overview of the XAI system’s underlying knowledge structure and interaction steps. Emphasis is placed on mixed-initiative interaction in which the system can lead or respond based on the context and the explainee’s reactions while asserting the importance of maintaining coherence across consecutive explanations. These advances aim to make XAI systems more flexible, interactive, and user-centric. An operationalization section outlines how such social XAI systems could be implemented based on the XAI capabilities provided by the Contextual Importance and Utility XAI method described in the previous chapter.</jats:p>","lang":"eng"}],"status":"public","project":[{"name":"TRR 318 - Subproject A3","_id":"113"},{"name":"TRR 318 - Subproject C2","_id":"125"}],"_id":"65091","user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"language":[{"iso":"eng"}],"publication_status":"published","publication_identifier":{"isbn":["9789819652891","9789819652907"]},"place":"Singapore","year":"2026","citation":{"ieee":"K. Främling, B. Wrede, and K. Thommes, “Exploration of Explaining Content,” in <i>Social Explainable AI</i>, Singapore: Springer Nature Singapore, 2026.","chicago":"Främling, Kary, Britta Wrede, and Kirsten Thommes. “Exploration of Explaining Content.” In <i>Social Explainable AI</i>. Singapore: Springer Nature Singapore, 2026. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_16\">https://doi.org/10.1007/978-981-96-5290-7_16</a>.","apa":"Främling, K., Wrede, B., &#38; Thommes, K. (2026). Exploration of Explaining Content. In <i>Social Explainable AI</i>. Springer Nature Singapore. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_16\">https://doi.org/10.1007/978-981-96-5290-7_16</a>","ama":"Främling K, Wrede B, Thommes K. Exploration of Explaining Content. In: <i>Social Explainable AI</i>. Springer Nature Singapore; 2026. doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_16\">10.1007/978-981-96-5290-7_16</a>","short":"K. Främling, B. Wrede, K. Thommes, in: Social Explainable AI, Springer Nature Singapore, Singapore, 2026.","bibtex":"@inbook{Främling_Wrede_Thommes_2026, place={Singapore}, title={Exploration of Explaining Content}, DOI={<a href=\"https://doi.org/10.1007/978-981-96-5290-7_16\">10.1007/978-981-96-5290-7_16</a>}, booktitle={Social Explainable AI}, publisher={Springer Nature Singapore}, author={Främling, Kary and Wrede, Britta and Thommes, Kirsten}, year={2026} }","mla":"Främling, Kary, et al. “Exploration of Explaining Content.” <i>Social Explainable AI</i>, Springer Nature Singapore, 2026, doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_16\">10.1007/978-981-96-5290-7_16</a>."},"publisher":"Springer Nature Singapore","date_updated":"2026-03-23T08:52:46Z","author":[{"first_name":"Kary","full_name":"Främling, Kary","last_name":"Främling"},{"full_name":"Wrede, Britta","last_name":"Wrede","first_name":"Britta"},{"first_name":"Kirsten","last_name":"Thommes","full_name":"Thommes, Kirsten","id":"72497"}],"date_created":"2026-03-23T08:52:18Z","title":"Exploration of Explaining Content","doi":"10.1007/978-981-96-5290-7_16"},{"publication":"Social Explainable AI","type":"book_chapter","status":"public","abstract":[{"lang":"eng","text":"<jats:title>Abstract</jats:title>\r\n                  <jats:p>Much research in XAI focuses on single, one-shot interactions, implicitly assuming that interactions have no past, no future, and no surroundings. Although this assumption may be necessary for many empirical research settings, it is overly simplifying and unrealistic. Whereas empirical research focuses on a world in which no social context exists, real applications are embedded in a temporal (past and future) and social context. Social science research shows that repeated interactions and secondhand knowledge in the social space massively affect human attitudes and behaviors. This chapter explains how not only repeated interactions between XAI and humans but also the social space and secondhand information may affect social XAI research.</jats:p>"}],"department":[{"_id":"178"},{"_id":"184"}],"user_id":"72497","_id":"65087","project":[{"_id":"125","name":"TRR 318 - Subproject C2"},{"name":"TRR 318 - Subproject A3","_id":"113"}],"language":[{"iso":"eng"}],"publication_identifier":{"isbn":["9789819652891","9789819652907"]},"publication_status":"published","citation":{"bibtex":"@inbook{Thommes_Främling_Wrede_Kubler_2026, place={Singapore}, title={Interaction History in Social XAI}, DOI={<a href=\"https://doi.org/10.1007/978-981-96-5290-7_17\">10.1007/978-981-96-5290-7_17</a>}, booktitle={Social Explainable AI}, publisher={Springer Nature Singapore}, author={Thommes, Kirsten and Främling, Kary and Wrede, Britta and Kubler, Sylvain}, year={2026} }","short":"K. Thommes, K. Främling, B. Wrede, S. Kubler, in: Social Explainable AI, Springer Nature Singapore, Singapore, 2026.","mla":"Thommes, Kirsten, et al. “Interaction History in Social XAI.” <i>Social Explainable AI</i>, Springer Nature Singapore, 2026, doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_17\">10.1007/978-981-96-5290-7_17</a>.","apa":"Thommes, K., Främling, K., Wrede, B., &#38; Kubler, S. (2026). Interaction History in Social XAI. In <i>Social Explainable AI</i>. Springer Nature Singapore. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_17\">https://doi.org/10.1007/978-981-96-5290-7_17</a>","ama":"Thommes K, Främling K, Wrede B, Kubler S. Interaction History in Social XAI. In: <i>Social Explainable AI</i>. Springer Nature Singapore; 2026. doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_17\">10.1007/978-981-96-5290-7_17</a>","chicago":"Thommes, Kirsten, Kary Främling, Britta Wrede, and Sylvain Kubler. “Interaction History in Social XAI.” In <i>Social Explainable AI</i>. Singapore: Springer Nature Singapore, 2026. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_17\">https://doi.org/10.1007/978-981-96-5290-7_17</a>.","ieee":"K. Thommes, K. Främling, B. Wrede, and S. Kubler, “Interaction History in Social XAI,” in <i>Social Explainable AI</i>, Singapore: Springer Nature Singapore, 2026."},"year":"2026","place":"Singapore","date_created":"2026-03-23T08:43:50Z","author":[{"id":"72497","full_name":"Thommes, Kirsten","last_name":"Thommes","first_name":"Kirsten"},{"full_name":"Främling, Kary","last_name":"Främling","first_name":"Kary"},{"first_name":"Britta","full_name":"Wrede, Britta","last_name":"Wrede"},{"first_name":"Sylvain","last_name":"Kubler","full_name":"Kubler, Sylvain"}],"publisher":"Springer Nature Singapore","date_updated":"2026-03-23T08:44:37Z","doi":"10.1007/978-981-96-5290-7_17","title":"Interaction History in Social XAI"},{"project":[{"name":"TRR 318 - Subproject A3","_id":"113"},{"name":"TRR 318 - Subproject C2","_id":"125"}],"_id":"65089","user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"language":[{"iso":"eng"}],"type":"book_chapter","publication":"Social Explainable AI","abstract":[{"lang":"eng","text":"<jats:title>Abstract</jats:title>\r\n                  <jats:p>In the past, there has been much research aiming to evaluate XAI practices—that is, explanations that can add to a user’s understanding of “why” or “why not.” However, because there is such a huge amount of diversity in social contexts, optimizing for the mean neglects the social dimensions of to whom, what, why, when, and where explanations are provided. Nonetheless, these dimensions matter. We give some brief examples on the accuracy of the mental model (as an example for who?), on measuring explanation practices (as an example of what?), on human motivation (as an example of why?), on repeated interactions (as an example of when), and on bystander effects (as an example of where?). Importantly, controlling for these factors (or randomizing them) is as important as attempting to perform external validations.</jats:p>"}],"status":"public","date_updated":"2026-03-23T08:47:38Z","publisher":"Springer Nature Singapore","author":[{"full_name":"Thommes, Kirsten","id":"72497","last_name":"Thommes","first_name":"Kirsten"}],"date_created":"2026-03-23T08:47:10Z","title":"Evaluation Principles","doi":"10.1007/978-981-96-5290-7_26","publication_status":"published","publication_identifier":{"isbn":["9789819652891","9789819652907"]},"year":"2026","place":"Singapore","citation":{"apa":"Thommes, K. (2026). Evaluation Principles. In <i>Social Explainable AI</i>. Springer Nature Singapore. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_26\">https://doi.org/10.1007/978-981-96-5290-7_26</a>","short":"K. Thommes, in: Social Explainable AI, Springer Nature Singapore, Singapore, 2026.","bibtex":"@inbook{Thommes_2026, place={Singapore}, title={Evaluation Principles}, DOI={<a href=\"https://doi.org/10.1007/978-981-96-5290-7_26\">10.1007/978-981-96-5290-7_26</a>}, booktitle={Social Explainable AI}, publisher={Springer Nature Singapore}, author={Thommes, Kirsten}, year={2026} }","mla":"Thommes, Kirsten. “Evaluation Principles.” <i>Social Explainable AI</i>, Springer Nature Singapore, 2026, doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_26\">10.1007/978-981-96-5290-7_26</a>.","chicago":"Thommes, Kirsten. “Evaluation Principles.” In <i>Social Explainable AI</i>. Singapore: Springer Nature Singapore, 2026. <a href=\"https://doi.org/10.1007/978-981-96-5290-7_26\">https://doi.org/10.1007/978-981-96-5290-7_26</a>.","ieee":"K. Thommes, “Evaluation Principles,” in <i>Social Explainable AI</i>, Singapore: Springer Nature Singapore, 2026.","ama":"Thommes K. Evaluation Principles. In: <i>Social Explainable AI</i>. Springer Nature Singapore; 2026. doi:<a href=\"https://doi.org/10.1007/978-981-96-5290-7_26\">10.1007/978-981-96-5290-7_26</a>"}},{"language":[{"iso":"eng"}],"user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"project":[{"_id":"125","name":"TRR 318 - C2: TRR 318 - Subproject C2"}],"_id":"58939","status":"public","type":"journal_article","publication":"Plos One","doi":"10.1371/journal.pone.0318874","title":"Algorithm, expert, or both? Evaluating the role of feature selection methods on user preferences and reliance","date_created":"2025-03-08T12:09:50Z","author":[{"full_name":"Kornowicz, Jaroslaw","id":"44029","orcid":"0000-0002-5654-9911","last_name":"Kornowicz","first_name":"Jaroslaw"},{"first_name":"Kirsten","last_name":"Thommes","id":"72497","full_name":"Thommes, Kirsten"}],"date_updated":"2025-03-14T19:24:56Z","citation":{"ama":"Kornowicz J, Thommes K. Algorithm, expert, or both? Evaluating the role of feature selection methods on user preferences and reliance. <i>Plos One</i>. Published online 2025. doi:<a href=\"https://doi.org/10.1371/journal.pone.0318874\">10.1371/journal.pone.0318874</a>","ieee":"J. Kornowicz and K. Thommes, “Algorithm, expert, or both? Evaluating the role of feature selection methods on user preferences and reliance,” <i>Plos One</i>, 2025, doi: <a href=\"https://doi.org/10.1371/journal.pone.0318874\">10.1371/journal.pone.0318874</a>.","chicago":"Kornowicz, Jaroslaw, and Kirsten Thommes. “Algorithm, Expert, or Both? Evaluating the Role of Feature Selection Methods on User Preferences and Reliance.” <i>Plos One</i>, 2025. <a href=\"https://doi.org/10.1371/journal.pone.0318874\">https://doi.org/10.1371/journal.pone.0318874</a>.","apa":"Kornowicz, J., &#38; Thommes, K. (2025). Algorithm, expert, or both? Evaluating the role of feature selection methods on user preferences and reliance. <i>Plos One</i>. <a href=\"https://doi.org/10.1371/journal.pone.0318874\">https://doi.org/10.1371/journal.pone.0318874</a>","bibtex":"@article{Kornowicz_Thommes_2025, title={Algorithm, expert, or both? Evaluating the role of feature selection methods on user preferences and reliance}, DOI={<a href=\"https://doi.org/10.1371/journal.pone.0318874\">10.1371/journal.pone.0318874</a>}, journal={Plos One}, author={Kornowicz, Jaroslaw and Thommes, Kirsten}, year={2025} }","short":"J. Kornowicz, K. Thommes, Plos One (2025).","mla":"Kornowicz, Jaroslaw, and Kirsten Thommes. “Algorithm, Expert, or Both? Evaluating the Role of Feature Selection Methods on User Preferences and Reliance.” <i>Plos One</i>, 2025, doi:<a href=\"https://doi.org/10.1371/journal.pone.0318874\">10.1371/journal.pone.0318874</a>."},"year":"2025"},{"type":"journal_article","publication":"Arxiv","status":"public","abstract":[{"text":"Prior research shows that social norms can reduce algorithm aversion, but little is known about how such norms become established. Most accounts emphasize technological and individual determinants, yet AI adoption unfolds within organizational social contexts shaped by peers and supervisors. We ask whether the source of the norm-peers or supervisors-shapes AI usage behavior. This question is practically relevant for organizations seeking to promote effective AI adoption. We conducted an online vignette experiment, complemented by qualitative data on participants' feelings and justifications after (counter-)normative behavior. In line with the theory, counter-normative choices elicited higher regret than norm-adherent choices. On average, choosing AI increased regret compared to choosing an human. This aversion was weaker when AI use was presented as the prevailing norm, indicating a statistically significant interaction between AI use and an AI-favoring norm. Participants also attributed less blame to technology than to humans, which increased regret when AI was chosen over human expertise. Both peer and supervisor influence emerged as relevant factors, though contrary to expectations they did not significantly affect regret. Our findings suggest that regret aversion, embedded in social norms, is a central mechanism driving imitation in AI-related decision-making.","lang":"eng"}],"user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"project":[{"name":"TRR 318 - Subproject C2","_id":"125"}],"_id":"61137","language":[{"iso":"eng"}],"citation":{"chicago":"Kornowicz, Jaroslaw, Maurice Pape, and Kirsten Thommes. “Would I Regret Being Different? The Influence of Social Norms on Attitudes toward AI Usage.” <i>Arxiv</i>, 2025. <a href=\"https://doi.org/10.48550/ARXIV.2509.04241\">https://doi.org/10.48550/ARXIV.2509.04241</a>.","ieee":"J. Kornowicz, M. Pape, and K. Thommes, “Would I regret being different? The influence of social norms on attitudes toward AI usage,” <i>Arxiv</i>, 2025, doi: <a href=\"https://doi.org/10.48550/ARXIV.2509.04241\">10.48550/ARXIV.2509.04241</a>.","ama":"Kornowicz J, Pape M, Thommes K. Would I regret being different? The influence of social norms on attitudes toward AI usage. <i>Arxiv</i>. Published online 2025. doi:<a href=\"https://doi.org/10.48550/ARXIV.2509.04241\">10.48550/ARXIV.2509.04241</a>","short":"J. Kornowicz, M. Pape, K. Thommes, Arxiv (2025).","bibtex":"@article{Kornowicz_Pape_Thommes_2025, title={Would I regret being different? The influence of social norms on attitudes toward AI usage}, DOI={<a href=\"https://doi.org/10.48550/ARXIV.2509.04241\">10.48550/ARXIV.2509.04241</a>}, journal={Arxiv}, author={Kornowicz, Jaroslaw and Pape, Maurice and Thommes, Kirsten}, year={2025} }","mla":"Kornowicz, Jaroslaw, et al. “Would I Regret Being Different? The Influence of Social Norms on Attitudes toward AI Usage.” <i>Arxiv</i>, 2025, doi:<a href=\"https://doi.org/10.48550/ARXIV.2509.04241\">10.48550/ARXIV.2509.04241</a>.","apa":"Kornowicz, J., Pape, M., &#38; Thommes, K. (2025). Would I regret being different? The influence of social norms on attitudes toward AI usage. <i>Arxiv</i>. <a href=\"https://doi.org/10.48550/ARXIV.2509.04241\">https://doi.org/10.48550/ARXIV.2509.04241</a>"},"year":"2025","date_created":"2025-09-05T06:42:48Z","author":[{"first_name":"Jaroslaw","full_name":"Kornowicz, Jaroslaw","id":"44029","orcid":"0000-0002-5654-9911","last_name":"Kornowicz"},{"last_name":"Pape","full_name":"Pape, Maurice","first_name":"Maurice"},{"first_name":"Kirsten","id":"72497","full_name":"Thommes, Kirsten","last_name":"Thommes"}],"date_updated":"2025-09-05T06:45:01Z","doi":"10.48550/ARXIV.2509.04241","title":"Would I regret being different? The influence of social norms on attitudes toward AI usage"},{"citation":{"apa":"Heid, S., Kornowicz, J., Hanselle, J., Thommes, K., &#38; Hüllermeier, E. (2025). MSL: Multi-class Scoring Lists for Interpretable Incremental Decision-Making. In <i>Communications in Computer and Information Science</i>. Springer Nature Switzerland. <a href=\"https://doi.org/10.1007/978-3-032-08327-2_6\">https://doi.org/10.1007/978-3-032-08327-2_6</a>","short":"S. Heid, J. Kornowicz, J. Hanselle, K. Thommes, E. Hüllermeier, in: Communications in Computer and Information Science, Springer Nature Switzerland, Cham, 2025.","mla":"Heid, Stefan, et al. “MSL: Multi-Class Scoring Lists for Interpretable Incremental Decision-Making.” <i>Communications in Computer and Information Science</i>, Springer Nature Switzerland, 2025, doi:<a href=\"https://doi.org/10.1007/978-3-032-08327-2_6\">10.1007/978-3-032-08327-2_6</a>.","bibtex":"@inbook{Heid_Kornowicz_Hanselle_Thommes_Hüllermeier_2025, place={Cham}, title={MSL: Multi-class Scoring Lists for Interpretable Incremental Decision-Making}, DOI={<a href=\"https://doi.org/10.1007/978-3-032-08327-2_6\">10.1007/978-3-032-08327-2_6</a>}, booktitle={Communications in Computer and Information Science}, publisher={Springer Nature Switzerland}, author={Heid, Stefan and Kornowicz, Jaroslaw and Hanselle, Jonas and Thommes, Kirsten and Hüllermeier, Eyke}, year={2025} }","chicago":"Heid, Stefan, Jaroslaw Kornowicz, Jonas Hanselle, Kirsten Thommes, and Eyke Hüllermeier. “MSL: Multi-Class Scoring Lists for Interpretable Incremental Decision-Making.” In <i>Communications in Computer and Information Science</i>. Cham: Springer Nature Switzerland, 2025. <a href=\"https://doi.org/10.1007/978-3-032-08327-2_6\">https://doi.org/10.1007/978-3-032-08327-2_6</a>.","ieee":"S. Heid, J. Kornowicz, J. Hanselle, K. Thommes, and E. Hüllermeier, “MSL: Multi-class Scoring Lists for Interpretable Incremental Decision-Making,” in <i>Communications in Computer and Information Science</i>, Cham: Springer Nature Switzerland, 2025.","ama":"Heid S, Kornowicz J, Hanselle J, Thommes K, Hüllermeier E. MSL: Multi-class Scoring Lists for Interpretable Incremental Decision-Making. In: <i>Communications in Computer and Information Science</i>. Springer Nature Switzerland; 2025. doi:<a href=\"https://doi.org/10.1007/978-3-032-08327-2_6\">10.1007/978-3-032-08327-2_6</a>"},"year":"2025","place":"Cham","publication_status":"published","publication_identifier":{"isbn":["9783032083265","9783032083272"],"issn":["1865-0929","1865-0937"]},"doi":"10.1007/978-3-032-08327-2_6","title":"MSL: Multi-class Scoring Lists for Interpretable Incremental Decision-Making","date_created":"2025-10-13T13:34:36Z","author":[{"id":"39640","full_name":"Heid, Stefan","orcid":"0000-0002-9461-7372","last_name":"Heid","first_name":"Stefan"},{"first_name":"Jaroslaw","orcid":"0000-0002-5654-9911","last_name":"Kornowicz","full_name":"Kornowicz, Jaroslaw","id":"44029"},{"last_name":"Hanselle","full_name":"Hanselle, Jonas","first_name":"Jonas"},{"id":"72497","full_name":"Thommes, Kirsten","last_name":"Thommes","first_name":"Kirsten"},{"first_name":"Eyke","last_name":"Hüllermeier","full_name":"Hüllermeier, Eyke","id":"48129"}],"publisher":"Springer Nature Switzerland","date_updated":"2025-10-13T13:35:11Z","status":"public","abstract":[{"text":"<jats:title>Abstract</jats:title>\r\n          <jats:p>A scoring list is a sequence of simple decision models, where features are incrementally evaluated and scores of satisfied features are summed to be used for threshold-based decisions or for calculating class probabilities. In this paper, we introduce a new multi-class variant and compare it against previously introduced binary classification variants for incremental decisions, as well as multi-class variants for classical decision-making using all features. Furthermore, we introduce a new multi-class dataset to assess collaborative human-machine decision-making, which is suitable for user studies with non-expert participants. We demonstrate the usefulness of our approach by evaluating predictive performance and compared to the performance of participants without AI help.</jats:p>","lang":"eng"}],"type":"book_chapter","publication":"Communications in Computer and Information Science","language":[{"iso":"eng"}],"user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"project":[{"_id":"125","name":"TRR 318 - Subproject C2"}],"_id":"61820"},{"department":[{"_id":"178"},{"_id":"184"}],"user_id":"72497","_id":"62213","project":[{"_id":"125","name":"TRR 318 - Subproject C2"}],"language":[{"iso":"eng"}],"publication":"International Journal of Human–Computer Interaction","type":"journal_article","status":"public","date_created":"2025-11-17T12:27:42Z","author":[{"full_name":"Kornowicz, Jaroslaw","id":"44029","last_name":"Kornowicz","orcid":"0000-0002-5654-9911","first_name":"Jaroslaw"}],"publisher":"Informa UK Limited","date_updated":"2025-11-17T12:28:12Z","doi":"10.1080/10447318.2025.2581260","title":"An Empirical Examination of the Evaluative AI Framework","publication_identifier":{"issn":["1044-7318","1532-7590"]},"publication_status":"published","page":"1-19","citation":{"apa":"Kornowicz, J. (2025). An Empirical Examination of the Evaluative AI Framework. <i>International Journal of Human–Computer Interaction</i>, 1–19. <a href=\"https://doi.org/10.1080/10447318.2025.2581260\">https://doi.org/10.1080/10447318.2025.2581260</a>","mla":"Kornowicz, Jaroslaw. “An Empirical Examination of the Evaluative AI Framework.” <i>International Journal of Human–Computer Interaction</i>, Informa UK Limited, 2025, pp. 1–19, doi:<a href=\"https://doi.org/10.1080/10447318.2025.2581260\">10.1080/10447318.2025.2581260</a>.","short":"J. Kornowicz, International Journal of Human–Computer Interaction (2025) 1–19.","bibtex":"@article{Kornowicz_2025, title={An Empirical Examination of the Evaluative AI Framework}, DOI={<a href=\"https://doi.org/10.1080/10447318.2025.2581260\">10.1080/10447318.2025.2581260</a>}, journal={International Journal of Human–Computer Interaction}, publisher={Informa UK Limited}, author={Kornowicz, Jaroslaw}, year={2025}, pages={1–19} }","ama":"Kornowicz J. An Empirical Examination of the Evaluative AI Framework. <i>International Journal of Human–Computer Interaction</i>. Published online 2025:1-19. doi:<a href=\"https://doi.org/10.1080/10447318.2025.2581260\">10.1080/10447318.2025.2581260</a>","ieee":"J. Kornowicz, “An Empirical Examination of the Evaluative AI Framework,” <i>International Journal of Human–Computer Interaction</i>, pp. 1–19, 2025, doi: <a href=\"https://doi.org/10.1080/10447318.2025.2581260\">10.1080/10447318.2025.2581260</a>.","chicago":"Kornowicz, Jaroslaw. “An Empirical Examination of the Evaluative AI Framework.” <i>International Journal of Human–Computer Interaction</i>, 2025, 1–19. <a href=\"https://doi.org/10.1080/10447318.2025.2581260\">https://doi.org/10.1080/10447318.2025.2581260</a>."},"year":"2025"},{"title":"Learning decision catalogues for situated decision making: The case of scoring systems","doi":"10.1016/j.ijar.2024.109190","publisher":"Elsevier BV","date_updated":"2024-06-26T14:28:47Z","volume":171,"date_created":"2024-06-26T14:28:07Z","author":[{"first_name":"Stefan","last_name":"Heid","orcid":"0000-0002-9461-7372","id":"39640","full_name":"Heid, Stefan"},{"orcid":"0000-0002-1231-4985","last_name":"Hanselle","id":"43980","full_name":"Hanselle, Jonas Manuel","first_name":"Jonas Manuel"},{"first_name":"Johannes","full_name":"Fürnkranz, Johannes","last_name":"Fürnkranz"},{"first_name":"Eyke","full_name":"Hüllermeier, Eyke","id":"48129","last_name":"Hüllermeier"}],"year":"2024","intvolume":"       171","citation":{"ama":"Heid S, Hanselle JM, Fürnkranz J, Hüllermeier E. Learning decision catalogues for situated decision making: The case of scoring systems. <i>International Journal of Approximate Reasoning</i>. 2024;171. doi:<a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>","chicago":"Heid, Stefan, Jonas Manuel Hanselle, Johannes Fürnkranz, and Eyke Hüllermeier. “Learning Decision Catalogues for Situated Decision Making: The Case of Scoring Systems.” <i>International Journal of Approximate Reasoning</i> 171 (2024). <a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">https://doi.org/10.1016/j.ijar.2024.109190</a>.","ieee":"S. Heid, J. M. Hanselle, J. Fürnkranz, and E. Hüllermeier, “Learning decision catalogues for situated decision making: The case of scoring systems,” <i>International Journal of Approximate Reasoning</i>, vol. 171, Art. no. 109190, 2024, doi: <a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>.","bibtex":"@article{Heid_Hanselle_Fürnkranz_Hüllermeier_2024, title={Learning decision catalogues for situated decision making: The case of scoring systems}, volume={171}, DOI={<a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>}, number={109190}, journal={International Journal of Approximate Reasoning}, publisher={Elsevier BV}, author={Heid, Stefan and Hanselle, Jonas Manuel and Fürnkranz, Johannes and Hüllermeier, Eyke}, year={2024} }","mla":"Heid, Stefan, et al. “Learning Decision Catalogues for Situated Decision Making: The Case of Scoring Systems.” <i>International Journal of Approximate Reasoning</i>, vol. 171, 109190, Elsevier BV, 2024, doi:<a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>.","short":"S. Heid, J.M. Hanselle, J. Fürnkranz, E. Hüllermeier, International Journal of Approximate Reasoning 171 (2024).","apa":"Heid, S., Hanselle, J. M., Fürnkranz, J., &#38; Hüllermeier, E. (2024). Learning decision catalogues for situated decision making: The case of scoring systems. <i>International Journal of Approximate Reasoning</i>, <i>171</i>, Article 109190. <a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">https://doi.org/10.1016/j.ijar.2024.109190</a>"},"publication_identifier":{"issn":["0888-613X"]},"publication_status":"published","article_number":"109190","language":[{"iso":"eng"}],"_id":"54911","project":[{"_id":"125","name":"TRR 318 - C2: TRR 318 - Subproject C2"}],"department":[{"_id":"660"}],"user_id":"72497","status":"public","publication":"International Journal of Approximate Reasoning","type":"journal_article"},{"publisher":"Elsevier BV","date_updated":"2024-11-20T09:44:54Z","volume":171,"author":[{"first_name":"Stefan","last_name":"Heid","orcid":"0000-0002-9461-7372","id":"39640","full_name":"Heid, Stefan"},{"first_name":"Jonas Manuel","last_name":"Hanselle","orcid":"0000-0002-1231-4985","id":"43980","full_name":"Hanselle, Jonas Manuel"},{"first_name":"Johannes","last_name":"Fürnkranz","full_name":"Fürnkranz, Johannes"},{"last_name":"Hüllermeier","full_name":"Hüllermeier, Eyke","id":"48129","first_name":"Eyke"}],"date_created":"2024-06-26T14:19:19Z","title":"Learning decision catalogues for situated decision making: The case of scoring systems","doi":"10.1016/j.ijar.2024.109190","publication_identifier":{"issn":["0888-613X"]},"publication_status":"published","year":"2024","intvolume":"       171","citation":{"apa":"Heid, S., Hanselle, J. M., Fürnkranz, J., &#38; Hüllermeier, E. (2024). Learning decision catalogues for situated decision making: The case of scoring systems. <i>International Journal of Approximate Reasoning</i>, <i>171</i>, Article 109190. <a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">https://doi.org/10.1016/j.ijar.2024.109190</a>","bibtex":"@article{Heid_Hanselle_Fürnkranz_Hüllermeier_2024, title={Learning decision catalogues for situated decision making: The case of scoring systems}, volume={171}, DOI={<a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>}, number={109190}, journal={International Journal of Approximate Reasoning}, publisher={Elsevier BV}, author={Heid, Stefan and Hanselle, Jonas Manuel and Fürnkranz, Johannes and Hüllermeier, Eyke}, year={2024} }","mla":"Heid, Stefan, et al. “Learning Decision Catalogues for Situated Decision Making: The Case of Scoring Systems.” <i>International Journal of Approximate Reasoning</i>, vol. 171, 109190, Elsevier BV, 2024, doi:<a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>.","short":"S. Heid, J.M. Hanselle, J. Fürnkranz, E. Hüllermeier, International Journal of Approximate Reasoning 171 (2024).","ieee":"S. Heid, J. M. Hanselle, J. Fürnkranz, and E. Hüllermeier, “Learning decision catalogues for situated decision making: The case of scoring systems,” <i>International Journal of Approximate Reasoning</i>, vol. 171, Art. no. 109190, 2024, doi: <a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>.","chicago":"Heid, Stefan, Jonas Manuel Hanselle, Johannes Fürnkranz, and Eyke Hüllermeier. “Learning Decision Catalogues for Situated Decision Making: The Case of Scoring Systems.” <i>International Journal of Approximate Reasoning</i> 171 (2024). <a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">https://doi.org/10.1016/j.ijar.2024.109190</a>.","ama":"Heid S, Hanselle JM, Fürnkranz J, Hüllermeier E. Learning decision catalogues for situated decision making: The case of scoring systems. <i>International Journal of Approximate Reasoning</i>. 2024;171. doi:<a href=\"https://doi.org/10.1016/j.ijar.2024.109190\">10.1016/j.ijar.2024.109190</a>"},"_id":"54907","project":[{"_id":"125","name":"TRR 318 - C2: TRR 318 - Subproject C2"}],"department":[{"_id":"660"}],"user_id":"72497","article_number":"109190","language":[{"iso":"eng"}],"publication":"International Journal of Approximate Reasoning","type":"journal_article","status":"public"},{"user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"project":[{"name":"TRR 318 - C2: TRR 318 - Subproject C2","_id":"125"}],"_id":"57645","language":[{"iso":"eng"}],"type":"conference","publication":"PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE","status":"public","author":[{"first_name":"Stefan","full_name":"Heid, Stefan","id":"39640","orcid":"0000-0002-9461-7372","last_name":"Heid"},{"last_name":"Kornowicz","orcid":"0000-0002-5654-9911","full_name":"Kornowicz, Jaroslaw","id":"44029","first_name":"Jaroslaw"},{"full_name":"Hanselle, Jonas Manuel","id":"43980","orcid":"0000-0002-1231-4985","last_name":"Hanselle","first_name":"Jonas Manuel"},{"last_name":"Hüllermeier","id":"48129","full_name":"Hüllermeier, Eyke","first_name":"Eyke"},{"first_name":"Kirsten","full_name":"Thommes, Kirsten","id":"72497","last_name":"Thommes"}],"date_created":"2024-12-09T08:05:54Z","volume":21,"date_updated":"2024-12-09T08:06:37Z","title":"Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems","citation":{"ieee":"S. Heid, J. Kornowicz, J. M. Hanselle, E. Hüllermeier, and K. Thommes, “Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems,” in <i>PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE</i>, 2024, vol. 21, p. 233.","chicago":"Heid, Stefan, Jaroslaw Kornowicz, Jonas Manuel Hanselle, Eyke Hüllermeier, and Kirsten Thommes. “Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems.” In <i>PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE</i>, 21:233, 2024.","ama":"Heid S, Kornowicz J, Hanselle JM, Hüllermeier E, Thommes K. Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems. In: <i>PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE</i>. Vol 21. ; 2024:233.","mla":"Heid, Stefan, et al. “Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems.” <i>PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE</i>, vol. 21, 2024, p. 233.","bibtex":"@inproceedings{Heid_Kornowicz_Hanselle_Hüllermeier_Thommes_2024, title={Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems}, volume={21}, booktitle={PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE}, author={Heid, Stefan and Kornowicz, Jaroslaw and Hanselle, Jonas Manuel and Hüllermeier, Eyke and Thommes, Kirsten}, year={2024}, pages={233} }","short":"S. Heid, J. Kornowicz, J.M. Hanselle, E. Hüllermeier, K. Thommes, in: PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE, 2024, p. 233.","apa":"Heid, S., Kornowicz, J., Hanselle, J. M., Hüllermeier, E., &#38; Thommes, K. (2024). Human-AI Co-Construction of Interpretable Predictive Models: The Case of Scoring Systems. <i>PROCEEDINGS 34. WORKSHOP COMPUTATIONAL INTELLIGENCE</i>, <i>21</i>, 233."},"intvolume":"        21","page":"233","year":"2024"},{"year":"2024","quality_controlled":"1","title":"Towards a Computational Architecture for Co-Constructive Explainable Systems","date_created":"2024-07-26T11:57:31Z","publisher":"ACM","abstract":[{"text":"In this paper we consider the interactive processes by which an explainer and an explainee cooperate to produce an explanation, which we refer to as co-construction. Explainable Artificial Intelligence (XAI) is concerned with the development of intelligent systems and robots that can explain and justify their actions, decisions, recommendations, and so on. However, the cooperative construction of explanations remains a key but under-explored issue. This short paper proposes an architecture for intelligent systems that promotes a co-constructive and interactive approach to explanation generation. By outlining its basic components and their specific roles, we aim to contribute to the advancement of XAI computational frameworks that actively engage users in the explanation process.","lang":"eng"}],"publication":"Proceedings of the 2024 Workshop on Explainability Engineering","language":[{"iso":"eng"}],"page":"20-25","citation":{"ieee":"H. Buschmeier <i>et al.</i>, “Towards a Computational Architecture for Co-Constructive Explainable Systems,” in <i>Proceedings of the 2024 Workshop on Explainability Engineering</i>, Lisbon, Portugal, 2024, pp. 20–25, doi: <a href=\"https://doi.org/10.1145/3648505.3648509\">10.1145/3648505.3648509</a>.","chicago":"Buschmeier, Hendrik, Philipp Cimiano, Stefan Kopp, Jaroslaw Kornowicz, Olesja Lammert, Marco Matarese, Dimitry Mindlin, et al. “Towards a Computational Architecture for Co-Constructive Explainable Systems.” In <i>Proceedings of the 2024 Workshop on Explainability Engineering</i>, 20–25. ACM, 2024. <a href=\"https://doi.org/10.1145/3648505.3648509\">https://doi.org/10.1145/3648505.3648509</a>.","ama":"Buschmeier H, Cimiano P, Kopp S, et al. Towards a Computational Architecture for Co-Constructive Explainable Systems. In: <i>Proceedings of the 2024 Workshop on Explainability Engineering</i>. ACM; 2024:20-25. doi:<a href=\"https://doi.org/10.1145/3648505.3648509\">10.1145/3648505.3648509</a>","mla":"Buschmeier, Hendrik, et al. “Towards a Computational Architecture for Co-Constructive Explainable Systems.” <i>Proceedings of the 2024 Workshop on Explainability Engineering</i>, ACM, 2024, pp. 20–25, doi:<a href=\"https://doi.org/10.1145/3648505.3648509\">10.1145/3648505.3648509</a>.","short":"H. Buschmeier, P. Cimiano, S. Kopp, J. Kornowicz, O. Lammert, M. Matarese, D. Mindlin, A.S. Robrecht, A.-L. Vollmer, P. Wagner, B. Wrede, M. Booshehri, in: Proceedings of the 2024 Workshop on Explainability Engineering, ACM, 2024, pp. 20–25.","bibtex":"@inproceedings{Buschmeier_Cimiano_Kopp_Kornowicz_Lammert_Matarese_Mindlin_Robrecht_Vollmer_Wagner_et al._2024, title={Towards a Computational Architecture for Co-Constructive Explainable Systems}, DOI={<a href=\"https://doi.org/10.1145/3648505.3648509\">10.1145/3648505.3648509</a>}, booktitle={Proceedings of the 2024 Workshop on Explainability Engineering}, publisher={ACM}, author={Buschmeier, Hendrik and Cimiano, Philipp and Kopp, Stefan and Kornowicz, Jaroslaw and Lammert, Olesja and Matarese, Marco and Mindlin, Dimitry and Robrecht, Amelie Sophie and Vollmer, Anna-Lisa and Wagner, Petra and et al.}, year={2024}, pages={20–25} }","apa":"Buschmeier, H., Cimiano, P., Kopp, S., Kornowicz, J., Lammert, O., Matarese, M., Mindlin, D., Robrecht, A. S., Vollmer, A.-L., Wagner, P., Wrede, B., &#38; Booshehri, M. (2024). Towards a Computational Architecture for Co-Constructive Explainable Systems. <i>Proceedings of the 2024 Workshop on Explainability Engineering</i>, 20–25. <a href=\"https://doi.org/10.1145/3648505.3648509\">https://doi.org/10.1145/3648505.3648509</a>"},"publication_status":"published","doi":"10.1145/3648505.3648509","conference":{"name":"2024 Workshop on Explainability Engineering","location":"Lisbon, Portugal"},"main_file_link":[{"open_access":"1"}],"author":[{"first_name":"Hendrik","full_name":"Buschmeier, Hendrik","id":"76456","orcid":"0000-0002-9613-5713","last_name":"Buschmeier"},{"last_name":"Cimiano","full_name":"Cimiano, Philipp","first_name":"Philipp"},{"last_name":"Kopp","full_name":"Kopp, Stefan","first_name":"Stefan"},{"first_name":"Jaroslaw","id":"44029","full_name":"Kornowicz, Jaroslaw","last_name":"Kornowicz","orcid":"0000-0002-5654-9911"},{"orcid":"0000-0001-8201-5166","last_name":"Lammert","id":"47384","full_name":"Lammert, Olesja","first_name":"Olesja"},{"first_name":"Marco","full_name":"Matarese, Marco","last_name":"Matarese"},{"first_name":"Dimitry","last_name":"Mindlin","full_name":"Mindlin, Dimitry"},{"last_name":"Robrecht","full_name":"Robrecht, Amelie Sophie","first_name":"Amelie Sophie"},{"first_name":"Anna-Lisa","full_name":"Vollmer, Anna-Lisa","id":"86589","last_name":"Vollmer"},{"last_name":"Wagner","full_name":"Wagner, Petra","id":"74505","first_name":"Petra"},{"first_name":"Britta","full_name":"Wrede, Britta","last_name":"Wrede"},{"first_name":"Meisam","id":"93424","full_name":"Booshehri, Meisam","last_name":"Booshehri"}],"oa":"1","date_updated":"2025-03-14T19:24:35Z","status":"public","type":"conference","department":[{"_id":"184"},{"_id":"178"},{"_id":"660"}],"user_id":"44029","_id":"55403","project":[{"name":"TRR 318 - C2: TRR 318 - Subproject C2","_id":"125"},{"name":"TRR 318 - A3: TRR 318 - Subproject A3","_id":"113"},{"name":"TRR 318 - A02: TRR 318 - Verstehensprozess einer Erklärung beobachten und auswerten (Teilprojekt A02)","_id":"112","grant_number":"438445824"},{"name":"TRR 318 - INF: TRR 318 - Project Area INF","_id":"118"}]},{"title":"An Empirical Examination of the Evaluative AI Framework","doi":"10.48550/ARXIV.2411.08583","date_updated":"2025-09-15T12:41:53Z","author":[{"id":"44029","full_name":"Kornowicz, Jaroslaw","orcid":"0000-0002-5654-9911","last_name":"Kornowicz","first_name":"Jaroslaw"}],"date_created":"2024-11-27T16:20:25Z","year":"2024","citation":{"ieee":"J. Kornowicz, “An Empirical Examination of the Evaluative AI Framework,” <i>arXiv</i>, 2024, doi: <a href=\"https://doi.org/10.48550/ARXIV.2411.08583\">10.48550/ARXIV.2411.08583</a>.","chicago":"Kornowicz, Jaroslaw. “An Empirical Examination of the Evaluative AI Framework.” <i>ArXiv</i>, 2024. <a href=\"https://doi.org/10.48550/ARXIV.2411.08583\">https://doi.org/10.48550/ARXIV.2411.08583</a>.","ama":"Kornowicz J. An Empirical Examination of the Evaluative AI Framework. <i>arXiv</i>. Published online 2024. doi:<a href=\"https://doi.org/10.48550/ARXIV.2411.08583\">10.48550/ARXIV.2411.08583</a>","mla":"Kornowicz, Jaroslaw. “An Empirical Examination of the Evaluative AI Framework.” <i>ArXiv</i>, 2024, doi:<a href=\"https://doi.org/10.48550/ARXIV.2411.08583\">10.48550/ARXIV.2411.08583</a>.","bibtex":"@article{Kornowicz_2024, title={An Empirical Examination of the Evaluative AI Framework}, DOI={<a href=\"https://doi.org/10.48550/ARXIV.2411.08583\">10.48550/ARXIV.2411.08583</a>}, journal={arXiv}, author={Kornowicz, Jaroslaw}, year={2024} }","short":"J. Kornowicz, ArXiv (2024).","apa":"Kornowicz, J. (2024). An Empirical Examination of the Evaluative AI Framework. <i>ArXiv</i>. <a href=\"https://doi.org/10.48550/ARXIV.2411.08583\">https://doi.org/10.48550/ARXIV.2411.08583</a>"},"article_type":"original","language":[{"iso":"eng"}],"project":[{"_id":"125","name":"TRR 318 - C2: TRR 318 - Subproject C2"}],"_id":"57461","user_id":"44029","department":[{"_id":"184"},{"_id":"178"}],"abstract":[{"lang":"eng","text":"This study empirically examines the \"Evaluative AI\" framework, which aims to enhance the decision-making process for AI users by transitioning from a recommendation-based approach to a hypothesis-driven one. Rather than offering direct recommendations, this framework presents users pro and con evidence for hypotheses to support more informed decisions. However, findings from the current behavioral experiment reveal no significant improvement in decision-making performance and limited user engagement with the evidence provided, resulting in cognitive processes similar to those observed in traditional AI systems. Despite these results, the framework still holds promise for further exploration in future research.\r\n"}],"status":"public","type":"journal_article","publication":"arXiv"},{"type":"conference","publication":"Artificial Intelligence in HCI","status":"public","user_id":"44648","series_title":"Artificial Intelligence in HCI","department":[{"_id":"178"},{"_id":"184"}],"project":[{"name":"TRR 318 - A3: TRR 318 - Subproject A3","_id":"113"},{"name":"TRR 318 - C2: TRR 318 - Subproject C2","_id":"125"},{"grant_number":"02L19C115","name":"KIAM: KIAM: Kompetenzzentrum KI in der Arbeitswelt des industriellen Mittelstands in OstWestfalenLippe","_id":"409"}],"_id":"48285","language":[{"iso":"eng"}],"citation":{"ieee":"A. Lebedeva, J. Kornowicz, O. Lammert, and J. Papenkordt, “The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks,” 2023, doi: <a href=\"https://doi.org/10.1007/978-3-031-35891-3_9\">10.1007/978-3-031-35891-3_9</a>.","chicago":"Lebedeva, Anastasia, Jaroslaw Kornowicz, Olesja Lammert, and Jörg Papenkordt. “The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks.” In <i>Artificial Intelligence in HCI</i>. Artificial Intelligence in HCI, 2023. <a href=\"https://doi.org/10.1007/978-3-031-35891-3_9\">https://doi.org/10.1007/978-3-031-35891-3_9</a>.","ama":"Lebedeva A, Kornowicz J, Lammert O, Papenkordt J. The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks. In: <i>Artificial Intelligence in HCI</i>. Artificial Intelligence in HCI. ; 2023. doi:<a href=\"https://doi.org/10.1007/978-3-031-35891-3_9\">10.1007/978-3-031-35891-3_9</a>","apa":"Lebedeva, A., Kornowicz, J., Lammert, O., &#38; Papenkordt, J. (2023). The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks. <i>Artificial Intelligence in HCI</i>. <a href=\"https://doi.org/10.1007/978-3-031-35891-3_9\">https://doi.org/10.1007/978-3-031-35891-3_9</a>","mla":"Lebedeva, Anastasia, et al. “The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks.” <i>Artificial Intelligence in HCI</i>, 2023, doi:<a href=\"https://doi.org/10.1007/978-3-031-35891-3_9\">10.1007/978-3-031-35891-3_9</a>.","short":"A. Lebedeva, J. Kornowicz, O. Lammert, J. Papenkordt, in: Artificial Intelligence in HCI, 2023.","bibtex":"@inproceedings{Lebedeva_Kornowicz_Lammert_Papenkordt_2023, series={Artificial Intelligence in HCI}, title={The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks}, DOI={<a href=\"https://doi.org/10.1007/978-3-031-35891-3_9\">10.1007/978-3-031-35891-3_9</a>}, booktitle={Artificial Intelligence in HCI}, author={Lebedeva, Anastasia and Kornowicz, Jaroslaw and Lammert, Olesja and Papenkordt, Jörg}, year={2023}, collection={Artificial Intelligence in HCI} }"},"year":"2023","author":[{"full_name":"Lebedeva, Anastasia","last_name":"Lebedeva","first_name":"Anastasia"},{"last_name":"Kornowicz","full_name":"Kornowicz, Jaroslaw","id":"44029","first_name":"Jaroslaw"},{"first_name":"Olesja","last_name":"Lammert","id":"47384","full_name":"Lammert, Olesja"},{"first_name":"Jörg","last_name":"Papenkordt","full_name":"Papenkordt, Jörg","id":"44648"}],"date_created":"2023-10-19T07:53:23Z","date_updated":"2023-11-23T08:06:17Z","doi":"10.1007/978-3-031-35891-3_9","title":"The Role of Response Time for Algorithm Aversion in Fast and Slow Thinking Tasks"},{"citation":{"mla":"Kornowicz, Jaroslaw, and Kirsten Thommes. “Aggregating Human Domain Knowledge for Feature Ranking.” <i>Artificial Intelligence in HCI</i>, Springer Nature Switzerland, 2023, doi:<a href=\"https://doi.org/10.1007/978-3-031-35891-3_7\">10.1007/978-3-031-35891-3_7</a>.","short":"J. Kornowicz, K. Thommes, Artificial Intelligence in HCI (2023).","bibtex":"@article{Kornowicz_Thommes_2023, title={Aggregating Human Domain Knowledge for Feature Ranking}, DOI={<a href=\"https://doi.org/10.1007/978-3-031-35891-3_7\">10.1007/978-3-031-35891-3_7</a>}, journal={Artificial Intelligence in HCI}, publisher={Springer Nature Switzerland}, author={Kornowicz, Jaroslaw and Thommes, Kirsten}, year={2023} }","apa":"Kornowicz, J., &#38; Thommes, K. (2023). Aggregating Human Domain Knowledge for Feature Ranking. <i>Artificial Intelligence in HCI</i>. <a href=\"https://doi.org/10.1007/978-3-031-35891-3_7\">https://doi.org/10.1007/978-3-031-35891-3_7</a>","ama":"Kornowicz J, Thommes K. Aggregating Human Domain Knowledge for Feature Ranking. <i>Artificial Intelligence in HCI</i>. Published online 2023. doi:<a href=\"https://doi.org/10.1007/978-3-031-35891-3_7\">10.1007/978-3-031-35891-3_7</a>","ieee":"J. Kornowicz and K. Thommes, “Aggregating Human Domain Knowledge for Feature Ranking,” <i>Artificial Intelligence in HCI</i>, 2023, doi: <a href=\"https://doi.org/10.1007/978-3-031-35891-3_7\">10.1007/978-3-031-35891-3_7</a>.","chicago":"Kornowicz, Jaroslaw, and Kirsten Thommes. “Aggregating Human Domain Knowledge for Feature Ranking.” <i>Artificial Intelligence in HCI</i>, 2023. <a href=\"https://doi.org/10.1007/978-3-031-35891-3_7\">https://doi.org/10.1007/978-3-031-35891-3_7</a>."},"year":"2023","publication_identifier":{"issn":["0302-9743","1611-3349"],"isbn":["9783031358906","9783031358913"]},"publication_status":"published","doi":"10.1007/978-3-031-35891-3_7","title":"Aggregating Human Domain Knowledge for Feature Ranking","date_created":"2023-10-11T08:01:00Z","author":[{"id":"44029","full_name":"Kornowicz, Jaroslaw","last_name":"Kornowicz","first_name":"Jaroslaw"},{"last_name":"Thommes","id":"72497","full_name":"Thommes, Kirsten","first_name":"Kirsten"}],"date_updated":"2023-12-05T10:15:37Z","publisher":"Springer Nature Switzerland","status":"public","publication":"Artificial Intelligence in HCI","type":"journal_article","language":[{"iso":"eng"}],"department":[{"_id":"184"},{"_id":"178"}],"user_id":"42933","_id":"47953","project":[{"_id":"125","name":"TRR 318 - C2: TRR 318 - Subproject C2"}]},{"publication_status":"published","publication_identifier":{"issn":["0302-9743","1611-3349"],"isbn":["9783031452741","9783031452758"]},"year":"2023","place":"Cham","citation":{"ama":"Hanselle JM, Fürnkranz J, Hüllermeier E. Probabilistic Scoring Lists for Interpretable Machine Learning. In: <i>Discovery Science</i>. Springer Nature Switzerland; 2023. doi:<a href=\"https://doi.org/10.1007/978-3-031-45275-8_13\">10.1007/978-3-031-45275-8_13</a>","apa":"Hanselle, J. M., Fürnkranz, J., &#38; Hüllermeier, E. (2023). Probabilistic Scoring Lists for Interpretable Machine Learning. In <i>Discovery Science</i>. Springer Nature Switzerland. <a href=\"https://doi.org/10.1007/978-3-031-45275-8_13\">https://doi.org/10.1007/978-3-031-45275-8_13</a>","short":"J.M. Hanselle, J. Fürnkranz, E. Hüllermeier, in: Discovery Science, Springer Nature Switzerland, Cham, 2023.","mla":"Hanselle, Jonas Manuel, et al. “Probabilistic Scoring Lists for Interpretable Machine Learning.” <i>Discovery Science</i>, Springer Nature Switzerland, 2023, doi:<a href=\"https://doi.org/10.1007/978-3-031-45275-8_13\">10.1007/978-3-031-45275-8_13</a>.","bibtex":"@inbook{Hanselle_Fürnkranz_Hüllermeier_2023, place={Cham}, title={Probabilistic Scoring Lists for Interpretable Machine Learning}, DOI={<a href=\"https://doi.org/10.1007/978-3-031-45275-8_13\">10.1007/978-3-031-45275-8_13</a>}, booktitle={Discovery Science}, publisher={Springer Nature Switzerland}, author={Hanselle, Jonas Manuel and Fürnkranz, Johannes and Hüllermeier, Eyke}, year={2023} }","chicago":"Hanselle, Jonas Manuel, Johannes Fürnkranz, and Eyke Hüllermeier. “Probabilistic Scoring Lists for Interpretable Machine Learning.” In <i>Discovery Science</i>. Cham: Springer Nature Switzerland, 2023. <a href=\"https://doi.org/10.1007/978-3-031-45275-8_13\">https://doi.org/10.1007/978-3-031-45275-8_13</a>.","ieee":"J. M. Hanselle, J. Fürnkranz, and E. Hüllermeier, “Probabilistic Scoring Lists for Interpretable Machine Learning,” in <i>Discovery Science</i>, Cham: Springer Nature Switzerland, 2023."},"date_updated":"2024-06-26T14:25:50Z","publisher":"Springer Nature Switzerland","date_created":"2024-06-26T14:24:29Z","author":[{"full_name":"Hanselle, Jonas Manuel","id":"43980","last_name":"Hanselle","orcid":"0000-0002-1231-4985","first_name":"Jonas Manuel"},{"first_name":"Johannes","last_name":"Fürnkranz","full_name":"Fürnkranz, Johannes"},{"first_name":"Eyke","id":"48129","full_name":"Hüllermeier, Eyke","last_name":"Hüllermeier"}],"title":"Probabilistic Scoring Lists for Interpretable Machine Learning","doi":"10.1007/978-3-031-45275-8_13","type":"book_chapter","publication":"Discovery Science","status":"public","project":[{"name":"TRR 318 - C2: TRR 318 - Subproject C2","_id":"125"}],"_id":"54909","user_id":"72497","department":[{"_id":"660"}],"language":[{"iso":"eng"}]},{"language":[{"iso":"eng"}],"user_id":"72497","department":[{"_id":"178"},{"_id":"184"}],"project":[{"name":"TRR 318 - C2: TRR 318 - Subproject C2","_id":"125"}],"_id":"51209","status":"public","editor":[{"first_name":"M","full_name":"Leyer, M","last_name":"Leyer"},{"first_name":"J","last_name":"Wichmann","full_name":"Wichmann, J"}],"type":"conference","publication":"LWDA’23: Learning, Knowledge, Data, Analysis. ","main_file_link":[{"url":"https://ceur-ws.org/Vol-3630/LWDA2023-paper38.pdf"}],"title":"Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain","date_created":"2024-02-07T09:24:37Z","author":[{"last_name":"Hanselle","orcid":"0000-0002-1231-4985","full_name":"Hanselle, Jonas Manuel","id":"43980","first_name":"Jonas Manuel"},{"full_name":"Kornowicz, Jaroslaw","id":"44029","last_name":"Kornowicz","orcid":"0000-0002-5654-9911","first_name":"Jaroslaw"},{"first_name":"Stefan","last_name":"Heid","orcid":"0000-0002-9461-7372","id":"39640","full_name":"Heid, Stefan"},{"last_name":"Thommes","full_name":"Thommes, Kirsten","id":"72497","first_name":"Kirsten"},{"full_name":"Hüllermeier, Eyke","id":"48129","last_name":"Hüllermeier","first_name":"Eyke"}],"date_updated":"2024-12-09T08:09:28Z","citation":{"ieee":"J. M. Hanselle, J. Kornowicz, S. Heid, K. Thommes, and E. Hüllermeier, “Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain,” in <i>LWDA’23: Learning, Knowledge, Data, Analysis. </i>, 2023.","chicago":"Hanselle, Jonas Manuel, Jaroslaw Kornowicz, Stefan Heid, Kirsten Thommes, and Eyke Hüllermeier. “Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain.” In <i>LWDA’23: Learning, Knowledge, Data, Analysis. </i>, edited by M Leyer and J Wichmann, 2023.","ama":"Hanselle JM, Kornowicz J, Heid S, Thommes K, Hüllermeier E. Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain. In: Leyer M, Wichmann J, eds. <i>LWDA’23: Learning, Knowledge, Data, Analysis. </i>. ; 2023.","short":"J.M. Hanselle, J. Kornowicz, S. Heid, K. Thommes, E. Hüllermeier, in: M. Leyer, J. Wichmann (Eds.), LWDA’23: Learning, Knowledge, Data, Analysis. , 2023.","bibtex":"@inproceedings{Hanselle_Kornowicz_Heid_Thommes_Hüllermeier_2023, title={Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain}, booktitle={LWDA’23: Learning, Knowledge, Data, Analysis. }, author={Hanselle, Jonas Manuel and Kornowicz, Jaroslaw and Heid, Stefan and Thommes, Kirsten and Hüllermeier, Eyke}, editor={Leyer, M and Wichmann, J}, year={2023} }","mla":"Hanselle, Jonas Manuel, et al. “Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain.” <i>LWDA’23: Learning, Knowledge, Data, Analysis. </i>, edited by M Leyer and J Wichmann, 2023.","apa":"Hanselle, J. M., Kornowicz, J., Heid, S., Thommes, K., &#38; Hüllermeier, E. (2023). Comparing Humans and Algorithms in Feature Ranking: A Case-Study in the Medical Domain. In M. Leyer &#38; J. Wichmann (Eds.), <i>LWDA’23: Learning, Knowledge, Data, Analysis. </i>."},"year":"2023","publication_identifier":{"issn":["1613-0073"]}}]
