[{"status":"public","type":"journal_article","article_type":"original","user_id":"118059","department":[{"_id":"756"},{"_id":"26"}],"_id":"62028","citation":{"apa":"Fahimi, M., State, L., &#38; Kasirzadeh, A. (2025). From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection. <i>Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</i>, <i>8</i>(1), 879–892. <a href=\"https://doi.org/10.1609/aies.v8i1.36597\">https://doi.org/10.1609/aies.v8i1.36597</a>","bibtex":"@article{Fahimi_State_Kasirzadeh_2025, title={From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection}, volume={8}, DOI={<a href=\"https://doi.org/10.1609/aies.v8i1.36597\">10.1609/aies.v8i1.36597</a>}, number={1}, journal={Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society}, publisher={Association for the Advancement of Artificial Intelligence (AAAI)}, author={Fahimi, Miriam and State, Laura and Kasirzadeh, Atoosa}, year={2025}, pages={879–892} }","mla":"Fahimi, Miriam, et al. “From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection.” <i>Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</i>, vol. 8, no. 1, Association for the Advancement of Artificial Intelligence (AAAI), 2025, pp. 879–92, doi:<a href=\"https://doi.org/10.1609/aies.v8i1.36597\">10.1609/aies.v8i1.36597</a>.","short":"M. Fahimi, L. State, A. Kasirzadeh, Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society 8 (2025) 879–892.","ama":"Fahimi M, State L, Kasirzadeh A. From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection. <i>Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</i>. 2025;8(1):879-892. doi:<a href=\"https://doi.org/10.1609/aies.v8i1.36597\">10.1609/aies.v8i1.36597</a>","chicago":"Fahimi, Miriam, Laura State, and Atoosa Kasirzadeh. “From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection.” <i>Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</i> 8, no. 1 (2025): 879–92. <a href=\"https://doi.org/10.1609/aies.v8i1.36597\">https://doi.org/10.1609/aies.v8i1.36597</a>.","ieee":"M. Fahimi, L. State, and A. Kasirzadeh, “From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection,” <i>Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</i>, vol. 8, no. 1, pp. 879–892, 2025, doi: <a href=\"https://doi.org/10.1609/aies.v8i1.36597\">10.1609/aies.v8i1.36597</a>."},"page":"879-892","intvolume":"         8","publication_status":"published","publication_identifier":{"issn":["3065-8365"]},"doi":"10.1609/aies.v8i1.36597","author":[{"full_name":"Fahimi, Miriam","id":"118059","last_name":"Fahimi","orcid":"0000-0002-0619-3160","first_name":"Miriam"},{"full_name":"State, Laura","last_name":"State","first_name":"Laura"},{"last_name":"Kasirzadeh","full_name":"Kasirzadeh, Atoosa","first_name":"Atoosa"}],"volume":8,"date_updated":"2025-11-18T10:09:40Z","abstract":[{"lang":"eng","text":"Explainable AI (XAI) methods can support the identification of biases in automated decision-making (ADM) systems. However, existing research does not sufficiently address whether these biases originate from the ADM system or mirror underlying societal inequalities. This distinction is important because it has major implications for how to act upon an explanation: while the societal bias produced by the ADM system can be algorithmically fixed, societal inequalities demand societal actions. To address this gap, we propose the RR-XAI-framework (recognition-redistribution through XAI) that builds on a distinction between socio-technical and societal bias and Nancy Fraser's justice theory of recognition and redistribution. In our framework, explanations can play two distinct roles: as a socio-technical diagnosis when they reveal biases produced by the ADM system itself, or as a societal diagnosis when they expose biases that reflect broader societal inequalities. We then outline the operationalization of the framework and discuss its applicability for cases in algorithmic hiring and credit scoring. Based on our findings, we argue that the diagnostic functions of XAI are contingent on the provision of such explanations, the resources of the audiences, as well as the current limits of XAI techniques."}],"publication":"Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society","language":[{"iso":"eng"}],"year":"2025","issue":"1","title":"From Explaining to Diagnosing: A Justice-Oriented Framework of Explainable AI for Bias Detection","date_created":"2025-10-31T15:05:38Z","publisher":"Association for the Advancement of Artificial Intelligence (AAAI)"},{"status":"public","abstract":[{"text":"In 2024, the EU adopted the AI Act, a new set of rules for trustworthy artificial intelligence. This legal instrument carves a large place for standardisation, a regulatory technique that consists in crafting so-called harmonised technical standards, to facilitate legal compliance by industry stakeholders. While EU technical standards have been used in the past for ensuring product safety, for the first time the AI Act relies on standardisation to facilitate compliance with fundamental rights, including the right to non-discrimination and equality. The attempt to translate inherently open-textured rights and ethical principles into operationalizable standards raises critical questions. In particular, how will standardisation practices under the new EU AI Act affect, transform, contest and stabilise notions of equality and non-discrimination in an increasingly algorithmic society? This paper proposes a research agenda to address this question and unpack the black box of AI standardisation.","lang":"eng"}],"type":"conference","publication":"Proceedings of Fourth European Workshop on Algorithmic Fairness","language":[{"iso":"eng"}],"user_id":"118059","department":[{"_id":"756"},{"_id":"26"}],"_id":"62229","citation":{"mla":"Xenidis, Raphaële, and Miriam Fahimi. “Standardising Equality in the Algorithmic Society? A Research Agenda.” <i>Proceedings of Fourth European Workshop on Algorithmic Fairness</i>, PMLR, 2025, pp. 310–314.","bibtex":"@inproceedings{Xenidis_Fahimi_2025, title={Standardising Equality in the Algorithmic Society? A Research Agenda}, booktitle={Proceedings of Fourth European Workshop on Algorithmic Fairness}, publisher={PMLR}, author={Xenidis, Raphaële and Fahimi, Miriam}, year={2025}, pages={310–314} }","short":"R. Xenidis, M. Fahimi, in: Proceedings of Fourth European Workshop on Algorithmic Fairness, PMLR, 2025, pp. 310–314.","apa":"Xenidis, R., &#38; Fahimi, M. (2025). Standardising Equality in the Algorithmic Society? A Research Agenda. <i>Proceedings of Fourth European Workshop on Algorithmic Fairness</i>, 310–314.","chicago":"Xenidis, Raphaële, and Miriam Fahimi. “Standardising Equality in the Algorithmic Society? A Research Agenda.” In <i>Proceedings of Fourth European Workshop on Algorithmic Fairness</i>, 310–314. PMLR, 2025.","ieee":"R. Xenidis and M. Fahimi, “Standardising Equality in the Algorithmic Society? A Research Agenda,” in <i>Proceedings of Fourth European Workshop on Algorithmic Fairness</i>, 2025, pp. 310–314.","ama":"Xenidis R, Fahimi M. Standardising Equality in the Algorithmic Society? A Research Agenda. In: <i>Proceedings of Fourth European Workshop on Algorithmic Fairness</i>. PMLR; 2025:310–314."},"page":"310–314","year":"2025","title":"Standardising Equality in the Algorithmic Society? A Research Agenda","author":[{"first_name":"Raphaële","full_name":"Xenidis, Raphaële","last_name":"Xenidis"},{"orcid":"0000-0002-0619-3160","last_name":"Fahimi","id":"118059","full_name":"Fahimi, Miriam","first_name":"Miriam"}],"date_created":"2025-11-18T09:59:34Z","publisher":"PMLR","date_updated":"2025-11-18T10:02:20Z"},{"status":"public","type":"journal_article","publication":"Digital Culture & Society","language":[{"iso":"ger"}],"extern":"1","user_id":"118059","_id":"62029","citation":{"apa":"Fahimi, M., &#38; Kinder-Kurlanda, K. (2025). Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring. <i>Digital Culture &#38; Society</i>, <i>9</i>(2), 141–160. <a href=\"https://doi.org/10.14361/dcs-2023-0208\">https://doi.org/10.14361/dcs-2023-0208</a>","short":"M. Fahimi, K. Kinder-Kurlanda, Digital Culture &#38; Society 9 (2025) 141–160.","bibtex":"@article{Fahimi_Kinder-Kurlanda_2025, title={Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring}, volume={9}, DOI={<a href=\"https://doi.org/10.14361/dcs-2023-0208\">10.14361/dcs-2023-0208</a>}, number={2}, journal={Digital Culture &#38; Society}, publisher={Transcript Verlag}, author={Fahimi, Miriam and Kinder-Kurlanda, Katharina}, year={2025}, pages={141–160} }","mla":"Fahimi, Miriam, and Katharina Kinder-Kurlanda. “Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring.” <i>Digital Culture &#38; Society</i>, vol. 9, no. 2, Transcript Verlag, 2025, pp. 141–60, doi:<a href=\"https://doi.org/10.14361/dcs-2023-0208\">10.14361/dcs-2023-0208</a>.","ama":"Fahimi M, Kinder-Kurlanda K. Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring. <i>Digital Culture &#38; Society</i>. 2025;9(2):141-160. doi:<a href=\"https://doi.org/10.14361/dcs-2023-0208\">10.14361/dcs-2023-0208</a>","chicago":"Fahimi, Miriam, and Katharina Kinder-Kurlanda. “Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring.” <i>Digital Culture &#38; Society</i> 9, no. 2 (2025): 141–60. <a href=\"https://doi.org/10.14361/dcs-2023-0208\">https://doi.org/10.14361/dcs-2023-0208</a>.","ieee":"M. Fahimi and K. Kinder-Kurlanda, “Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring,” <i>Digital Culture &#38; Society</i>, vol. 9, no. 2, pp. 141–160, 2025, doi: <a href=\"https://doi.org/10.14361/dcs-2023-0208\">10.14361/dcs-2023-0208</a>."},"intvolume":"         9","page":"141-160","year":"2025","issue":"2","publication_status":"published","publication_identifier":{"issn":["2364-2114","2364-2122"]},"doi":"10.14361/dcs-2023-0208","title":"Friction in the Materialities of Value. Relating Transparency, Algorithms and Credit Scoring","date_created":"2025-10-31T15:07:42Z","author":[{"first_name":"Miriam","last_name":"Fahimi","orcid":"0000-0002-0619-3160","id":"118059","full_name":"Fahimi, Miriam"},{"full_name":"Kinder-Kurlanda, Katharina","last_name":"Kinder-Kurlanda","first_name":"Katharina"}],"volume":9,"publisher":"Transcript Verlag","date_updated":"2025-12-04T18:42:03Z"},{"volume":26,"author":[{"full_name":"Alvarez, Jose M.","last_name":"Alvarez","first_name":"Jose M."},{"last_name":"Colmenarejo","full_name":"Colmenarejo, Alejandra Bringas","first_name":"Alejandra Bringas"},{"first_name":"Alaa","full_name":"Elobaid, Alaa","last_name":"Elobaid"},{"full_name":"Fabbrizzi, Simone","last_name":"Fabbrizzi","first_name":"Simone"},{"first_name":"Miriam","orcid":"0000-0002-0619-3160","last_name":"Fahimi","id":"118059","full_name":"Fahimi, Miriam"},{"last_name":"Ferrara","full_name":"Ferrara, Antonio","first_name":"Antonio"},{"last_name":"Ghodsi","full_name":"Ghodsi, Siamak","first_name":"Siamak"},{"first_name":"Carlos","full_name":"Mougan, Carlos","last_name":"Mougan"},{"last_name":"Papageorgiou","full_name":"Papageorgiou, Ioanna","first_name":"Ioanna"},{"first_name":"Paula","full_name":"Reyero, Paula","last_name":"Reyero"},{"last_name":"Russo","full_name":"Russo, Mayra","first_name":"Mayra"},{"full_name":"Scott, Kristen M.","last_name":"Scott","first_name":"Kristen M."},{"first_name":"Laura","last_name":"State","full_name":"State, Laura"},{"last_name":"Zhao","full_name":"Zhao, Xuan","first_name":"Xuan"},{"full_name":"Ruggieri, Salvatore","last_name":"Ruggieri","first_name":"Salvatore"}],"date_created":"2025-10-31T15:17:11Z","publisher":"Springer Science and Business Media LLC","date_updated":"2025-11-18T09:56:30Z","doi":"10.1007/s10676-024-09746-w","title":"Policy advice and best practices on bias and fairness in AI","issue":"2","publication_identifier":{"issn":["1388-1957","1572-8439"]},"publication_status":"published","intvolume":"        26","citation":{"mla":"Alvarez, Jose M., et al. “Policy Advice and Best Practices on Bias and Fairness in AI.” <i>Ethics and Information Technology</i>, vol. 26, no. 2, 31, Springer Science and Business Media LLC, 2024, doi:<a href=\"https://doi.org/10.1007/s10676-024-09746-w\">10.1007/s10676-024-09746-w</a>.","short":"J.M. Alvarez, A.B. Colmenarejo, A. Elobaid, S. Fabbrizzi, M. Fahimi, A. Ferrara, S. Ghodsi, C. Mougan, I. Papageorgiou, P. Reyero, M. Russo, K.M. Scott, L. State, X. Zhao, S. Ruggieri, Ethics and Information Technology 26 (2024).","bibtex":"@article{Alvarez_Colmenarejo_Elobaid_Fabbrizzi_Fahimi_Ferrara_Ghodsi_Mougan_Papageorgiou_Reyero_et al._2024, title={Policy advice and best practices on bias and fairness in AI}, volume={26}, DOI={<a href=\"https://doi.org/10.1007/s10676-024-09746-w\">10.1007/s10676-024-09746-w</a>}, number={231}, journal={Ethics and Information Technology}, publisher={Springer Science and Business Media LLC}, author={Alvarez, Jose M. and Colmenarejo, Alejandra Bringas and Elobaid, Alaa and Fabbrizzi, Simone and Fahimi, Miriam and Ferrara, Antonio and Ghodsi, Siamak and Mougan, Carlos and Papageorgiou, Ioanna and Reyero, Paula and et al.}, year={2024} }","apa":"Alvarez, J. M., Colmenarejo, A. B., Elobaid, A., Fabbrizzi, S., Fahimi, M., Ferrara, A., Ghodsi, S., Mougan, C., Papageorgiou, I., Reyero, P., Russo, M., Scott, K. M., State, L., Zhao, X., &#38; Ruggieri, S. (2024). Policy advice and best practices on bias and fairness in AI. <i>Ethics and Information Technology</i>, <i>26</i>(2), Article 31. <a href=\"https://doi.org/10.1007/s10676-024-09746-w\">https://doi.org/10.1007/s10676-024-09746-w</a>","ama":"Alvarez JM, Colmenarejo AB, Elobaid A, et al. Policy advice and best practices on bias and fairness in AI. <i>Ethics and Information Technology</i>. 2024;26(2). doi:<a href=\"https://doi.org/10.1007/s10676-024-09746-w\">10.1007/s10676-024-09746-w</a>","chicago":"Alvarez, Jose M., Alejandra Bringas Colmenarejo, Alaa Elobaid, Simone Fabbrizzi, Miriam Fahimi, Antonio Ferrara, Siamak Ghodsi, et al. “Policy Advice and Best Practices on Bias and Fairness in AI.” <i>Ethics and Information Technology</i> 26, no. 2 (2024). <a href=\"https://doi.org/10.1007/s10676-024-09746-w\">https://doi.org/10.1007/s10676-024-09746-w</a>.","ieee":"J. M. Alvarez <i>et al.</i>, “Policy advice and best practices on bias and fairness in AI,” <i>Ethics and Information Technology</i>, vol. 26, no. 2, Art. no. 31, 2024, doi: <a href=\"https://doi.org/10.1007/s10676-024-09746-w\">10.1007/s10676-024-09746-w</a>."},"year":"2024","user_id":"118059","_id":"62033","extern":"1","language":[{"iso":"eng"}],"article_number":"31","publication":"Ethics and Information Technology","type":"journal_article","status":"public","abstract":[{"lang":"eng","text":"<jats:title>Abstract</jats:title><jats:p>The literature addressing bias and fairness in AI models (<jats:italic>fair-AI</jats:italic>) is growing at a fast pace, making it difficult for novel researchers and practitioners to have a bird’s-eye view picture of the field. In particular, many policy initiatives, standards, and best practices in fair-AI have been proposed for setting principles, procedures, and knowledge bases to guide and operationalize the management of bias and fairness. The first objective of this paper is to concisely survey the state-of-the-art of fair-AI methods and resources, and the main policies on bias in AI, with the aim of providing such a bird’s-eye guidance for both researchers and practitioners. The second objective of the paper is to contribute to the policy advice and best practices state-of-the-art by leveraging from the results of the NoBIAS research project. We present and discuss a few relevant topics organized around the NoBIAS architecture, which is made up of a Legal Layer, focusing on the European Union context, and a Bias Management Layer, focusing on understanding, mitigating, and accounting for bias.</jats:p>"}]},{"issue":"CSCW2","publication_identifier":{"issn":["2573-0142"]},"publication_status":"published","page":"1-23","intvolume":"         8","citation":{"chicago":"Fahimi, Miriam, Mayra Russo, Kristen M. Scott, Maria-Esther Vidal, Bettina Berendt, and Katharina Kinder-Kurlanda. “Articulation Work and Tinkering for Fairness in Machine Learning.” <i>Proceedings of the ACM on Human-Computer Interaction</i> 8, no. CSCW2 (2024): 1–23. <a href=\"https://doi.org/10.1145/3686973\">https://doi.org/10.1145/3686973</a>.","ieee":"M. Fahimi, M. Russo, K. M. Scott, M.-E. Vidal, B. Berendt, and K. Kinder-Kurlanda, “Articulation Work and Tinkering for Fairness in Machine Learning,” <i>Proceedings of the ACM on Human-Computer Interaction</i>, vol. 8, no. CSCW2, pp. 1–23, 2024, doi: <a href=\"https://doi.org/10.1145/3686973\">10.1145/3686973</a>.","ama":"Fahimi M, Russo M, Scott KM, Vidal M-E, Berendt B, Kinder-Kurlanda K. Articulation Work and Tinkering for Fairness in Machine Learning. <i>Proceedings of the ACM on Human-Computer Interaction</i>. 2024;8(CSCW2):1-23. doi:<a href=\"https://doi.org/10.1145/3686973\">10.1145/3686973</a>","apa":"Fahimi, M., Russo, M., Scott, K. M., Vidal, M.-E., Berendt, B., &#38; Kinder-Kurlanda, K. (2024). Articulation Work and Tinkering for Fairness in Machine Learning. <i>Proceedings of the ACM on Human-Computer Interaction</i>, <i>8</i>(CSCW2), 1–23. <a href=\"https://doi.org/10.1145/3686973\">https://doi.org/10.1145/3686973</a>","bibtex":"@article{Fahimi_Russo_Scott_Vidal_Berendt_Kinder-Kurlanda_2024, title={Articulation Work and Tinkering for Fairness in Machine Learning}, volume={8}, DOI={<a href=\"https://doi.org/10.1145/3686973\">10.1145/3686973</a>}, number={CSCW2}, journal={Proceedings of the ACM on Human-Computer Interaction}, publisher={Association for Computing Machinery (ACM)}, author={Fahimi, Miriam and Russo, Mayra and Scott, Kristen M. and Vidal, Maria-Esther and Berendt, Bettina and Kinder-Kurlanda, Katharina}, year={2024}, pages={1–23} }","short":"M. Fahimi, M. Russo, K.M. Scott, M.-E. Vidal, B. Berendt, K. Kinder-Kurlanda, Proceedings of the ACM on Human-Computer Interaction 8 (2024) 1–23.","mla":"Fahimi, Miriam, et al. “Articulation Work and Tinkering for Fairness in Machine Learning.” <i>Proceedings of the ACM on Human-Computer Interaction</i>, vol. 8, no. CSCW2, Association for Computing Machinery (ACM), 2024, pp. 1–23, doi:<a href=\"https://doi.org/10.1145/3686973\">10.1145/3686973</a>."},"year":"2024","volume":8,"author":[{"full_name":"Fahimi, Miriam","id":"118059","last_name":"Fahimi","orcid":"0000-0002-0619-3160","first_name":"Miriam"},{"first_name":"Mayra","full_name":"Russo, Mayra","last_name":"Russo"},{"full_name":"Scott, Kristen M.","last_name":"Scott","first_name":"Kristen M."},{"last_name":"Vidal","full_name":"Vidal, Maria-Esther","first_name":"Maria-Esther"},{"first_name":"Bettina","full_name":"Berendt, Bettina","last_name":"Berendt"},{"full_name":"Kinder-Kurlanda, Katharina","last_name":"Kinder-Kurlanda","first_name":"Katharina"}],"date_created":"2025-10-31T15:10:31Z","date_updated":"2025-11-18T09:56:04Z","publisher":"Association for Computing Machinery (ACM)","doi":"10.1145/3686973","title":"Articulation Work and Tinkering for Fairness in Machine Learning","publication":"Proceedings of the ACM on Human-Computer Interaction","type":"journal_article","status":"public","abstract":[{"lang":"eng","text":"The field of fair AI aims to counter biased algorithms through computational modelling. However, it faces increasing criticism for perpetuating the use of overly technical and reductionist methods. As a result, novel approaches appear in the field to address more socially-oriented and interdisciplinary (SOI) perspectives on fair AI. In this paper, we take this dynamic as the starting point to study the tension between computer science (CS) and SOI research. By drawing on STS and CSCW theory, we position fair AI research as a matter of 'organizational alignment': what makes research 'doable' is the successful alignment of three levels of work organization (the social world, the laboratory, and the experiment). Based on qualitative interviews with CS researchers, we analyze the tasks, resources, and actors required for doable research in the case of fair AI. We find that CS researchers engage with SOI research to some extent, but organizational conditions, articulation work, and ambiguities of the social world constrain the doability of SOI research for them. Based on our findings, we identify and discuss problems for aligning CS and SOI as fair AI continues to evolve."}],"user_id":"118059","_id":"62031","language":[{"iso":"eng"}],"extern":"1"},{"title":"In/visibilities in Data Studies: Methods, Tools, and Interventions","publisher":"Bristol University Press","date_updated":"2025-11-18T10:02:15Z","date_created":"2025-11-18T09:58:30Z","author":[{"first_name":"Miriam","id":"118059","full_name":"Fahimi, Miriam","orcid":"0000-0002-0619-3160","last_name":"Fahimi"},{"first_name":"Petter","last_name":"Falk","full_name":"Falk, Petter"},{"first_name":"Jonathan W. Y.","full_name":"Gray, Jonathan W. Y.","last_name":"Gray"},{"full_name":"Jarke, Juliane","last_name":"Jarke","first_name":"Juliane"},{"full_name":"Kinder-Kurlanda, Katharina","last_name":"Kinder-Kurlanda","first_name":"Katharina"},{"full_name":"Light, Evan","last_name":"Light","first_name":"Evan"},{"full_name":"McGeachey, Ellouise","last_name":"McGeachey","first_name":"Ellouise"},{"full_name":"Perea, Itzelle Medina","last_name":"Perea","first_name":"Itzelle Medina"},{"first_name":"Nikolaus","last_name":"Poechhacker","full_name":"Poechhacker, Nikolaus"},{"last_name":"Poirier","full_name":"Poirier, Lindsay","first_name":"Lindsay"},{"full_name":"Röhle, Theo","last_name":"Röhle","first_name":"Theo"},{"first_name":"Tamar","full_name":"Sharon, Tamar","last_name":"Sharon"},{"first_name":"Marthe","last_name":"Stevens","full_name":"Stevens, Marthe"},{"first_name":"Bernard van","full_name":"Gastel, Bernard van","last_name":"Gastel"},{"last_name":"White","full_name":"White, Quinn","first_name":"Quinn"},{"last_name":"Zakharova","full_name":"Zakharova, Irina","first_name":"Irina"}],"year":"2024","citation":{"short":"M. Fahimi, P. Falk, J.W.Y. Gray, J. Jarke, K. Kinder-Kurlanda, E. Light, E. McGeachey, I.M. Perea, N. Poechhacker, L. Poirier, T. Röhle, T. Sharon, M. Stevens, B. van Gastel, Q. White, I. Zakharova, in: Dialogues in Data Power, Bristol University Press, 2024, pp. 52–79.","mla":"Fahimi, Miriam, et al. “In/Visibilities in Data Studies: Methods, Tools, and Interventions.” <i>Dialogues in Data Power</i>, Bristol University Press, 2024, pp. 52–79.","bibtex":"@inbook{Fahimi_Falk_Gray_Jarke_Kinder-Kurlanda_Light_McGeachey_Perea_Poechhacker_Poirier_et al._2024, title={In/visibilities in Data Studies: Methods, Tools, and Interventions}, booktitle={Dialogues in Data Power}, publisher={Bristol University Press}, author={Fahimi, Miriam and Falk, Petter and Gray, Jonathan W. Y. and Jarke, Juliane and Kinder-Kurlanda, Katharina and Light, Evan and McGeachey, Ellouise and Perea, Itzelle Medina and Poechhacker, Nikolaus and Poirier, Lindsay and et al.}, year={2024}, pages={52–79} }","apa":"Fahimi, M., Falk, P., Gray, J. W. Y., Jarke, J., Kinder-Kurlanda, K., Light, E., McGeachey, E., Perea, I. M., Poechhacker, N., Poirier, L., Röhle, T., Sharon, T., Stevens, M., Gastel, B. van, White, Q., &#38; Zakharova, I. (2024). In/visibilities in Data Studies: Methods, Tools, and Interventions. In <i>Dialogues in Data Power</i> (pp. 52–79). Bristol University Press.","ieee":"M. Fahimi <i>et al.</i>, “In/visibilities in Data Studies: Methods, Tools, and Interventions,” in <i>Dialogues in Data Power</i>, Bristol University Press, 2024, pp. 52–79.","chicago":"Fahimi, Miriam, Petter Falk, Jonathan W. Y. Gray, Juliane Jarke, Katharina Kinder-Kurlanda, Evan Light, Ellouise McGeachey, et al. “In/Visibilities in Data Studies: Methods, Tools, and Interventions.” In <i>Dialogues in Data Power</i>, 52–79. Bristol University Press, 2024.","ama":"Fahimi M, Falk P, Gray JWY, et al. In/visibilities in Data Studies: Methods, Tools, and Interventions. In: <i>Dialogues in Data Power</i>. Bristol University Press; 2024:52–79."},"page":"52–79","publication_identifier":{"isbn":["978-1-5292-3832-7"]},"language":[{"iso":"eng"}],"_id":"62228","user_id":"118059","department":[{"_id":"756"},{"_id":"26"}],"abstract":[{"text":"This chapter highlights the intricate nature of data and their profound social implications. It examines the acts of rendering data visible and the inherent power dynamics and imbalances that accompany such processes. Our dialogue unfolds in three interconnected parts, each focusing on the intersection of in/visibility and power. Part 1 attends to the challenges of producing knowledge about and with data, emphasizing the relativity, fluidity, and instability inherent in data. It explores frameworks that uncover the often invisible infrastructures of algorithms, rendering visible the actors, technologies, and divergent values involved in data manipulation. Part 2 presents empirical case studies that analyse the consequences of data visibility while contemplating the methodological opportunities and challenges of foregrounding the embedded values and norms within data. Part 3 discusses tool-based interventions aimed at bringing alternative data framings and narratives to the fore. It examines the complexities of tracing data across various contexts and the value, utility, and obstacles associated with creating visual representations of data and their flows. By critically engaging with the complexities of data in/visibility, this chapter challenges existing gatekeepers and fosters a deeper understanding of the multifaceted nature of data and its socio-political ramifications.","lang":"eng"}],"status":"public","type":"book_chapter","publication":"Dialogues in Data Power"},{"publication_identifier":{"isbn":["978-94-6372-848-5"]},"page":"309–330","citation":{"apa":"Kinder-Kurlanda, K., &#38; Fahimi, M. (2024). Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions. In J. Jarke, B. Prietl, S. Egbert, Y. Boeva, H. Heuer, &#38; M. Arnold (Eds.), <i>Algorithmic Regimes. Methods, Interactions, and Politics.</i> (pp. 309–330). Amsterdam University Press.","mla":"Kinder-Kurlanda, Katharina, and Miriam Fahimi. “Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions.” <i>Algorithmic Regimes. Methods, Interactions, and Politics.</i>, edited by Juliane Jarke et al., Amsterdam University Press, 2024, pp. 309–330.","short":"K. Kinder-Kurlanda, M. Fahimi, in: J. Jarke, B. Prietl, S. Egbert, Y. Boeva, H. Heuer, M. Arnold (Eds.), Algorithmic Regimes. Methods, Interactions, and Politics., Amsterdam University Press, Amsterdam, 2024, pp. 309–330.","bibtex":"@inbook{Kinder-Kurlanda_Fahimi_2024, place={Amsterdam}, title={Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions}, booktitle={Algorithmic Regimes. Methods, Interactions, and Politics.}, publisher={Amsterdam University Press}, author={Kinder-Kurlanda, Katharina and Fahimi, Miriam}, editor={Jarke, Juliane and Prietl, Bianca and Egbert, Simon and Boeva, Yana and Heuer, Hendrik and Arnold, Maike}, year={2024}, pages={309–330} }","ama":"Kinder-Kurlanda K, Fahimi M. Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions. In: Jarke J, Prietl B, Egbert S, Boeva Y, Heuer H, Arnold M, eds. <i>Algorithmic Regimes. Methods, Interactions, and Politics.</i> Amsterdam University Press; 2024:309–330.","ieee":"K. Kinder-Kurlanda and M. Fahimi, “Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions,” in <i>Algorithmic Regimes. Methods, Interactions, and Politics.</i>, J. Jarke, B. Prietl, S. Egbert, Y. Boeva, H. Heuer, and M. Arnold, Eds. Amsterdam: Amsterdam University Press, 2024, pp. 309–330.","chicago":"Kinder-Kurlanda, Katharina, and Miriam Fahimi. “Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions.” In <i>Algorithmic Regimes. Methods, Interactions, and Politics.</i>, edited by Juliane Jarke, Bianca Prietl, Simon Egbert, Yana Boeva, Hendrik Heuer, and Maike Arnold, 309–330. Amsterdam: Amsterdam University Press, 2024."},"place":"Amsterdam","year":"2024","date_created":"2025-11-18T10:00:38Z","author":[{"full_name":"Kinder-Kurlanda, Katharina","last_name":"Kinder-Kurlanda","first_name":"Katharina"},{"first_name":"Miriam","orcid":"0000-0002-0619-3160","last_name":"Fahimi","id":"118059","full_name":"Fahimi, Miriam"}],"publisher":"Amsterdam University Press","date_updated":"2025-11-18T10:02:25Z","title":"Making Algorithms Fair: Ethnographic Insights from Machine Learning Interventions","publication":"Algorithmic Regimes. Methods, Interactions, and Politics.","type":"book_chapter","status":"public","editor":[{"last_name":"Jarke","full_name":"Jarke, Juliane","first_name":"Juliane"},{"last_name":"Prietl","full_name":"Prietl, Bianca","first_name":"Bianca"},{"full_name":"Egbert, Simon","last_name":"Egbert","first_name":"Simon"},{"first_name":"Yana","full_name":"Boeva, Yana","last_name":"Boeva"},{"full_name":"Heuer, Hendrik","last_name":"Heuer","first_name":"Hendrik"},{"full_name":"Arnold, Maike","last_name":"Arnold","first_name":"Maike"}],"abstract":[{"text":"Algorithms have risen to become one, if not the central technology for producing, circulating, and evaluating knowledge in multiple societal arenas. In this book, scholars from the social sciences, humanities, and computer science argue that this shift has, and will continue to have, profound implications for how knowledge is produced and what and whose knowledge is valued and deemed valid. To attend to this fundamental change, the authors propose the concept of algorithmic regimes and demonstrate how they transform the epistemological, methodological, and political foundations of knowledge production, sensemaking, and decision-making in contemporary societies. Across sixteen chapters, the volume offers a diverse collection of contributions along three perspectives on algorithmic regimes: the methods necessary to research and design algorithmic regimes, the ways in which algorithmic regimes reconfigure sociotechnical interactions, and the politics engrained in algorithmic regimes.","lang":"eng"}],"department":[{"_id":"756"},{"_id":"26"}],"user_id":"118059","_id":"62230","language":[{"iso":"eng"}]},{"publication":"Proceedings of the 2nd European Workshop on Algorithmic Fairness","type":"conference","status":"public","abstract":[{"text":"Explainable artificial intelligence (XAI) is a rapidly growing research field that has received a lot of attention during the last few years. An important goal of the field is to use its methods to detect (social) bias and discrimination. Despite these positive intentions, aspects of XAI can be in conflict with feminist approaches and values. Therefore, our conceptual contribution brings forward both a careful assessment of current XAI methods, as well as visions for carefully doing XAI from a feminist perspective. We conclude with a discussion on the possibilities for caring XAI, and the challenges that might lie along the way.","lang":"eng"}],"department":[{"_id":"756"},{"_id":"26"}],"user_id":"118059","_id":"62231","language":[{"iso":"eng"}],"citation":{"short":"L. State, M. Fahimi, in: Proceedings of the 2nd European Workshop on Algorithmic Fairness, CEUR Workshop Proceedings, Winterthur, Switzerland, 2023.","bibtex":"@inproceedings{State_Fahimi_2023, place={Winterthur, Switzerland}, title={Careful Explanations: A Feminist Perspective on XAI}, booktitle={Proceedings of the 2nd European Workshop on Algorithmic Fairness}, publisher={CEUR Workshop Proceedings}, author={State, Laura and Fahimi, Miriam}, year={2023} }","mla":"State, Laura, and Miriam Fahimi. “Careful Explanations: A Feminist Perspective on XAI.” <i>Proceedings of the 2nd European Workshop on Algorithmic Fairness</i>, CEUR Workshop Proceedings, 2023.","apa":"State, L., &#38; Fahimi, M. (2023). Careful Explanations: A Feminist Perspective on XAI. <i>Proceedings of the 2nd European Workshop on Algorithmic Fairness</i>.","ama":"State L, Fahimi M. Careful Explanations: A Feminist Perspective on XAI. In: <i>Proceedings of the 2nd European Workshop on Algorithmic Fairness</i>. CEUR Workshop Proceedings; 2023.","ieee":"L. State and M. Fahimi, “Careful Explanations: A Feminist Perspective on XAI,” 2023.","chicago":"State, Laura, and Miriam Fahimi. “Careful Explanations: A Feminist Perspective on XAI.” In <i>Proceedings of the 2nd European Workshop on Algorithmic Fairness</i>. Winterthur, Switzerland: CEUR Workshop Proceedings, 2023."},"place":"Winterthur, Switzerland","year":"2023","date_created":"2025-11-18T10:01:00Z","author":[{"last_name":"State","full_name":"State, Laura","first_name":"Laura"},{"first_name":"Miriam","id":"118059","full_name":"Fahimi, Miriam","last_name":"Fahimi","orcid":"0000-0002-0619-3160"}],"date_updated":"2025-11-18T10:02:09Z","publisher":"CEUR Workshop Proceedings","title":"Careful Explanations: A Feminist Perspective on XAI"},{"extern":"1","language":[{"iso":"eng"}],"alternative_title":["Theoretical Approaches and Empirical Studies"],"_id":"62032","user_id":"118059","editor":[{"first_name":"Miriam","id":"118059","full_name":"Fahimi, Miriam","last_name":"Fahimi","orcid":"0000-0002-0619-3160"},{"first_name":"Elmar","full_name":"Flatschart, Elmar","last_name":"Flatschart"},{"first_name":"Wolfram","full_name":"Schaffar, Wolfram","last_name":"Schaffar"}],"status":"public","type":"book_editor","title":"State and Statehood in the Global South","doi":"10.1007/978-3-030-94000-3","date_updated":"2025-11-18T09:56:13Z","publisher":"Springer International Publishing","date_created":"2025-10-31T15:16:00Z","place":"Cham","year":"2022","citation":{"bibtex":"@book{Fahimi_Flatschart_Schaffar_2022, place={Cham}, title={State and Statehood in the Global South}, DOI={<a href=\"https://doi.org/10.1007/978-3-030-94000-3\">10.1007/978-3-030-94000-3</a>}, publisher={Springer International Publishing}, year={2022} }","short":"M. Fahimi, E. Flatschart, W. Schaffar, eds., State and Statehood in the Global South, Springer International Publishing, Cham, 2022.","mla":"Fahimi, Miriam, et al., editors. <i>State and Statehood in the Global South</i>. Springer International Publishing, 2022, doi:<a href=\"https://doi.org/10.1007/978-3-030-94000-3\">10.1007/978-3-030-94000-3</a>.","apa":"Fahimi, M., Flatschart, E., &#38; Schaffar, W. (Eds.). (2022). <i>State and Statehood in the Global South</i>. Springer International Publishing. <a href=\"https://doi.org/10.1007/978-3-030-94000-3\">https://doi.org/10.1007/978-3-030-94000-3</a>","ama":"Fahimi M, Flatschart E, Schaffar W, eds. <i>State and Statehood in the Global South</i>. Springer International Publishing; 2022. doi:<a href=\"https://doi.org/10.1007/978-3-030-94000-3\">10.1007/978-3-030-94000-3</a>","chicago":"Fahimi, Miriam, Elmar Flatschart, and Wolfram Schaffar, eds. <i>State and Statehood in the Global South</i>. Cham: Springer International Publishing, 2022. <a href=\"https://doi.org/10.1007/978-3-030-94000-3\">https://doi.org/10.1007/978-3-030-94000-3</a>.","ieee":"M. Fahimi, E. Flatschart, and W. Schaffar, Eds., <i>State and Statehood in the Global South</i>. Cham: Springer International Publishing, 2022."},"publication_identifier":{"isbn":["9783030939991","9783030940003"]},"publication_status":"published"}]
