[{"doi":"10.1007/s10994-023-06385-y","title":"Incremental permutation feature importance (iPFI): towards online explanations on data streams","author":[{"last_name":"Fumagalli","full_name":"Fumagalli, Fabian","first_name":"Fabian"},{"first_name":"Maximilian","last_name":"Muschalik","full_name":"Muschalik, Maximilian"},{"full_name":"Hüllermeier, Eyke","last_name":"Hüllermeier","first_name":"Eyke"},{"first_name":"Barbara","full_name":"Hammer, Barbara","last_name":"Hammer"}],"date_created":"2023-11-10T14:15:36Z","date_updated":"2023-11-10T14:24:27Z","publisher":"Springer Science and Business Media LLC","citation":{"ama":"Fumagalli F, Muschalik M, Hüllermeier E, Hammer B. Incremental permutation feature importance (iPFI): towards online explanations on data streams. <i>Machine Learning</i>. Published online 2023. doi:<a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>","ieee":"F. Fumagalli, M. Muschalik, E. Hüllermeier, and B. Hammer, “Incremental permutation feature importance (iPFI): towards online explanations on data streams,” <i>Machine Learning</i>, 2023, doi: <a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>.","chicago":"Fumagalli, Fabian, Maximilian Muschalik, Eyke Hüllermeier, and Barbara Hammer. “Incremental Permutation Feature Importance (IPFI): Towards Online Explanations on Data Streams.” <i>Machine Learning</i>, 2023. <a href=\"https://doi.org/10.1007/s10994-023-06385-y\">https://doi.org/10.1007/s10994-023-06385-y</a>.","short":"F. Fumagalli, M. Muschalik, E. Hüllermeier, B. Hammer, Machine Learning (2023).","bibtex":"@article{Fumagalli_Muschalik_Hüllermeier_Hammer_2023, title={Incremental permutation feature importance (iPFI): towards online explanations on data streams}, DOI={<a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>}, journal={Machine Learning}, publisher={Springer Science and Business Media LLC}, author={Fumagalli, Fabian and Muschalik, Maximilian and Hüllermeier, Eyke and Hammer, Barbara}, year={2023} }","mla":"Fumagalli, Fabian, et al. “Incremental Permutation Feature Importance (IPFI): Towards Online Explanations on Data Streams.” <i>Machine Learning</i>, Springer Science and Business Media LLC, 2023, doi:<a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>.","apa":"Fumagalli, F., Muschalik, M., Hüllermeier, E., &#38; Hammer, B. (2023). Incremental permutation feature importance (iPFI): towards online explanations on data streams. <i>Machine Learning</i>. <a href=\"https://doi.org/10.1007/s10994-023-06385-y\">https://doi.org/10.1007/s10994-023-06385-y</a>"},"year":"2023","publication_status":"published","publication_identifier":{"issn":["0885-6125","1573-0565"]},"language":[{"iso":"eng"}],"keyword":["Artificial Intelligence","Software"],"user_id":"55908","department":[{"_id":"424"},{"_id":"660"}],"_id":"48777","status":"public","abstract":[{"text":"<jats:title>Abstract</jats:title><jats:p>Explainable artificial intelligence has mainly focused on static learning scenarios so far. We are interested in dynamic scenarios where data is sampled progressively, and learning is done in an incremental rather than a batch mode. We seek efficient incremental algorithms for computing feature importance (FI). Permutation feature importance (PFI) is a well-established model-agnostic measure to obtain global FI based on feature marginalization of absent features. We propose an efficient, model-agnostic algorithm called iPFI to estimate this measure incrementally and under dynamic modeling conditions including concept drift. We prove theoretical guarantees on the approximation quality in terms of expectation and variance. To validate our theoretical findings and the efficacy of our approaches in incremental scenarios dealing with streaming data rather than traditional batch settings, we conduct multiple experimental studies on benchmark data with and without concept drift.</jats:p>","lang":"eng"}],"type":"journal_article","publication":"Machine Learning"},{"doi":"10.1007/s10994-023-06385-y","volume":112,"author":[{"first_name":"Fabian","last_name":"Fumagalli","full_name":"Fumagalli, Fabian"},{"first_name":"Maximilian","last_name":"Muschalik","full_name":"Muschalik, Maximilian"},{"first_name":"Eyke","last_name":"Hüllermeier","full_name":"Hüllermeier, Eyke"},{"last_name":"Hammer","full_name":"Hammer, Barbara","first_name":"Barbara"}],"date_updated":"2025-01-16T16:20:12Z","page":"4863-4903","intvolume":"       112","citation":{"mla":"Fumagalli, Fabian, et al. “Incremental Permutation Feature Importance (IPFI): Towards Online Explanations on Data Streams.” <i>Machine Learning</i>, vol. 112, no. 12, Springer Science and Business Media LLC, 2023, pp. 4863–903, doi:<a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>.","short":"F. Fumagalli, M. Muschalik, E. Hüllermeier, B. Hammer, Machine Learning 112 (2023) 4863–4903.","bibtex":"@article{Fumagalli_Muschalik_Hüllermeier_Hammer_2023, title={Incremental permutation feature importance (iPFI): towards online explanations on data streams}, volume={112}, DOI={<a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>}, number={12}, journal={Machine Learning}, publisher={Springer Science and Business Media LLC}, author={Fumagalli, Fabian and Muschalik, Maximilian and Hüllermeier, Eyke and Hammer, Barbara}, year={2023}, pages={4863–4903} }","apa":"Fumagalli, F., Muschalik, M., Hüllermeier, E., &#38; Hammer, B. (2023). Incremental permutation feature importance (iPFI): towards online explanations on data streams. <i>Machine Learning</i>, <i>112</i>(12), 4863–4903. <a href=\"https://doi.org/10.1007/s10994-023-06385-y\">https://doi.org/10.1007/s10994-023-06385-y</a>","ama":"Fumagalli F, Muschalik M, Hüllermeier E, Hammer B. Incremental permutation feature importance (iPFI): towards online explanations on data streams. <i>Machine Learning</i>. 2023;112(12):4863-4903. doi:<a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>","chicago":"Fumagalli, Fabian, Maximilian Muschalik, Eyke Hüllermeier, and Barbara Hammer. “Incremental Permutation Feature Importance (IPFI): Towards Online Explanations on Data Streams.” <i>Machine Learning</i> 112, no. 12 (2023): 4863–4903. <a href=\"https://doi.org/10.1007/s10994-023-06385-y\">https://doi.org/10.1007/s10994-023-06385-y</a>.","ieee":"F. Fumagalli, M. Muschalik, E. Hüllermeier, and B. Hammer, “Incremental permutation feature importance (iPFI): towards online explanations on data streams,” <i>Machine Learning</i>, vol. 112, no. 12, pp. 4863–4903, 2023, doi: <a href=\"https://doi.org/10.1007/s10994-023-06385-y\">10.1007/s10994-023-06385-y</a>."},"publication_identifier":{"issn":["0885-6125","1573-0565"]},"publication_status":"published","department":[{"_id":"660"}],"user_id":"93420","_id":"50262","project":[{"name":"TRR 318 - C3: TRR 318 - Subproject C3","_id":"126"},{"_id":"117","name":"TRR 318 - C: TRR 318 - Project Area C"},{"name":"TRR 318: TRR 318 - Erklärbarkeit konstruieren","_id":"109","grant_number":"438445824"}],"status":"public","type":"journal_article","title":"Incremental permutation feature importance (iPFI): towards online explanations on data streams","date_created":"2024-01-05T21:52:28Z","publisher":"Springer Science and Business Media LLC","year":"2023","issue":"12","language":[{"iso":"eng"}],"keyword":["Artificial Intelligence","Software"],"abstract":[{"lang":"eng","text":"<jats:title>Abstract</jats:title><jats:p>Explainable artificial intelligence has mainly focused on static learning scenarios so far. We are interested in dynamic scenarios where data is sampled progressively, and learning is done in an incremental rather than a batch mode. We seek efficient incremental algorithms for computing feature importance (FI). Permutation feature importance (PFI) is a well-established model-agnostic measure to obtain global FI based on feature marginalization of absent features. We propose an efficient, model-agnostic algorithm called iPFI to estimate this measure incrementally and under dynamic modeling conditions including concept drift. We prove theoretical guarantees on the approximation quality in terms of expectation and variance. To validate our theoretical findings and the efficacy of our approaches in incremental scenarios dealing with streaming data rather than traditional batch settings, we conduct multiple experimental studies on benchmark data with and without concept drift.</jats:p>"}],"publication":"Machine Learning"},{"language":[{"iso":"eng"}],"_id":"25035","user_id":"38261","abstract":[{"text":"<jats:title>Abstract</jats:title><jats:p>The efficiency of state-of-the-art algorithms for the dueling bandits problem is essentially due to a clever exploitation of (stochastic) transitivity properties of pairwise comparisons: If one arm is likely to beat a second one, which in turn is likely to beat a third one, then the first is also likely to beat the third one. By now, however, there is no way to test the validity of corresponding assumptions, although this would be a key prerequisite to guarantee the meaningfulness of the results produced by an algorithm. In this paper, we investigate the problem of testing different forms of stochastic transitivity in an online manner. We derive lower bounds on the expected sample complexity of any sequential hypothesis testing algorithm for various forms of stochastic transitivity, thereby providing additional motivation to focus on weak stochastic transitivity. To this end, we introduce an algorithmic framework for the dueling bandits problem, in which the statistical validity of weak stochastic transitivity can be tested, either actively or passively, based on a multiple binomial hypothesis test. Moreover, by exploiting a connection between weak stochastic transitivity and graph theory, we suggest an enhancement to further improve the efficiency of the testing algorithm. In the active setting, both variants achieve an expected sample complexity that is optimal up to a logarithmic factor.</jats:p>","lang":"eng"}],"status":"public","publication":"Machine Learning","type":"journal_article","title":"On testing transitivity in online preference learning","doi":"10.1007/s10994-021-06026-2","date_updated":"2022-01-06T06:56:44Z","date_created":"2021-09-24T11:07:54Z","author":[{"full_name":"Haddenhorst, Björn","last_name":"Haddenhorst","first_name":"Björn"},{"first_name":"Viktor","last_name":"Bengs","full_name":"Bengs, Viktor"},{"first_name":"Eyke","last_name":"Hüllermeier","full_name":"Hüllermeier, Eyke"}],"year":"2021","page":"2063-2084","citation":{"mla":"Haddenhorst, Björn, et al. “On Testing Transitivity in Online Preference Learning.” <i>Machine Learning</i>, 2021, pp. 2063–84, doi:<a href=\"https://doi.org/10.1007/s10994-021-06026-2\">10.1007/s10994-021-06026-2</a>.","bibtex":"@article{Haddenhorst_Bengs_Hüllermeier_2021, title={On testing transitivity in online preference learning}, DOI={<a href=\"https://doi.org/10.1007/s10994-021-06026-2\">10.1007/s10994-021-06026-2</a>}, journal={Machine Learning}, author={Haddenhorst, Björn and Bengs, Viktor and Hüllermeier, Eyke}, year={2021}, pages={2063–2084} }","short":"B. Haddenhorst, V. Bengs, E. Hüllermeier, Machine Learning (2021) 2063–2084.","apa":"Haddenhorst, B., Bengs, V., &#38; Hüllermeier, E. (2021). On testing transitivity in online preference learning. <i>Machine Learning</i>, 2063–2084. <a href=\"https://doi.org/10.1007/s10994-021-06026-2\">https://doi.org/10.1007/s10994-021-06026-2</a>","ama":"Haddenhorst B, Bengs V, Hüllermeier E. On testing transitivity in online preference learning. <i>Machine Learning</i>. Published online 2021:2063-2084. doi:<a href=\"https://doi.org/10.1007/s10994-021-06026-2\">10.1007/s10994-021-06026-2</a>","ieee":"B. Haddenhorst, V. Bengs, and E. Hüllermeier, “On testing transitivity in online preference learning,” <i>Machine Learning</i>, pp. 2063–2084, 2021, doi: <a href=\"https://doi.org/10.1007/s10994-021-06026-2\">10.1007/s10994-021-06026-2</a>.","chicago":"Haddenhorst, Björn, Viktor Bengs, and Eyke Hüllermeier. “On Testing Transitivity in Online Preference Learning.” <i>Machine Learning</i>, 2021, 2063–84. <a href=\"https://doi.org/10.1007/s10994-021-06026-2\">https://doi.org/10.1007/s10994-021-06026-2</a>."},"publication_identifier":{"issn":["0885-6125","1573-0565"]},"publication_status":"published"},{"date_updated":"2022-01-06T06:59:21Z","oa":"1","author":[{"full_name":"Mohr, Felix","last_name":"Mohr","first_name":"Felix"},{"last_name":"Wever","orcid":" https://orcid.org/0000-0001-9782-6818","full_name":"Wever, Marcel Dominik","id":"33176","first_name":"Marcel Dominik"},{"last_name":"Hüllermeier","id":"48129","full_name":"Hüllermeier, Eyke","first_name":"Eyke"}],"doi":"10.1007/s10994-018-5735-z","conference":{"location":"Dublin, Ireland","end_date":"2018-09-14","start_date":"2018-09-10","name":"European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases"},"main_file_link":[{"open_access":"1","url":"https://rdcu.be/3Nc2"}],"publication_identifier":{"eissn":["1573-0565"],"issn":["0885-6125"]},"has_accepted_license":"1","publication_status":"epub_ahead","page":"1495-1515","citation":{"ieee":"F. Mohr, M. D. Wever, and E. Hüllermeier, “ML-Plan: Automated Machine Learning via Hierarchical Planning,” <i>Machine Learning</i>, pp. 1495–1515, 2018, doi: <a href=\"https://doi.org/10.1007/s10994-018-5735-z\">10.1007/s10994-018-5735-z</a>.","chicago":"Mohr, Felix, Marcel Dominik Wever, and Eyke Hüllermeier. “ML-Plan: Automated Machine Learning via Hierarchical Planning.” <i>Machine Learning</i>, 2018, 1495–1515. <a href=\"https://doi.org/10.1007/s10994-018-5735-z\">https://doi.org/10.1007/s10994-018-5735-z</a>.","ama":"Mohr F, Wever MD, Hüllermeier E. ML-Plan: Automated Machine Learning via Hierarchical Planning. <i>Machine Learning</i>. Published online 2018:1495-1515. doi:<a href=\"https://doi.org/10.1007/s10994-018-5735-z\">10.1007/s10994-018-5735-z</a>","mla":"Mohr, Felix, et al. “ML-Plan: Automated Machine Learning via Hierarchical Planning.” <i>Machine Learning</i>, Springer, 2018, pp. 1495–515, doi:<a href=\"https://doi.org/10.1007/s10994-018-5735-z\">10.1007/s10994-018-5735-z</a>.","bibtex":"@article{Mohr_Wever_Hüllermeier_2018, title={ML-Plan: Automated Machine Learning via Hierarchical Planning}, DOI={<a href=\"https://doi.org/10.1007/s10994-018-5735-z\">10.1007/s10994-018-5735-z</a>}, journal={Machine Learning}, publisher={Springer}, author={Mohr, Felix and Wever, Marcel Dominik and Hüllermeier, Eyke}, year={2018}, pages={1495–1515} }","short":"F. Mohr, M.D. Wever, E. Hüllermeier, Machine Learning (2018) 1495–1515.","apa":"Mohr, F., Wever, M. D., &#38; Hüllermeier, E. (2018). ML-Plan: Automated Machine Learning via Hierarchical Planning. <i>Machine Learning</i>, 1495–1515. <a href=\"https://doi.org/10.1007/s10994-018-5735-z\">https://doi.org/10.1007/s10994-018-5735-z</a>"},"_id":"3510","project":[{"name":"SFB 901","_id":"1"},{"name":"SFB 901 - Project Area B","_id":"3"},{"name":"SFB 901 - Subproject B2","_id":"10"},{"name":"Computing Resources Provided by the Paderborn Center for Parallel Computing","_id":"52"}],"department":[{"_id":"355"},{"_id":"34"},{"_id":"7"},{"_id":"26"}],"user_id":"5786","article_type":"original","file_date_updated":"2018-11-02T15:32:16Z","type":"journal_article","status":"public","publisher":"Springer","date_created":"2018-07-08T14:06:14Z","title":"ML-Plan: Automated Machine Learning via Hierarchical Planning","year":"2018","keyword":["AutoML","Hierarchical Planning","HTN planning","ML-Plan"],"ddc":["000"],"language":[{"iso":"eng"}],"publication":"Machine Learning","abstract":[{"lang":"eng","text":"Automated machine learning (AutoML) seeks to automatically select, compose, and parametrize machine learning algorithms, so as to achieve optimal performance on a given task (dataset). Although current approaches to AutoML have already produced impressive results, the field is still far from mature, and new techniques are still being developed. In this paper, we present ML-Plan, a new approach to AutoML based on hierarchical planning. To highlight the potential of this approach, we compare ML-Plan to the state-of-the-art frameworks Auto-WEKA, auto-sklearn, and TPOT. In an extensive series of experiments, we show that ML-Plan is highly competitive and often outperforms existing approaches."}],"file":[{"date_updated":"2018-11-02T15:32:16Z","date_created":"2018-11-02T15:32:16Z","creator":"ups","file_size":1070937,"access_level":"closed","file_name":"ML-PlanAutomatedMachineLearnin.pdf","file_id":"5306","content_type":"application/pdf","success":1,"relation":"main_file"}]}]
