Hubert Baniecki

University of WarsawMI2.AI, Warsaw University of Technology

h.baniecki (at) uw.edu.pl

I am a 4th (last) year PhD student in Computer Science at the University of Warsaw, advised by Przemyslaw Biecek. During my PhD, I have been a visiting researcher at LMU Munich, hosted by Bernd Bischl (2024) and Eyke Hüllermeier (2025). Previously, I earned a Master’s degree in Data Science from Warsaw University of Technology.

My research focuses on machine learning interpretability and explainable AI:

  • interpreting vision–language models like CLIP (ICML 2025, NeurIPS 2025),
  • statistical foundations of explainable machine learning (ECML 2024, ICLR 2025),
  • open-source software and benchmarks (JMLR 2021, NeurIPS 2024),
  • applications in medicine and beyond (PNAS 2024, WACV 2025).

I actively contribute to the academic community by serving as a reviewer for conferences like NeurIPS, ECML, ICLR, with their workshops on interpretability/xAI, and journals like the Journal of Machine Learning Research, Machine Learning, Nature Communications.

I am on the job market for positions starting in 2026/2027; feel free to contact me if there is an opportunity that fits.


recent news [previous]

2025 Sep A paper Explaining similarity in vision-language encoders with weighted Banzhaf interactions is accepted at NeurIPS 2025.
2025 Sep A paper Birds look like cars: Adversarial analysis of intrinsically interpretable deep learning is accepted for publication in the Machine Learning journal.
2025 May Foundation for Polish Science awarded me the START scholarship for young scientists.
2025 May A paper Interpreting CLIP with hierarchical sparse autoencoders is accepted at ICML 2025.
2025 Mar I stay in Germany until April for a 1-month research visit at LMU Munich hosted by Eyke Hüllermeier.
2025 Jan A paper Efficient and accurate explanation estimation with distribution compression is accepted as a Spotlight at ICLR 2025 (notable 5% of submissions).
2024 Nov A paper Increasing phosphorus loss despite widespread concentration decline in US rivers is published in the Proceedings of the National Academy of Sciences.

selected publications [full list]

    1. Explaining similarity in vision-language encoders with weighted Banzhaf interactions
      H. Baniecki, M. Muschalik, F. Fumagalli, B. Hammer, E. Hüllermeier, P. Biecek
      Advances in Neural Information Processing Systems (NeurIPS), 2025
      We introduce faithful interaction explanations of CLIP and SigLIP models (FIxLIP), offering a unique game-theoretic perspective on interpreting image–text similarity predictions.
    2. Interpreting CLIP with hierarchical sparse autoencoders
      V. Zaigrajew, H. Baniecki, P. Biecek
      International Conference on Machine Learning (ICML), 2025
      We introduce the Matryoshka sparse autoencoder (MSAE) that establishes a state-of-the-art Pareto frontier between reconstruction quality and sparsity for interpreting CLIP models.
    3. ICLR Spotlight
      Efficient and accurate explanation estimation with distribution compression
      H. Baniecki, G. Casalicchio, B. Bischl, P. Biecek
      International Conference on Learning Representations (ICLR), 2025 (Spotlight)
      We introduce compress then explain (CTE) as a new paradigm for sample-efficient estimation of post-hoc explanations, including feature attributions, importance, and effects.
    4. WACV Oral
      Aggregated attributions for explanatory analysis of 3D segmentation models
      M. Chrabaszcz=, H. Baniecki=, P. Komorowski, S. Plotka, P. Biecek
      IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2025 (Oral)
      We discover knowledge acquired by the TotalSegmentator foundation model trained to segment all anatomical structures in computed tomography medical images.
    5. shapiq: Shapley interactions for machine learning
      M. Muschalik, H. Baniecki, F. Fumagalli, P. Kolpaczki, B. Hammer, E. Hullermeier
      Advances in Neural Information Processing Systems (NeurIPS), 2024
      We develop {shapiq}, an open-source Python package that implements several algorithms and benchmarks for efficiently approximating game-theoretic attribution and interaction indices.
    6. Increasing phosphorus loss despite widespread concentration decline in US rivers
      W. Zhi, H. Baniecki, J. Liu, E. Boyer, C. Shen, G. Shenk, X. Liu, L. Li
      Proceedings of the National Academy of Sciences, 2024
      We reveal a paradox in US rivers with deep learning: phosphorus concentration is down over the last 40 years, particularly in urban areas, but total phosphorus loss is up due to climate change.
    7. On the robustness of global feature effect explanations
      H. Baniecki, G. Casalicchio, B. Bischl, P. Biecek
      European Conference on Machine Learning (ECML PKDD), 2024
      Theoretical bounds for the robustness of feature effects to data and model perturbations.
    8. dalex: Responsible machine learning with interactive explainability and fairness in Python
      H. Baniecki, W. Kretowicz, P. Piatyszek, J. Wisniewski, P. Biecek
      Journal of Machine Learning Research, 2021
      2022 John M. Chambers Statistical Software Award by the American Statistical Association