Three papers accepted at WACV 2026 Main Conference

Applications Track:

  • Knowledge to Sight: Reasoning over Visual Attributes via Knowledge Decomposition for Abnormality Grounding
    Jun Li, Che Liu, Wenjia Bai, Mingxuan Liu, Rossella Arcucci, Cosmin Bercea, Julia Schnabel
    (https://arxiv.org/abs/2508.04572)

  • Tables Guide Vision: Learning to See the Heart through Tabular Data
    Marta Hasny, Maxime Di Folco, Keno Bressem, Julia Schnabel
    (https://arxiv.org/abs/2503.14998)

Algorithms Track:

  • Hierarchical Adaptive networks with Task vectors for Test-Time Adaptation
    Sameer Ambekar, Marta Hasny, Laura Daza, Daniel Lang, Julia Schnabel
    (https://www.arxiv.org/abs/2508.09223)



Sameer Ambekar
Sameer Ambekar
Doctoral Researcher

My research interests include Domain Generalization, Meta learning, Variational Inference

Cosmin I. Bercea
Cosmin I. Bercea
Research Scientist

I am a postdoctoral researcher specializing in vision and multimodal learning for medical image analysis, with the current focus on developing vision-language models for generative downstream tasks.

Laura Daza
Laura Daza
Research Scientist

My research interests include machine learning for medical image segmentation.

Maxime Di Folco
Maxime Di Folco
Research Scientist

My research interest is the study of the cardiac function via machine learning methods, in particular representation learning methods that aim to acquire low dimensional representation of high dimensional data. I have a strong interest in cardiac remodelling (adaptation of the heart to its environment or a disease), notably the study of the deformation and shape aspects.

Marta Hasny
Marta Hasny
Doctoral Researcher

My research interests include the application of foundation models and generative AI in cardiology.

Daniel M. Lang
Daniel M. Lang
Research Scientist

My current research focuses on the development of deep generative models for dynamic settings in cancer imaging.

Jun Li
Jun Li
Doctoral Researcher

My research interests include Vision and Language, Multi-Modal Learning, and Cross-Modality Generation.