I'm a machine learning researcher with a PhD from ETH Zurich, focused on multimodal generative models, self-supervised representation learning, and foundation models for real-world applications in healthcare.
Most recently, Iβve been working on scalable multimodal learning as part of the Swiss AI initiative, and developing new approaches to representation learning across healthcare and biology datasets.
- Multimodal & cross-modal generative modeling
- Vision-language and structure-function learning
- Probabilistic modeling and self-supervised representation learning
- Applications in healthcare, computational biology, and scientific discovery
A scalable VAE for heterogeneous data sources β improves representation learning across modalities. Featured at NeurIPS 2024.
Training and evaluating large-scale representation models across clinical, imaging, and omics data. Focus on missing data robustness and downstream task generalization.
A probabilistic approach to modeling group structure and importance. Published at ICLR 2023 (Spotlight) and NeurIPS 2023.
- Unity by Diversity: Improved Representation Learning in Multimodal VAEs β NeurIPS 2024
- Learning Group Importance using the Differentiable Hypergeometric Distribution β ICLR 2023 (Spotlight)
- Generalized Multimodal ELBO β ICLR 2021
- Full publication list β
Languages: Python, C++, R, Matlab
Frameworks: PyTorch, TensorFlow, Hugging Face, Scikit-learn, OpenCV, Pandas
Tools: Weights & Biases, Docker, HPC, Git
Other: Probabilistic modeling, multimodal generative models, multimodal representation learning, vision-language architectures
- Generative AI for Good, Panelist β ICLR 2024
- Unity by Diversity: VAEs in Biomedicine β UCSF, Abbasi Lab
- Scalable Multimodal VAEs β Freiburg Young Scientist AI Network
- π thomassutter.github.io
- π§ thomasmarcosutter@gmail.com
- πΌ LinkedIn
- π§ͺ Google Scholar
Thanks for stopping by!