Tassilo Klein, Ph.D.

tjk

LinkedIn

Google scholar

GitHub

Tassilo J. Klein, Ph.D.

Director of Machine Learning Research & Principal Scientist

📍 SAP AI CTO Office | 💡 LLMs, Foundation Models, Structured Data AI
🌐 LinkedInGoogle Scholar

Building intelligent systems that connect language, knowledge, and enterprise data.


🧠 About Me

I am a Director of Machine Learning Research & Principal Scientist in the SAP AI CTO Office, where I lead research on large language models (LLMs), foundation models, and structured data AI.
My mission is to advance how enterprises leverage AI — by designing systems that can reason over complex structured data, adapt to limited supervision, and integrate domain-specific knowledge with the general capabilities of LLMs.

I oversee strategic research programs that bridge fundamental machine learning and enterprise-scale deployment, turning scientific innovation into robust, trustworthy AI platforms.
This includes developing foundation models for structured and relational data, knowledge-augmented LLMs, and agentic AI systems that coordinate decision-making and automation across business processes.

Over the past decade, I’ve built and led multidisciplinary research teams translating state-of-the-art methods into product-ready technologies — shaping how AI is applied in large-scale enterprise systems.
Before joining SAP, I conducted postdoctoral research at MIT CSAIL and Harvard Medical School, focusing on representation learning and multimodal modeling for medical imaging.
I hold a Ph.D. in Computer Science (summa cum laude) from the Technical University of Munich (TUM) and am a member of the European Laboratory for Learning and Intelligent Systems (ELLIS).


🔬 Research Focus


📄 Selected Publications & Projects

[2025.07] - SALT officially integrated into RelBench.

[2025.05] - New pre-print available on foundation models for tabular data in enterprises

arXiv

[2025.05] - Paper accepted at ACL 2025 — Contrastive Perplexity for Controlled Generation: An Application in Detoxifying Large Language Model

arXiv

[2024.10] - Two papers accepted at the NeurIPS’24 Table Representation Learning Workshop

[2023.05] - Paper accepted at ACL 2023 on low-shot contrastive learning of sentence representations.

arXiv View on GitHub Download Model

[2022.02] Paper accepted at ACL 2022 on self-supervised sentence representation learning

arXiv View on GitHub

[2021.08] Paper at EMNLP 2021 on Contrastive Language Model Refinement for Commonsense Reasoning

arXiv View on GitHub video

[2021.08] Paper at EMNLP 2021 on Contrastive Self-Supervised Learning for Commonsense Reasoning

arXiv View on GitHub

[2021.04] Acceptance of co-organized at ICML 2021 workshop on Self-Supervised Learning for Reasoning and Perception

[2021.02] Paper accepted at IPMI 2021 on self-supervised representation learning for medical imaging (acceptance rate 30.0%)

arXiv

[2020.09] Presentation on commonsense reasoning in AI

video Medium

[2020.04] Paper accepted at ACL 2020 on contrastive self-supervised commonsense reasoning (acceptance rate of 17.6%)

arXiv View on GitHub video

[2020.02] Paper accepted at NeuroImage

arXiv

[2019.10.20] Paper on Multi-Domain Learning accepted at ICCV 2019 (acceptance rate 25.0%)

arXiv

[2019.05.14] Short-paper on commonsense reasoning accepted at ACL 2019 (acceptance rate 18.2%)

arXiv View on GitHub Open Notebook

[2019.02.25] Paper accepted at CVPR 2019 (acceptance rate 25.2%)

arXiv View on GitHub

[2017.02.01] Paper accepted at NeuroImage

arXiv View on GitHub


🧭 Vision & Collaboration

I aim to advance AI systems that reason over structured data, adapt efficiently, and integrate domain knowledge to support enterprise decision-making.
At SAP, I drive the interface between fundamental AI research and strategic enterprise transformation, shaping how LLMs evolve for real-world impact.


🤝 Community & Mentorship


Last updated — 22 October 2025