Florian Mai
Welcome
I am the Junior Research Group Leader of the mAI-alignment group at the CAISA group at University of Bonn as part of The Lamarr Institute for Machine Learning and Artificial Intelligence.
My current research focuses on AI alignment and safety issues, exploring how to ensure that current and future advanced AI systems are beneficial and safe for humanity.
Read more about my research and background →
News
Our paper “AI Alignment Strategies from a Risk Perspective: Independent Safety Mechanisms or Shared Failures?” will be presented at IASEAI’26.
Our workshop paper “Pluralistic AI Alignment: A Cross-Cultural Pilot Survey” will be presented at the Second Workshop on Language Models for Underserved Communities (LM4UC).
The AI alignment lab at Uni Bonn has started! Learn more on the course page.
New preprint! Leonard Dung and Florian Mai analyze AI alignment strategies from a risk perspective and compare overlaps in failure modes across alignment techniques. Read the preprint on arXiv.
Our JQL paper has been accepted at EMNLP 2025! Read the preprint on arXiv.
Selected Publications
IASEAI'26: International Association for Safe and Ethical AI Conference, 2026
We analyze overlap in failure modes across alignment techniques to assess the limits of defense-in-depth risk mitigation.
ICLR 2025 Workshop on Bidirectional Human-AI Alignment (BiAlign), 2025
This paper sketches a roadmap for training a superhuman reasoning model to decompose complex tasks into subtasks amenable to human-level guidance, addressing scalable oversight and dynamic human values in AI alignment.
COLM, 2024
We propose a method to learn planning for language modeling using unlabeled data.
