Florian Mai

Welcome

I am the Junior Research Group Leader of the mAI-alignment group at the CAISA group at University of Bonn as part of The Lamarr Institute for Machine Learning and Artificial Intelligence.

My current research focuses on AI alignment and safety issues, exploring how to ensure that current and future advanced AI systems are beneficial and safe for humanity.

Visit the lab page →

Read more about my research and background →

News

14-12-2025

Our workshop paper “Pluralistic AI Alignment: A Cross-Cultural Pilot Survey” will be presented at the Second Workshop on Language Models for Underserved Communities (LM4UC).

27-10-2025

The AI alignment lab at Uni Bonn has started! Learn more on the course page.

13-10-2025

New preprint! Leonard Dung and Florian Mai analyze AI alignment strategies from a risk perspective and compare overlaps in failure modes across alignment techniques. Read the preprint on arXiv.

21-08-2025

Our JQL paper has been accepted at EMNLP 2025! Read the preprint on arXiv.

View all news →

Selected Publications

AI Alignment Strategies from a Risk Perspective: Independent Safety Mechanisms or Shared Failures?
IASEAI'26: International Association for Safe and Ethical AI Conference, 2026
We analyze overlap in failure modes across alignment techniques to assess the limits of defense-in-depth risk mitigation.
Superalignment with Dynamic Human Values
ICLR 2025 Workshop on Bidirectional Human-AI Alignment (BiAlign), 2025
This paper sketches a roadmap for training a superhuman reasoning model to decompose complex tasks into subtasks amenable to human-level guidance, addressing scalable oversight and dynamic human values in AI alignment.
Learning to Plan for Language Modeling from Unlabeled Data
COLM, 2024
We propose a method to learn planning for language modeling using unlabeled data.

View all publications →