Publications

Selected Publications

Superalignment with Dynamic Human Values

Published in ICLR 2025 Workshop on Bidirectional Human-AI Alignment (BiAlign), 2025

This paper sketches a roadmap for training a superhuman reasoning model to decompose complex tasks into subtasks amenable to human-level guidance, addressing scalable oversight and dynamic human values in AI alignment.

Recommended citation: Florian Mai, David Kaczér, Nicholas Kluge Corrêa, Lucie Flek. (2025). "Superalignment with Dynamic Human Values." ICLR 2025 Workshop on Bidirectional Human-AI Alignment (BiAlign).

All Peer-Reviewed Publications

Open-Source Conversational AI with SpeechBrain 1.0

Published in JMLR, MLOSS, 2024

We present SpeechBrain 1.0, an open-source toolkit for speech and language processing.

Recommended citation: Mirco Ravanelli, Titouan Parcollet, Adel Moumen, Sylvain de Langen, Cem Subakan, Peter Plantinga, Yingzhi Wang, Pooneh Mousavi, Luca Della Libera, Artem Ploujnikov, Francesco Paissan, Davide Borra, Salah Zaiem, Zeyu Zhao, Shucong Zhang, Georgios Karakasidis, Sung-Lin Yeh, Pierre Champion, Aku Rouhe, Rudolf Braun, Florian Mai, Juan Zuluaga-Gomez, Seyed Mahed Mousavi, Andreas Nautsch, Xuechen Liu, Sangeet Sagar, Jarod Duret, Salima Mdhaffar, Gaelle Laperriere, Mickael Rouvier, Renato De Mori, Yannick Esteve. (2024). "Open-Source Conversational AI with SpeechBrain 1.0." arXiv. https://www.jmlr.org/papers/volume25/24-0991/24-0991.pdf

Using Deep Learning for Title-Based Semantic Subject Indexing to Reach Competitive Performance to Full-Text

Published in JCDL, 2018

We show that title-based text classification can outperform classification based on the full text due to the larger number of available training data.

Recommended citation: Florian Mai, Lukas Galke and Ansgar Scherp. (2018). "Using Deep Learning for Title-Based Semantic Subject Indexing to Reach Competitive Performance to Full-Text." JCDL 2018. https://arxiv.org/abs/1801.06717

Preprints

End-to-end Planner Training for Language Modeling

Published in arXiv, 2024

We propose a differentiable method for joint fine-tuning of language models with planning modules by using predicted label probabilities as mixing weights.

Recommended citation: Nathan Cornille, Florian Mai, Jingyuan Sun, Marie-Francine Moens. (2024). "End-to-end Planner Training for Language Modeling." arXiv:2410.12492. https://arxiv.org/abs/2410.12492

Learning to Plan Long-Term for Language Modeling

Published in arXiv, 2024

We propose a planner that predicts a latent plan for many sentences into the future, allowing language models to trade computation time for better next token prediction accuracy.

Recommended citation: Florian Mai, Nathan Cornille, Marie-Francine Moens. (2024). "Learning to Plan Long-Term for Language Modeling." arXiv:2409.00070. https://arxiv.org/abs/2409.00070