The BilAI cluster is strongly represented at the most relevant flagship conferences in the fields of artificial intelligence and machine learning. Our researchers are presenting their latest work and contributing to key discussions that shape the future of the field.
These conferences also offer a great opportunity to connect with BilAI experts, exchange ideas, and explore possibilities for collaboration.
We’re proud to share our presence at these important venues—feel free to reach out and meet us there!
The 42nd International Conference on Machine Learning (ICML)
July 13-19, 2025 | Vancouver
Accepted papers by BilAI key researchers:
A Large Recurrent Action Model: xLSTM enables Fast Inference for Robotics Tasks by Thomas Schmied, Thomas Adler, Vihang Patil, Maximilian Beck, Korbinian Pöppel, Johannes Brandstetter, Günter Klambauer, Razvan Pascanu, Sepp Hochreiter
Cache Me If You Must: Adaptive Key-Value Quantization for Large Language Models by Alina Shutova, Vladimir Malinovskii, Ivan Ermakov, Surkov Nikita, Denis Mazur, Denis Kuznedelev, Vage Egiazarian, Dan Alistarh
DEALing with image reconstruction: Deep Attentive Least squares by Mehrsa Pourya, Erich Kobler, Michael Unser, Sebastian Neumayer
Differentially Private Federated k-Means Clustering with Server-Side Data by Jonathan Scott, Christoph Lampert, David Saulpic
EvoPress: Accurate Dynamic Model Compression via Evolutionary Search by Oliver Sieberling, Denis Kuznedelev, Eldar Kurtic, Dan Alistarh
Geometry-Informed Neural Networks by Arturs Berzins, Andreas Radler, Eric Volkmann, Sebastian Sanokowski, Sepp Hochreiter, Johannes Brandstetter
Layer-wise Quantization for Quantized Optimistic Dual Averaging by Anh Duc Nguyen, Ilia Markov, Zhengqing Wu, Ali Ramezani-Kebrya, Kimon Antonakopoulos, Dan Alistarh, Volkan Cevher
Mechanistic PDE Networks for Discovery of Governing Equations by Adeel Pervez, Efstratios Gavves, Francesco Locatello
Neural Collapse Beyond the Unconstrainted Features Model: Landscape, Dynamics, and Generalization in the Mean-Field Regime by Diyuan Wu, Marco Mondelli
Optimal Decision Tree Pruning Revisited: Algorithms and Complexity by Juha Harviainen, Frank Sommer, Manuel Sorge, Stefan Szeider
QuEST: Quantized Gradient Estimation for Accurate Training of Extremely Low-Bitwith Large Language Models by Andrei Panferov, Jiale Chen, Soroush Tabesh, Mahdi Nikdan, Dan Alistarh
Spurious Correlations in High Dimensional Regression: The Roles of Regularization, Simplicity Bias and Over-Parameterization by Simone Bombari, Marco Mondelli
Test-Time Training Provably Improves Transformers as In-context Learners by Halil Alperen Gozeten, Muhammed Emrullah Ildiz, Xuechen Zhang, Mahdi Soltanolkotabi, Marco Mondelli, Samet Oymak
The Disparate Benefits of Deep Ensembles by Kajetan Schweighofer, Adrián Arnaiz-Rodríguez, Sepp Hochreiter, Nuria Oliver
xLSTM 7B: A Recurrent LLM for Fast and Efficient Inference by Maximilian Beck, Korbinian Pöppel, Phillip Lippe, Richard Kurle, Patrick Blies, Günter Klambauer, Sebastian Böck, Sepp Hochreiter
The 48th International ACM SIGIR Conference on Research and Development in Information Retrieval
July 13-18, 2025 | Padua
Accepted papers by BilAI key researchers:
Reassessing the Effectiveness of Reinforcement Learning based Recommender Systems for Sequential Recommendation by Dilina Chandika Rajapakse, Dietmar Jannach
A Worrying Reproducibility Study of Intent-Aware Recommendation Models by Faisal Shehzad, Maurizio Ferrari Dacrema, Dietmar Jannach
OnSET: Ontology and Semantic Exploration Toolkit by Benedikt Kantz, Kevin Innerebner, Peter Waldert, Stefan Lengauer, Elisabeth Lex, Tobias Schreck
The 13th International Conference on Learning Representations (ICLR)
April 24-28, 2025 | Singapore
Accepted papers by BilAI key researchers:
Benchmarking Predictive Coding Networks -- Made Simple by Luca Pinchetti, Chang Qi, Oleh Lokshyn, Cornelius Emde, Amine M'Charrak, Mufeng Tang, Simon Frieder, Bayar Menzat, Gaspard Oliviers, Rafal Bogacz, Thomas Lukasiewicz, Tommaso Salvatori
Bio-xLSTM: Generative modeling, representation and in-context learning of biological and chemical sequences by Niklas Schmidinger, Lisa Schneckenreiter, Philipp Seidl, Johannes Schimunek, Pieter-Jan Hoedt, Johannes Brandstetter, Andreas Mayr, Sohvi Luukkonen, Sepp Hochreiter, Günter Klambauer
Can LLMs Separate Instructions From Data? And What Do We Even Mean By That? by Egor Zverev, Sahar Abdelnabi, Soroush Tabesh, Mario Fritz, Christoph Lampert
FlashRNN: I/O-Aware Optimization of Traditional RNNs on modern hardware by Korbinian Pöppel, Maximilian Beck, Sepp Hochreiter
High-dimensional Analysis of Knowledge Distillation: Weak-to-Strong Generalization and Scaling Laws by Muhammed Ildiz, Halil Gozeten, Ege Taga, Marco Mondelli, Samet Oymak
How to Probe: Simple Yet Effective Techniques for Improving Post-hoc Explanations by Siddhartha Gairola, Moritz Böhle, Francesco Locatello, Bernt Schiele
Improving Uncertainty Estimation through Semantically Diverse Language Generation by Lukas Aichberger, Kajetan Schweighofer, Mykyta Ielanskyi, Sepp Hochreiter
LDAdam: Adaptive Optimization from Low-Dimensional Gradient Statistics by Thomas Robert, Mher Safaryan, Ionut-Vlad Modoranu, Dan Alistarh
MIM-Refiner: A Contrastive Learning Boost from Intermediate Pre-Trained Masked Image Modeling Representations by Benedikt Alkin, Lukas Miklautz, Sepp Hochreiter, Johannes Brandstetter
Near, far: Patch-ordering enhances vision foundation models' scene understanding by Valentinos Pariza, Mohammadreza Salehi, Gertjan J Burghouts, Francesco Locatello, Yuki Asano
Privacy-Aware Lifelong Learning by Ozan Özdenizci, Elmar Rueckert, Robert Legenstein
Scalable Discrete Diffusion Samplers: Combinatorial Optimization and Statistical Physics by Sebastian Sanokowski, Wilhelm Berghammer, Haoyu Wang, Martin Ennemoser, Sepp Hochreiter, Sebastian Lehner
Scalable Mechanistic Neural Networks by Jiale Chen, Dingling Yao, Adeel Pervez, Dan Alistarh, Francesco Locatello
Shh, don't say that! Domain Certification in LLMs by Cornelius Emde, Alasdair Paren, Preetham Arvind, Maxime Kayser, Tom Rainforth, Thomas Lukasiewicz, Philip Torr, Adel Bibi
The Computational Complexity of Positive Non-Clashing Teaching in Graphs by Robert Ganian, Liana Khazaliya, Fionn Mc Inerney, Mathis Rocton
The Journey Matters: Average Parameter Count over Pre-training Unifies Sparse and Dense Scaling Laws by Tian Jin, Ahmed Imtiaz Humayun, Utku Evci, Suvinay Subramanian, Amir Yazdanbakhsh, Dan Alistarh, Gintare Karolina Dziugaite
Towards Certification of Uncertainty Calibration under Adversarial Attacks by Cornelius Emde, Francesco Pinto, Thomas Lukasiewicz, Philip Torr, Adel Bibi
Training One-Dimensional Graph Neural Networks is NP-Hard by Robert Ganian, Mathis Rocton, Simon Wietheger
Unifying Causal Representation Learning with the Invariance Principle by Dingling Yao, Dario Rancati, Riccardo Cadei, Marco Fumero, Francesco Locatello
Vision-LSTM: xLSTM as Generic Vision Backbone by Benedikt Alkin, Maximilian Beck, Korbinian Pöppel, Sepp Hochreiter, Johannes Brandstetter
Wasserstein Distances, Neuronal Entanglement, and Sparsity by Shashata Sawmya, Linghao Kong, Ilia Markov, Dan Alistarh, Nir Shavit
Wide Neural Networks Trained with Weight Decay Provably Exhibit Neural Collapse by Arthur Jacot, Peter Súkeník, Zihan Wang, Marco Mondelli
Bilateral AI
c/o Johannes Kepler University Linz
Altenberger Straße 69
4040 Linz, Austria
office@bilateral-ai.net