Follow
Michael E. Sander
Michael E. Sander
Other namesMichael Eli Sander
Google DeepMind
Verified email at google.com - Homepage
Title
Cited by
Cited by
Year
Vision Transformers provably learn spatial structure
S Jelassi, ME Sander, Y Li
Advances in Neural Information Processing Systems 35, 2022
1012022
Sinkformers: Transformers with doubly stochastic attention
ME Sander, P Ablin, M Blondel, G Peyré
International Conference on Artificial Intelligence and Statistics, 2022
832022
Momentum Residual Neural Networks
ME Sander, P Ablin, M Blondel, G Peyré
International Conference on Machine Learning 139, 9276-9287, 2021
712021
Do Residual Neural Networks discretize Neural Ordinary Differential Equations?
ME Sander, P Ablin, G Peyré
Advances in Neural Information Processing Systems 35, 2022
272022
Fast, Differentiable and Sparse Top-k: a Convex Analysis Perspective
ME Sander, J Puigcerver, J Djolonga, G Peyré, M Blondel
International Conference on Machine Learning, 2023
212023
Implicit regularization of deep residual networks towards neural ODEs
P Marion, YH Wu, ME Sander, G Biau
International Conference on Learning Representations, 2024
172024
How do Transformers perform In-Context Autoregressive Learning?
ME Sander, R Giryes, T Suzuki, M Blondel, G Peyré
International Conference on Machine Learning, 2024
132024
Towards understanding the universality of transformers for next-token prediction
ME Sander, G Peyré
International Conference on Learning Representations, 2025
22025
Joint learning of energy-based models and their partition function
ME Sander, V Roulet, T Liu, M Blondel
arXiv preprint arXiv:2501.18528, 2025
12025
Loss Functions and Operators Generated by f-Divergences
V Roulet, T Liu, N Vieillard, ME Sander, M Blondel
arXiv preprint arXiv:2501.18537, 2025
2025
The system can't perform the operation now. Try again later.
Articles 1–10