NeurIPS
Linear Transformers are Versatile In-Context Learners
Max Vladymyrov, Johannes von Oswald, Mark Sandler, Rong Ge
38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)
ICML Workshop
Efficient Linear System Solver with Transformers
Max Vladymyrov, Johannes von Oswald, Nolan Miller, Mark Sandler
AI for Math Workshop, ICML 2024
Preprint
Narrowing the Focus: Learned Optimizers for Pretrained Models
Gus Kristiansen, Mark Sandler, Andrey Zhmoginov, Nolan Miller, Anirudh Goyal, Jihwan Lee, Max Vladymyrov
arXiv:2408.09310
ICML Workshop
Learning and Unlearning of Fabricated Knowledge in Language Models
Chen Sun, Nolan Miller, Andrey Zhmoginov, Max Vladymyrov, Mark Sandler
Mechanistic Interpretability Workshop, ICML 2024
ICML Workshop
Learning Fast and Slow: Representations for In-Context Weight Modulation
Andrey Zhmoginov, Jihwan Lee, Max Vladymyrov, Mark Sandler
Workshop on In-Context Learning, ICML 2024
TMLR
Continual HyperTransformer: A Meta-Learner for Continual Few-Shot Learning
Max Vladymyrov, Andrey Zhmoginov, Mark Sandler
Transactions on Machine Learning Research, 2024
Preprint
Uncovering Mesa-Optimization Algorithms in Transformers
Johannes von Oswald, Eyvind Niklasson, Maximilian Schlegel, Seijin Kobayashi, Nicolas Zucchet, Nino Scherrer, Nolan Miller, Mark Sandler, Max Vladymyrov, Razvan Pascanu, João Sacramento
arXiv:2309.05858
ICML
Transformers Learn In-Context by Gradient Descent
Johannes von Oswald, Eyvind Niklasson, Ettore Randazzo, João Sacramento, Alexander Mordvintsev, Andrey Zhmoginov, Max Vladymyrov
International Conference on Machine Learning (ICML 2023)
Preprint
Training Trajectories, Mini-Batch Losses and the Curious Role of the Learning Rate
Mark Sandler, Andrey Zhmoginov, Max Vladymyrov, Nolan Miller
arXiv preprint arXiv:2301.02312
CVPR
Decentralized Learning with Multi-Headed Distillation
Andrey Zhmoginov, Mark Sandler, Nolan Miller, Gus Kristiansen, Max Vladymyrov
Computer Vision and Pattern Recognition (CVPR 2023)
ICML
HyperTransformer: Model Generation for Supervised and Semi-Supervised Few-Shot Learning
Andrey Zhmoginov, Mark Sandler, Max Vladymyrov
International Conference on Machine Learning (ICML 2022)
ICLR
GradMax: Growing Neural Networks Using Gradient Information
Utku Evci, Bart van Merrienboer, Thomas Unterthiner, Max Vladymyrov, Fabian Pedregosa
International Conference on Learning Representations (ICLR 2022)
CVPR
Fine-Tuning Image Transformers Using Learnable Memory
Mark Sandler, Andrey Zhmoginov, Max Vladymyrov, Andrew Jackson
Computer Vision and Pattern Recognition (CVPR 2022)
JMLR
Underspecification Presents Challenges for Credibility in Modern Machine Learning
Alexander D'Amour, Katherine Heller, Dan Moldovan, Ben Adlam, Babak Alipanahi, Alex Beutel, Christina Chen, Jonathan Deaton, Jacob Eisenstein, Matthew D. Hoffman, Farhad Hormozdiari, Neil Houlsby, Shaobo Hou, Ghassen Jerfel, Alan Karthikesalingam, Mario Lucic, Yian Ma, Cory McLean, Diana Mincu, Akinori Mitani, Andrea Montanari, Zachary Nado, Vivek Natarajan, Christopher Nielson, Thomas F. Osborne, Rajiv Raman, Kim Ramasamy, Rory Sayres, Jessica Schrouff, Martin Seneviratne, Shannon Sequeira, Harini Suresh, Victor Veitch, Max Vladymyrov, Xuezhi Wang, Kellie Webster, Steve Yadlowsky, Taedong Yun, Xiaohua Zhai, D. Sculley
Journal of Machine Learning Research, 2022
ICML
Meta-Learning Bidirectional Update Rules
Mark Sandler, Max Vladymyrov, Andrey Zhmoginov, Nolan Miller, Andrew Jackson, Tom Madams, Blaise Agüera y Arcas
38th International Conference on Machine Learning (ICML 2021), pp. 9288-9300
NeurIPS
No Pressure! Addressing the Problem of Local Minima in Manifold Learning Algorithms
Max Vladymyrov
33th Annual Conference on Neural Information Processing Systems (NeurIPS 2019), pp. 678-687
IJCNN
Fast, Accurate Spectral Clustering Using Locally Linear Landmarks
Max Vladymyrov, Miguel Á. Carreira-Perpiñán
30th International Joint Conference on Neural Networks (IJCNN 2017), pp. 3870-3879
ICML
The Variational Nyström Method for Large-Scale Spectral Problems
Max Vladymyrov, Miguel Á. Carreira-Perpiñán
33th International Conference on Machine Learning (ICML 2016)
NeurIPS
A Fast, Universal Algorithm to Learn Parametric Nonlinear Embeddings
Miguel Á. Carreira-Perpiñán, Max Vladymyrov
29th Annual Conference on Neural Information Processing Systems (NIPS 2015), pp. 253-261
AISTATS
Linear-Time Training of Nonlinear Low-Dimensional Embeddings
Max Vladymyrov, Miguel Á. Carreira-Perpiñán
17th International Conference on Artificial Intelligence and Statistics (AISTATS 2014), pp. 968-977
ECML-PKDD
Locally Linear Landmarks for Large-Scale Manifold Learning
Max Vladymyrov, Miguel Á. Carreira-Perpiñán
24th European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD 2013), pp. 256-271
ICML
Entropic Affinities: Properties and Efficient Numerical Computation
Max Vladymyrov, Miguel Á. Carreira-Perpiñán
30th International Conference on Machine Learning (ICML 2013), pp. 477-485
ICML
Partial-Hessian Strategies for Fast Learning of Nonlinear Embeddings
Max Vladymyrov, Miguel Á. Carreira-Perpiñán
29th International Conference on Machine Learning (ICML 2012), pp. 345-352