I am a PhD student at the Vector Institute and the University of Calgary, supervised by Dr. Yani Ioannou and Dr. Rahul Krishnan. My research focuses on weight-space symmetries and their implications for optimization and generalization. I’m particularly interested in linear mode connectivity, model matching and merging, and training dynamics, as well as efficient adaptation methods for LLMs, such as low-rank representations and weight-space learning. My research is supported by NSERC, Borealis AI, Digital Research Alliance of Canada, and the Killam Fellowship. Feel free to reach out to discuss or collaborate.

SparseOpt: Addressing Normalization-induced Gradient Skew in Sparse Training
(To appear) ICML 2026, Seoul 🇰🇷
ICML 2026, Seoul 🇰🇷
What does it mean to align or match two neural networks, and when is such an alignment even possible?
Recent work like Git Re-Basin suggests that models trained independently from scratch can often be aligned through permutations, indicating a shared underlying structure. However, this phenomenon is primarily observed in wider networks. I'm interested in understanding how model width influences when such alignment is possible, and more broadly, how to improve model matching in modern architectures like transformers, especially at scale.
How can we leverage large collections of open-source model checkpoints (model zoo) to enable more efficient and scalable model adaptation?
Large collections of open-source model checkpoints offer a rich resource for understanding how models adapt across tasks. Methods like LoRA suggest that these adaptations can often be captured through low-dimensional updates rather than full retraining. However, from a weight-space perspective, such parameterizations are not unique—symmetries imply that many equivalent representations of these updates may exist across the model zoo, making weight-space learning challenging. I’m currently interested in developing weight-space learning approaches for LoRA that leverage this shared structure for efficient LLM adaptation.
Updates
Mar 2026 Gave a talk on the Role of Symmetries in Optimization at the Technical University of Munich (TUM), hosted by Dr. Stefanie Jegelka.
Mar 2026 Will be attending CPAL 2026 in Tübingen.
Oct 2025 Research visit to CISPA, Saarbrücken, Germany, hosted by Dr. Rebekka Burkholz.
May 2025 Joined Borealis AI as a Research Intern.
May 2025 Awarded the prestigious Killam Doctoral Fellowship in recognition of research contributions.
May 2025 Paper on Sparse Training accepted at ICML 2025 in Vancouver.
Dec 2024 Workshop proposal accepted at ICLR 2025. Co-organizing Sparsity in LLMs in Singapore.
Oct 2024 Research funding from the Digital Research Alliance of Canada to study LLM compression and model bias.
Sep 2024 Among 10 students awarded the Borealis AI Global Fellowship.
Apr 2024 Awarded the NSERC Doctoral Fellowship from the Government of Canada.
Jan 2023 Started PhD at Vector Institute / University of Calgary.
Sep 2022 Research internship at Borealis AI.
Jun 2021 Paper on whole-slide image classification accepted at MICCAI 2021.
Jul 2020 Paper on permutation-invariant representations accepted at ECCV 2020.
Apr 2020 Paper on histopathology GNNs accepted at CVPR Workshop 2020.
Apr 2019 Awarded Vector Scholarship in AI by the Vector Institute, Canada.