Past Seminars are listed here in chronological order
ABSTRACT
The problem of identifying geometric structure in heterogeneous, high-dimensional data is a cornerstone of Representation Learning. In this talk, we study this problem from the perspective of Discrete Geometry. We start by reviewing discrete notions of curvature with a focus on Ricci curvature. Then we discuss how curvature characterizations of graphs can be used to improve the efficiency of Graph Neural Networks. Specifically, we propose curvature-based rewiring and encoding approaches and study their impact on the Graph Neural Network’s downstream performance through theoretical and computational analysis. We further discuss applications of discrete Ricci curvature in Manifold Learning, where discrete-to-continuum consistency results allow for characterizing the geometry of a suitable embedding space both locally and in the sense of global curvature bounds.
Melanie is an Assistant Professor of Applied Mathematics and of Computer Science at Harvard University. Her research focuses on utilizing geometric structure in data for the design of efficient Machine Learning and Optimization methods. In 2021-2022, she was a Hooke Research Fellow at the Mathematical Institute in Oxford. Previously, she received her PhD from Princeton University (2021). She is the recipient of the 2023 IMA Leslie Fox Prize in Numerical Analysis and a 2024 Sloan Fellowship in Mathematics.
Acknowledgements The present talk has been made possible thanks to the help of Shubham Choudhary (Harvard) and Andy Keller (Harvard) who served as a volunteer for locally hosting and recording the seminar.
ABSTRACT
The performance of modern deep models relies on vast datasets to train, but in many domains data is scarce or difficult to collect. Incorporating symmetry constraints into neural networks has resulted in models called equivariant neural networks (ENNs) which have helped improve sample efficiency. As an application we consider equivariant policy learning which can be used to train robots using fewer iterations for reinforcement learning or fewer demonstrations for imitation learning. We will also discuss the limitations of standard equivariant learning, which assumes a known group action and suggest methods to circumvent this assumption.
Robin Walters is an assistant professor in the Khoury College of Computer Sciences at Northeastern University, where he leads the Geometric Learning Lab. Robin’s research seeks to develop a fundamental understanding of the role symmetry plays in deep learning and to exploit this to improve the generalization and data efficiency of deep learning methods. This includes designing equivariant neural networks, symmetry discovery methods, and creating a theory of symmetry for model parameters. He has applied these methods to improve models in domains with complex dynamics including climate science, transportation, and robotics
Dian Wang is a Ph.D. candidate at the Khoury College of Computer Sciences, Northeastern University, where he is co-advised by Prof. Robert Platt and Prof. Robin Walters. His research lies at the intersection of Machine Learning and Robotics, with a particular focus on Geometric Deep Learning and its applications in Robot Learning. Recently, Dian has focused on enhancing robotic manipulation through the use of equivariant methods to boost learning efficiency and performance. Dian has contributed to leading conferences and journals, including ICLR, NeurIPS, CoRL, ICRA, RSS, IJRR, AR, ISRR, and AAMAS. Dian was awarded the JPMorgan Ph.D. Fellowship in 2023 and the Khoury Research Fellowship in 2019.
Acknowledgements The present talk has been made possible thanks to the help of Behrooz Tahmasebi (MIT) who served as a volunteer for locally hosting and recording the seminar.
ABSTRACT
In this talk, we introduce a new class of problems related to integrating inertial measurements obtained from an IMU that play a significant role in navigation combined with visual data. While there have been tremendous technological advances in the precision of instrumentation, integrating acceleration and angular velocity still suffers from drift in the displacement estimates. Neural networks have come to the rescue in estimating displacement and the associated uncertainty covariance. However, such networks do not consider the physical roto reflective symmetries inherent in IMU data, leading to the need to memorize the same priors for every possible motion direction, which hinders generalization. In this work, we characterize these symmetries and show that the IMU data and the resulting displacement and covariance transform equivariantly when rotated around the gravity vector and reflected with respect to arbitrary planes parallel to gravity. We propose a network for predicting an equivariant gravity aligned frame from equivariant vectors and invariant scalars derived from IMU data, leveraging expressive linear and non-linear layers tailored to commute with the underlying symmetry transformation. Such a canonical frame can precede existing architectures that are end-to-end or filter-based. We will include an introduction to the inertial filtering problem and we will present results in real-world datasets.
Yinshuang Xu is currently pursuing her fifth year of PhD studies in Computer and Information Science at the University of Pennsylvania, where she is advised by Prof. Kostas Daniilidis. She previously graduated from Shanghai Jiao Tong University with a bachelor’s degree in Engineering Mechanics and later received her master’s in Robotics from the University of Pennsylvania. Her research interests include equivariance and geometric deep learning, with a focus on their use in computer vision and machine learning.
Kostas Daniilidis is the Ruth Yalom Stone Professor of Computer and Information Science at the University of Pennsylvania where he has been faculty since 1998. He is an IEEE Fellow. He obtained his undergraduate degree in Electrical Engineering from the National Technical University of Athens, 1986, and his PhD in Computer Science from the University of Karlsruhe, 1992, under the supervision of Hans-Hellmut Nagel. He received the Best Conference Paper Award at ICRA 2017. He co-chaired ECCV 2010 and 3DPVT 2006. His most cited works have been on event-based vision, equivariant learning, 3D human pose, and hand-eye calibration
Acknowledgements The present talk has been made possible thanks to the help of Ryan Chan (University of Pennsylvania) who served as a volunteer for locally hosting and recording the seminar. We'd like to further acknowledge René Vidal (University of Pennsylvania) and the Innovation in Data Engineering and Science (IDEAS) Initiative for supporting the NeurReps Speaker Series.
ABSTRACT
Invariant and equivariant networks have been the primary mechanism for imbuing machine learning with symmetry awareness. However, constraining architectures with symmetry may be infeasible or even undesirable. Infeasibility may be due to the inability of network design or lack of information on transformations, and undesirability may be due to the approximate nature of symmetries or suboptimal use of computing. In this talk, I'll briefly review several works from our group that use symmetries beyond equivariant networks. These examples explore symmetry in different learning paradigms ranging from reinforcement learning and self-supervised learning to physics-informed learning and generative modelling.
Siamak Ravanbakhsh is an Assistant Professor at the School of Computer Science at McGill University, a Canada CIFAR AI Chair, and a Core Member at Mila. His research explores the problem of representation learning, emphasizing symmetry, geometry, and probabilistic inference. Before joining McGill and Mila, he was an Assistant Professor at the University of British Columbia. He spent two years as a postdoctoral fellow at the Machine Learning Department and the Robotics Institute at Carnegie Mellon University, working with Barnabás Póczos and Jeff Schneider. He received his M.Sc. and Ph.D. from the University of Alberta as a member of the Alberta Ingenuity Center for Machine Learning, now Amii, working with Russ Greiner. He holds a B.Sc. from Sharif University of Technology.
Acknowledgements The present talk has been made possible thanks to the help of Colin Bredenberg (Mila) who served as a volunteer for locally hosting and recording the seminar.
ABSTRACT
There is now undeniable evidence for traveling waves and other structured spatiotemporal recurrent neural dynamics in cortical structures. However, to date, these observations have been difficult to reconcile with the dominant conceptual notions of topographically organized selectivity and feedforward receptive fields. This conceptual conflict has lead to the frequent marginalization of these dynamics to the category of ‘noise’, often being removed prior to neural data analysis. Our research proposes a new ‘spacetime’ perspective on neural computation in which structured selectivity and dynamics are not contradictory but instead may be complimentary. Through the analysis of state of the art models in the machine learning literature, along with the construction of novel spatiotemporal artificial neural network architectures, we show that spatiotemporal dynamics may be a mechanism by which natural neural systems encode approximate visual, temporal, and abstract symmetries of the world as conserved quantities, thereby enabling improved generalization and long-term working memory. Ultimately, we advocate that spacetime may be the fundamental ‘fabric of neural computation’, and believe our research calls for further empirical and theoretical study in this direction.
Andy Keller is a machine learning researcher who began studying computer science at Caltech, worked on deep learning research at Intel Nervana, and most recently completed his PhD at the University of Amsterdam under the supervision of Max Welling. He is now a Research Fellow at The Kempner Institute at Harvard University where he is attempting to discover the natural inductive biases imbued in biological neural networks, thereby enabling their improved generalization and reduced sample complexity when compared with modern artificial neural networks. Towards this goal, Andy’s research to-date has centered on biological mechanisms such as topographic organization and cortical traveling waves which may serve as natural mechanisms for structuring neural representations similar to the structure explicitly imposed in modern equivariant artificial neural network architectures.
Acknowledgements The present talk has been made possible thanks to the help of Manos Theodosis (Harvard University) who served as a volunteer for locally hosting and recording the seminar.