TEAM

Professor Barna Saha, Computer Science and Engineering, University of California San Diego
UCSD

Barna Saha is the E Guber Endowed Chair Associate Professor of UCSD CSE and HDSI. Before joining UCSD, she was an Associate Professor at UC Berkeley. Saha’s primary research focus is on Theoretical Computer Science, specifically Algorithm Design. She is passionate about diversity and teaching, and seeing students succeed from all backgrounds. She is a recipient of the Presidential Early Career Award (PECASE)- the highest honor given by the White House to early career scientists, a Sloan fellowship, an NSF CAREER Award, and multiple paper awards. Learn More.

CC
University of Sydney

I am a Lecturer in the School of Computer Science of the University of Sydney. Prior to that, I was a postdoc first in the Stanford Theory Group, then at IBM Research Almaden. Even prior to that, I obtained my Ph.D. from the Computer Science department of Columbia University, where I was advised by Prof. Rocco Servedio. Long ago, in a distant land, I received a M.Sc. in Computer Science from the Parisian Master of Research in Computer Science, and an engineering degree from one of France’s “Grand Schools,” the École Centrale Paris. Learn More.  

UCSD-JacobsSchool-20180928-Chaudhuri_Kamalika_thumb
UCSD

Kamalika Chaudhuri is a machine learning researcher. She is interested in the foundations of trustworthy machine learning — such as robust machine learning, learning with privacy and out-of-distribution generalization. She received her Ph.D. from University of California Berkeley. Learn More.  

chawla_1
UT Austin

Shuchi Chawla holds an Endowed Professorship in Computer Science at UT-Austin and is an Amazon Scholar. Shuchi is a theoretical computer scientist specializing in the areas of algorithm design, and economics and computation. Shuchi received a Ph.D. from Carnegie Mellon University and a B.Tech. from IIT, Delhi. Prior to joining UT-Austin, she spent 15 years as a professor of CS at the University of Wisconsin-Madison. She has also previously held visiting positions at the University of Washington and Microsoft Research. Shuchi recently served as the PC Chair of SODA’20 and EC’21, and currently serves on the editorial boards of the ACM Transactions on Algorithms and the ACM Transactions on Economics and Computation. Learn More.  

Professor Barna Saha, Computer Science and Engineering, University of California San Diego
UCSD

Sanjoy Dasgupta is a Professor in the Department of Computer Science and Engineering at UC San Diego. He received his PhD from Berkeley in 2000, and spent two years at AT&T Research Labs before joining UCSD. His area of research is algorithmic statistics, with a focus on unsupervised and minimally supervised learning. Learn More.  

allie
UCLA

Allie Fletcher is an Assistant Professor of UCLA Statistics Department. Previously, she was an assistant professor of UC Santa Cruz Professor. She was awarded the NSF CAREER award in January, 2013. She was previously a UC President’s Postdoctoral Fellow, a Henry Luce Foundation Claire Boothe Luce Fellow and won the UC Berkeley EECS Eugene Lawler award. She received the Ph.D. and M.S. degrees in electrical engineering, and the M.S. degree in mathematics, all from the University of California, Berkeley. She obtained the B.S. degree in mathematics and physics from the University of Iowa. Her research interests include Bayesian estimation, statistical signal processing, machine learning, and information theory with a particular interest in inference problems in neuroscience and computational models of the brain. Learn More.  

March 26, 2015 / Faculty and Staff Head shots / Photo by Bob Laramie
Upenn

Dr. Rajiv Gandhi is an Associate Professor of Computer Science at the Rutgers University-Camden. He also teaches pat-time at the University of Pennsylvania. He received his Ph.D. in Computer Science in 2003 from the University of Maryland, College Park. His research interests lie in the broad area of theoretical computer science. Specifically, he is interested in approximation and randomized algorithms. He is a passionate educator who loves working with students with diverse backgrounds, helping them achieve their potential. He has been the recipient of several teaching excellence awards, including the Warren I. Susman award for teaching excellence at Rutgers University in 2014. He also received the Chancellor’s award for Civic Engagement at Rutgers-Camden in 2013. In 2017 he was inducted in the Computer Science Alumni Hall of Fame at the University of Maryland. He was a Fulbright Fellow from Jan-June 2011, during which he worked with students in Mumbai, India. Since 2009, he has also been working with high school students as part of the Program in Algorithmic and Combinatorial Thinking (http://algorithmicthinking.org). Learn More.  

VENKATA GANDIKOTA
Syracuse University

Venkata Gandikota is senior personnel of EnCORE and an Assistant Professor in the Department of Electrical Engineering and Computer Science at Syracuse University.

Previously, he was a TRIPODS (Institute for Theoretical Foundations of Data Science) postdoctoral researcher at the University of Massachusetts at Amherst where he worked with Prof. Arya Mazumdar, and a postdoctoral fellow at Johns Hopkins University with Prof. Xin Li.

He received his Ph.D. in Computer Science from Purdue University under the supervision of Prof. Elena Grigorescu.

He works on the foundations of data science and machine learning. In the past, he has thoroughly enjoyed working on some classical problems in lattices and error-correcting codes. Learn More.

Fan Chung Graham
UCSD

Fan Chung Graham (professional name: Fan Chung , Chinese name) is a Distinguished Professor of Mathematics and Computer Science at UC San Diego. She holds the Paul Erdos Chair in Combinatorics. She paints watercolors, especially in seascape and portraits (click the Erdös’ painting below). A good part of her work in mathematics was told in the AMS Notics article The mathematical life of Fan Chung by Steve Butler. Learn More.  

Mustafa
Santa Clara University

Mustafa Hajij is senior personnel of EnCORE and assistant professor of data science at the department of mathematics and computer science SCU. His research interests include Topological and Geometric Deep Learning, Topological Data Analysis, and Geometric Data Processing. Learn More.

Abolfazl
Purdue

Abolfazl Hashemi is senior personnel of EnCORE and Assistant Professor at The Elmore Family School of Electrical and Computer Engineering at Purdue University and the director of MINDS Group. Previously, he was a Postdoctoral scholar at the Oden Institute for Computational Engineering and Sciences at the University of Texas at Austin. He received his PhD and MSE degrees in Electrical and Computer Engineering department at UT Austin in 2020 and 2016, and his BS degree in Electrical Engineering from Sharif University of Technology in 2014. Learn More.  

hassani-hamed300x240
UPenn

Hamed Hassani is an assistant professor in the Department of Electrical and Systems Engineering , and holds a secondary appointments in the Department of Computer and Information Systems as well as the Department of Statistics and Data Science at the Wharton Business School.
Before joining Penn, he was a research fellow at the Simons Institute, UC Berkeley, and a post-doctoral scholar and lecturer in the Institute for Machine Learning at ETH Zürich. He received his Ph.D. degree in Computer and Communication Sciences from EPFL. Learn More.

russell 2
UCSD

Russel Impagliazzo is a professor specializing in computational complexity theory. His research is in proof complexity, the theory of cryptography, computational randomness, structural complexity, and trying to analyze optimization heuristics and other approaches to solving hard problems. Learn More.

sandy
UC Irvine

Sandy Irani has research interests in Quantum Information and Computation, Online Algorithms, and Algorithms with applications to computer systems and resource allocation. Learn More.

daniel
UCSD

Daniel Kane is an associate professor at UCSD with a joint appointment between CSE and Mathematics. He has a broad research interests within mathematics and theoretical computer science. While his research covers a wide range of topics, his mathematics work tends to be within the broad areas of number theory and combinatorics, while his computer science research tends to be about derandomization, Boolean functions (with particular emphasis on polynomial threshold functions), or machine learning. Learn More.

Dominik
Stony Brook

Dominik Kempa is senior personnel of EnCORE and an assistant professor in the Department of Computer Science at Stony Brook University. Prior to joining Stony Brook, he was a postdoc at Johns Hopkins University fortunate to be hosted by Ben Langmead, and before that at UC Berkeley (Theory group), lucky to be advised by Barna Saha.

During his PhD at the University of Helsinki, he implemented a collection of parallel and external-memory algorithms on strings (code: here or Github). Earlier, he was an intern at Google, Zürich. As an undergraduate, he enjoyed algorithm competitions (and practiced here). He is a recipient of the Junior Researcher Award and Outstanding Doctoral Dissertation Award. Learn More.

6674
UCSD

Shachar Lovett is an Associate Professor in the CSE department in UC San Diego. He is part of the Theory of Computation group and helps run the theory seminar.

He has a broad interest in theoretical computer science and mathematics. In particular computational complexity, randomness and pseudo-randomness, algebraic constructions, coding theory, additive combinatorics and high-dimensional geometry. Learn More.

Robert
University of michigan

Robert Lunde is a senior personnel of EnCORE and is currently a Research Fellow in the Department of Statistics at the University of Michigan, working with Elizaveta Levina and Ji Zhu. Starting Fall 2022, he will be an assistant professor in the Department of Mathematics and Statistics at Washington University in St. Louis.

He is broadly interested in prediction and inference for data with complex dependence structures, such as networks and time series.

A major aim of his research has been to study the properties of resampling methods such as the bootstrap, subsampling, and the jackknife in these settings. Learn More.

arya300x240
UCSD

Arya Mazumdar is an associate professor of UCSD HDSI, and an affiliate faculty member of Computer Science & Engineering. He obtained his Ph.D. degree from University of Maryland, College Park (2011) specializing in information theory. Subsequently Arya was a postdoctoral scholar at Massachusetts Institute of Technology (2011-2012), an assistant professor in University of Minnesota (2013-2015), and an assistant followed by associate professor in University of Massachusetts Amherst (2015-2021). Arya is a recipient of multiple awards, including a Distinguished Dissertation Award for his Ph.D. thesis (2011), the NSF CAREER award (2015), an EURASIP JSAP Best Paper Award (2020), and the IEEE ISIT Jack K. Wolf Student Paper Award (2010). He is currently serving as an Associate Editor for the IEEE Transactions on Information Theory and as an Area editor for Now Publishers Foundation and Trends in Communication and Information Theory series. Arya’s research interests include coding theory (error-correcting codes and related combinatorics), information theory, statistical learning and distributed optimization.  Learn More. 

raghu
UCLA

Raghu Meka is an associate Professor in the CS department at UCLA. He is broadly interested in complexity theory, learning and probability theory. He got his PhD at UT Austin under the (wise) guidance of David Zuckerman in 2011. After that he spent two wonderful years in Princeton as a postdoctoral fellow at the Institute for Advanced Study with Avi Wigderson and at DIMACS, Rutgers. After that he spent an enjoyable year as a researcher at Microsoft Research, Silicon Valley Lab. he had a great time on sabbatical visiting the MIT math department from 2018-2020. Learn More.

me
UCSD

Gal Mishne is an Assistant Professor of HDSI. Mishne’s research is at the intersection of signal processing and machine learning for graph-based modeling, processing and analysis of large-scale high-dimensional real-world data. She develops unsupervised and generalizable methods that allow the data to reveal its own story in an unbiased manner. Her research includes anomaly detection and clustering in remote sensing imagery, manifold learning on multiway data tensors with biomedical applications, and computationally efficient application of spectral methods. Most recently her research has focused on unsupervised data analysis in neuroscience, from processing of raw neuroimaging data through discovery of neural manifolds to visualization of learning in neural networks. Learn More.  

Seth Neel
Harvard business school

Seth Neel is senior personnel of EnCORE and an Assistant Professor of Business Administration in the Department of Technology and Operations Management at Harvard Business School. His research involves the theory and application of machine learning, where he builds tools that incorporate (differential) privacy and fairness. He is a co-founder of Welligence, a data analytics company focused on the energy space. Learn More.  

pappas-george
UPenn

George J. Pappas is the UPS Foundation Professor and Chair of the Department of Electrical and Systems Engineering at the University of Pennsylvania. He also holds a secondary appointment in the Departments of Computer and Information Sciences, and Mechanical Engineering and Applied Mechanics. He is member of the GRASP Lab and the PRECISE Center. He has previously served as the Deputy Dean for Research in the School of Engineering and Applied Science. His research focuses on control theory and in particular, hybrid systems, embedded systems, hierarchical and distributed control systems, with applications to unmanned aerial vehicles, distributed robotics, green buildings, and biomolecular networks. He is a Fellow of IEEE, and has received various awards such as the Antonio Ruberti Young Researcher Prize, the George S. Axelby Award, the O. Hugo Schuck Best Paper Award, the National Science Foundation PECASE, and the George H. Heilmeier Faculty Excellence Award. Learn More.  

paturi 2
UCSD

Rammohan Paturi is a Professor in the department of Computer Science and Engineering at UCSD. His research interest areas include Algorithms and Complexity, Social Networks, Text Mining, and Medical Informatics.  Learn More.  

roth
UPenn

Aaron Roth is the Henry Salvatori Professor of Computer and Cognitive Science at the University of Pennsylvania computer science department. He also holds a secondary appointment at the Department of Statistics and Data Science at the Wharton School, and he is associated with the theory groupPRiML (Penn Research in Machine Learning) the Warren Center for Network and Data Sciences, and is a co-director of  Networked and Social Systems Engineering. He is also affiliated with the AMCS program (Applied Mathematics and Computational Science). He spent a year as a postdoc at Microsoft Research New England. Before that, he received his PhD from Carnegie Mellon University, where he was fortunate to have been advised by Avrim Blum. His main interests are in algorithms and machine learning, and specifically in the areas of private data analysis, fairness in machine learning, game theory and mechanism design, and learning theory. He is the recipient of a Presidential Early Career Award for Scientists and Engineers (PECASE), an Alfred P. Sloan Research Fellowship, an NSF CAREER award, a Google Faculty Research Award, an Amazon Research Award, and a Yahoo Academic Career Enhancement award. He is also an Amazon Scholar at Amazon Web Services (AWS). Previously, he was involved in advisory and consulting work related to differential privacy, algorithmic fairness, and machine learning, including with Apple and Facebook. He was also a scientific advisor for Leapyear and Spectrum LabsLearn More.  

sujay
UT Austin

Sujay Sanghavi is an Associate Professor and holds the Fluor Centennial Teaching Fellowship in Engineering # 2 in the Department of Electrical and Computer Engineering at The University of Texas at Austin. Dr. Sanghavi joined Texas ECE in July 2009. He got his PhD from UIUC, and a postdoc from MIT. His research interests lie at the intersection of two central challenges in modern systems: large-scale networking and high-dimensional data analysis, with a focus on algorithm design and evaluation. He received the NSF CAREER award in January 2010. Learn More.  

Sarkar_Purna
UT Austin

Purnamrita (Purna) Sarkar joined The University of Texas at Austin in 2014 as an assistant professor. Previously, she was a postdoctoral scholar at the University of California, Berkeley, jointly in the Department of Electrical Engineering and Computer Sciences and the Department of Statistics. She works on large-scale statistical machine-learning problems with a focus on statistical models, asymptotic theory, and scalable inference algorithms for large networks. Learn More.

Eric Tchetgen
UPenn

Eric J. Tchetgen Tchetgen is the Luddy Family President’s Distinguished Professor of Statistics and Data Science at UPenn. His primary area of interest is in semi-parametric efficiency theory with application to causal inference, missing data problems, statistical genetics and mixed model theory. He has received numerous awards including the Rousseeuw Prize for Statistics, 2022, the Myrto Lefkopoulou Distinguished Lectureship, 2020, the Society of Epidemiologic Research and American Journal of Epidemiology Article of the Year, 2014, Career Incubator Award, Harvard School of Public Health, 2013-2014, the Kenneth Rothman Epidemiology Prize, 2011, Best Poster Award: Gene Environment Initiative Symposium, Boston, MA, 2008, Yerby Fellowship, Harvard School of Public Health, 2006-2008, Mars Scholar, Yale University, 1995-1996 etc. Learn More.  

YYusu300x240
UCSD

Yusu Wang is a Professor of HDSI@UCSD, and also Director for NSF AI Institute TILOS. She obtained her PhD degree from Duke University in 2004, where she received the Best Dissertation Award from the CS Department. From 2004 – 2005, she was a post-doctoral fellow at Stanford University. Before joining UC San Diego, Yusu Wang was a Professor of the Computer Science and Engineering Department at the Ohio State University. Yusu Wang primarily works in the field of geometric and topological data analysis. She is particularly interested in developing effective and theoretically justified algorithms and machine learning models for data analysis using geometric and topological ideas and methods, as well as applying them to practical domains. She received DOE Early Career Principal Investigator Award in 2006, and NSF Career Award in 2008. Her work received several best paper awards. She is currently on the editorial boards for SIAM Journal on Computing (SICOMP), Computational Geometry: Theory and Applications (CGTA), and Journal of Computational Geometry (JoCG).  She also serves on the Computational Geometry Steering Committee and AATRN Advisory Committee. Learn More.

EmbeddedImage
Ut Austin

Rachel Ward is the W.A. “Tex” Moncrief Distinguished Professor in Computational Engineering and Sciences — Data Science and Professor of Mathematics at UT Austin. She is recognized for her contributions to stochastic gradient descent, compressive sensing, and randomized linear embeddings. From 2017-2018, Dr. Ward was a visiting researcher at Facebook AI Research. Prior to joining UT Austin in 2011, Dr. Ward received the PhD in Computational and Applied Mathematics at Princeton in 2009 and was a Courant Instructor at the Courant Institute, NYU, from 2009-2011. Among her awards are the Sloan research fellowship, NSF CAREER award, 2016 IMA prize in mathematics and its applications, 2020 Simons fellowship in mathematics. She is also an invited speaker at the 2022 International Congress of Mathematicians. Learn More.  

Product Manifold Learning with Independent Coordinate Selection

Jesse He (UCSD)

Abstract: In many dimensionality reduction tasks, we wish to identify the constituent components that explain our observations. For manifold learning, this can be formalized as factoring a Riemannian product manifold. Recovering this factorization, however, may suffer from certain difficulties in practice, especially when data is sparse or noisy, or when one factor is distorted by the other. To address these limitations, we propose identifying non-redundant coordinates on the product manifold before applying product manifold learning to identify which coordinates correspond to different factor manifolds.

Tear and Repulsion Enabled Registration of Point Clouds for Manifold Learning

Dhruv Kohli (UCSD)

Abstract: We present a framework for aligning the local views of a possibly closed/non-orientable data manifold to produce an embedding in its intrinsic dimension through tearing. Through a spectral coloring scheme, we render the embeddings of the points across the tear with matching colors, enabling a visual recovery of the topology of the data manifold. The embedding is further equipped with a tear-aware metric that enables computation of shortest paths while accounting for the tear. To measure the quality of an embedding, we propose two Lipschitz-type notions of global distortion—a stronger and a weaker one—along with their pointwise counterparts for a finer assessment of the embedding. Subsequently, we bound them using the distortion of the local views and the alignment error between them. We show that our theoretical result on strong distortion leads to a new perspective on the need for a repulsion term in manifold learning objectives. As a result, we enhance our alignment approach by incorporating repulsion. Finally, we compare various strategies for the tear and repulsion enabled alignment, with regard to their speed of convergence and the quality of the embeddings produced.


This is joint work with my advisors Gal Mishne and Alex Cloninger at UCSD.

Efficient Online Clustering with Moving Costs

Dimitrios Christou (UT Austin)

Abstract: In this talk, I will consider an online-learning problem, called Online Clustering with Moving Costs, at which a learner maintains a set of facilities over rounds so as to minimize the connection cost of an adversarially selected sequence of clients. The learner is informed on the positions of the clients at each round only after its facility-selection and can use this information to update its decision in the next round. However, updating the facility positions comes with an additional moving cost based on the moving distance of the facilities. I will be presenting the first (polynomial-time) approximate-regret algorithm for this setting through a combination of different algorithmic techniques such as HST embeddings, the FTRL framework with a dilated entropic regulariser as well as a novel rounding scheme.

Encoding Structural Symmetry is Key for Length Generalization in Arithmetic Tasks

Mahdi Sabbaghi (UPenn)

Abstract: Despite the success of Transformers on language understanding, code generation, and logical reasoning, they still fail to (length) generalize on basic arithmetic tasks such as addition and multiplication. A major reason behind this failure is the vast difference in structure of numbers versus text; for example, numbers are associated with a specific significance order that plays a role in calculating the answer. In contrast, for text, such symmetries are quite unnatural. In this work, we propose to encode these semantics explicitly into the model via appropriate data formatting and custom positional encodings. To further elucidate the importance of explicitly encoding structure, in a simplified linear setting, we prove that standard positional encodings even when trained with augmentations to implicitly induce structure fail at such generalization, whereas enforcing structure via positional encodings succeeds.

Bio: Mahdi Sabbaghi is a second year PhD student at UPenn, department of Electrical and System Engineering, supervised by Professors Hamed Hassani and George Pappas. Previously he obtained a B. Sc. degree in Electrical Engineering as well as a B. Sc. degree in Physics from the Sharif University of Technology, in Tehran.

Private Estimation of U Statistics

Shourya Pandey (UT Austin)

Abstract: We consider the problem of private estimation of U statistics. U statistics are widely used estimators that naturally arise in a broad class of problems, from nonparametric signed rank tests to subgraph counts in random networks. Despite the recent outpouring of interest in private mean estimation, private algorithms for more general U statistics have received little attention.  We propose a framework where, for a broad class of U statistics, one can use existing tools in private mean estimation to obtain confidence intervals where the private error does not overwhelm the irreducible error resulting from the variance of the U statistics. However, in specific cases that arise when the U statistics degenerate or have vanishing moments, the private error may be of a larger order than the non-private error. To remedy this, we propose a new thresholding-based approach that uses Hajek projections to re-weight different subsets. As we show,  this leads to more accurate inference in certain settings.

Finite-Time Logarithmic Bayes Regret Upper Bounds

Alexia Atsidakou (UT Austin)

Abstract: We derive the first finite-time logarithmic Bayes regret upper bounds for Bayesian bandits, for BayesUCB and Thompson Sampling. In Gaussian and Bernoulli multi-armed bandits, we obtain $O(c_\Delta \log n)$ and $O(c_h \log^2 n)$ upper bounds for an upper confidence bound algorithm, where $c_h$ and $c_\Delta$ are constants depending on the prior distribution and the gaps of bandit instances sampled from it, respectively. The latter bound asymptotically matches the lower bound of Lai (1987). Our proofs are a major technical departure from prior works, while being simple and general. The key idea in our proofs is to conduct a frequentist per-instance analysis with Bayesian confidence intervals, and then integrate it over the prior.

Our results provide insights on the value of prior in the Bayesian setting, both in the objective and as a side information given to the learner. They significantly improve upon existing $\tilde{O}(\sqrt{n})$ bounds, which have become standard in the literature despite the logarithmic lower bound of Lai (1987).

Pareto-Optimal Algorithms for Learning in Games

Natalie Collina and Eshwar Ram Arunachaleswaran

Abstract: We study the problem of characterizing optimal learning algorithms for playing repeated games against an adversary with unknown payoffs. In this problem, the first player (called the learner) commits to a learning algorithm against a second player (called the optimizer), and the optimizer best-responds by choosing the optimal dynamic strategy for their (unknown but well-defined) payoff. Classic learning algorithms (such as no-regret algorithms) provide some counterfactual guarantees for the learner, but might perform much more poorly than other learning algorithms against particular optimizer payoffs.

In this paper, we introduce the notion of asymptotically Pareto-optimal learning algorithms. Intuitively, if a learning algorithm is Pareto-optimal, then there is no other algorithm which performs asymptotically at least as well against all optimizers and performs strictly better (by at least $\Omega(T)$) against some optimizer. We show that well-known no-regret algorithms such as Multiplicative Weights and Follow The Regularized Leader are Pareto-dominated. However, while no-regret is not enough to ensure Pareto-optimality, we show that a strictly stronger property, no-swap-regret, is a sufficient condition for Pareto-optimality.

Proving these results requires us to address various technical challenges specific to repeated play, including the fact that there is no simple characterization of how optimizers who are rational in the long-term best-respond against a learning algorithm over multiple rounds of play. To address this, we introduce the idea of the asymptotic menu of a learning algorithm: the convex closure of all correlated distributions over strategy profiles that are asymptotically implementable by an adversary. Interestingly, we show that all no-swap-regret algorithms share the same asymptotic menu, implying that all no-swap-regret algorithms are “strategically equivalent”.

This talk is based on work with Jon Schneider.

Metric Learning from Lazy Crowds

Geelon So

Abstract: We consider crowd-based metric learning from preference comparisons, where given two items, a user prefers the item that is closer to their latent ideal item. Here, “closeness” is measured with respect to a shared but unknown Mahalanobis distance. Can we recover this distance when we can only obtain very few responses per user?

In this very low-budget regime, we show that generally, nothing at all about the metric is revealed, even with infinitely many users. But when the items have subspace cluster structure, we present a divide-and-conquer approach for metric recovery, and provide theoretical recovery guarantees and empirical validation.

This is joint work with Zhi Wang (UCSD) and Ramya Korlakai Vinayak (UW–Madison).

Random Walks, Conductance, and Resistance for the Connection Graph Laplacian

Zhengchao Wan

Abstract: We investigate the concept of effective resistance in connection graphs, expanding its traditional application from undirected graphs. We propose a robust definition of effective resistance in connection graphs by focusing on the duality of Dirichlet-type and Poisson-type problems on connection graphs. Additionally, we delve into random walks, taking into account both node transitions and vector rotations. This approach introduces novel concepts of effective conductance and resistance matrices for connection graphs, capturing mean rotation matrices corresponding to random walk transitions. Thereby, it provides new theoretical insights for network analysis and optimization.

This is based on a joint work with Alexander Cloninger, Gal Mishne, Andreas Oslandsbotn, Sawyer Jack Robertson and Yusu Wang.

Approximability and Inapproximability of Strict-CSPs

Akbar Rafiey

Abstract: We study the approximability and inapproximability of Strict-CSPs. An instance of the Strict-CSPs consists of a set of constraints over a set of variables and a cost function over the assignments. The goal is to find an assignment to the variables of minimum cost which satisfies all the constraints. Some prominent problems that this framework captures are (Hypergraph) Vertex Cover, Min Sum k-Coloring, Multiway Cut, Min Ones, and others.

We focus on a systematic study of Strict-CSPs of the form Strict-CSPs(H), that is, Strict-CSPs where the type of constraints is limited to predicates from a set H. Our first result is a dichotomy for approximation of Strict-CSPs(H), where H is a binary predicate, i.e., a digraph. We prove that if digraph H has bounded width, then Strict-CSPs(H) is approximable within a constant factor (depending on H); otherwise, there is no approximation for Strict-CSPs(H) unless P=NP.

Second, we study the inapproximability of Strict-CSP and present the first general hardness of approximation for Strict-CSP. More precisely, we prove a dichotomy theorem that states every instance of Strict-CSP(H) (H being a digraph) is either polynomial-time solvable or APX-complete. Moreover, we show the existence of a universal constant 0<\delta<1 such that it is NP-hard to approximate Strict-CSP(H) within a factor of (2-\delta) for all digraphs H where Strict-CSP(H) is NP-complete.

Buy-many Mechanisms for Many Unit-demand Buyers

Rojin Rezvan

Abstract: A recent line of research has established a novel desideratum for designing approximatelyrevenue-optimal multi-item mechanisms, namely the buy-many constraint. Under this constraint, prices for different allocations made by the mechanism must be subadditive implying that the price of a bundle cannot exceed the sum of prices of individual items it contains. This natural constraint has enabled several positive results in multi-item mechanism design bypassing well-established impossibility results. Our work addresses a main open question from this literature involving the design of buymany mechanisms for multiple buyers. Our main result is that a simple sequential item pricing mechanism with buyer-specific prices can achieve an O(log m) approximation to the revenue of any buy-many mechanism when all buyers have unit-demand preferences over m items. This is the best possible as it directly matches the previous results for the single-buyer setting where no simple mechanism can obtain a better approximation. Our result applies in full generality: even though there are many alternative ways buy-many mechanisms can be defined for multibuyer settings, our result captures all of them at the same time. We achieve this by directly competing with a more permissive upper-bound on the buy-many revenue, obtained via an ex-ante relaxation.

Streaming PCA for Markovian Data

Syamantak Kumar

Abstract: Since its inception in 1982, Oja’s algorithm has become an established method for streaming principle component analysis (PCA). We study the problem of streaming PCA, where the data-points are sampled from an irreducible, aperiodic, and reversible Markov chain. Our goal is to estimate the top eigenvector of the unknown covariance matrix of the stationary distribution. This setting has implications in scenarios where data can solely be sampled from a Markov Chain Monte Carlo (MCMC) type algorithm, and the objective is to perform inference on parameters of the stationary distribution. Most convergence guarantees for Oja’s algorithm in the literature assume that the data-points are sampled IID. For data streams with Markovian dependence, one typically downsamples the data to get a “nearly” independent data stream. In this paper, we obtain the first sharp rate for Oja’s algorithm on the entire data, where we remove the logarithmic dependence on the sample size, resulting from throwing data away in downsampling strategies.

A d^{1/2 + o(1)} Monotonicity Tester for Boolean Functions on d-Dimensional Hypergrids

Hadley Black

Abstract: Monotonicity testing of Boolean functions over the n-width, d-dimensional hypergrid is a classic problem in property testing, where the goal is to design a randomized algorithm which can distinguish monotone functions from those which are far from any monotone function while making as few queries as possible. The special case of n = 2 corresponds to the hypercube domain. Here a long line of works exploiting a very interesting connection with isoperimetric inequalities for Boolean functions culminated in a non-adaptive tester making ~O(d^{1/2}) queries in a celebrated paper by Khot, Minzer, and Safra (SICOMP 2018). This is known to be optimal for non-adaptive testers. However, the general case of hypergrids for n > 2 remained open. Very recently, two papers (Black-Chakrabarty-Seshadhri STOC 2023 and Braverman-Khot-Kindler-Minzer ITCS 2023) independently obtained ~O(poly(n) d^{1/2}) query testers for hypergrids. These results are essentially optimal for n < polylog(d), but are far from optimal for n >> polylog(d).

This talk covers our most recent result (appearing at FOCS 2023) which obtains a non-adaptive d^{1/2+o(1)} query tester for all n, resolving the non-adaptive monotonicity testing problem for hypergrids, up to a factor of d^{o(1)}. Our proof relies on many new techniques as well as two key theorems which we proved in earlier works from SODA 2020 and STOC 2023.

SmoothLLMs: Defending LLMs against Adversarial Attacks

Alex Robey

Abstract: Despite efforts to align large language models (LLMs) with human values, widely-used LLMs such as GPT, Llama, Claude, and PaLM are susceptible to jailbreaking attacks, wherein an adversary fools a targeted LLM into generating objectionable content.  To address this vulnerability, we propose SmoothLLM, the first algorithm designed to mitigate jailbreaking attacks on LLMs.  Based on our finding that adversarially-generated prompts are brittle to character-level changes, our defense first randomly perturbs multiple copies of a given input prompt, and then aggregates the corresponding predictions to detect adversarial inputs.  SmoothLLM reduces the attack success rate on numerous popular LLMs to below one percentage point, avoids unnecessary conservatism, and admits provable guarantees on attack mitigation.  Moreover, our defense uses exponentially fewer queries than existing attacks and is compatible with any LLM.

Composition of Nested Embeddings with an Application to Outlier Removal

Kristin Sheridan

Abstract: We study the design of embeddings into Euclidean space with outliers. Given a metric space $(X,d)$ and an integer $k$, the goal is to embed all but $k$ points in $X$ (called the “”outliers””) into $\ell_2$ with the smallest possible distortion $c$. Finding the optimal distortion $c$ for a given outlier set size $k$, or alternately the smallest $k$ for a given target distortion $c$ are both NP-hard problems. In fact, it is UGC-hard to approximate $k$ to within a factor smaller than $2$ even when the metric sans outliers is isometrically embeddable into $\ell_2$. We consider bi-criteria approximations. Our main result is a polynomial time algorithm that approximates the outlier set size to within an $O(\log^2 k)$ factor and the distortion to within a constant factor.

The main technical component in our result is an approach for constructing a composition of two given embeddings from subsets of $X$ into $\ell_2$ which inherits the distortions of each to within small multiplicative factors. Specifically, given a low $c_S$ distortion embedding from $S\subset X$ into $\ell_2$ and a high(er) $c_X$ distortion embedding from the entire set $X$ into $\ell_2$, we construct a single embedding that achieves the same  distortion $c_S$ over pairs of points in $S$ and an expansion of at most $O(\log k)\cdot c_X$ over the remaining pairs of points, where $k=|X\setminus S|$. Our composition theorem extends to embeddings into arbitrary $\ell_p$ metrics for $p\ge 1$, and may be of independent interest. While unions of embeddings over disjoint sets have been studied previously, to our knowledge, this is the first work to consider compositions of {\em nested} embeddings.

Graph Sparsification by Approximate Matrix Multiplication

Neo Charalambides

Abstract:  Graphs arising in statistical problems, signal processing, large networks, combinatorial optimization, and data analysis are often dense, which causes both computational and storage bottlenecks. One way of sparsifying a weighted graph, while sharing the same vertices as the original graph but reducing the number of edges, is through spectral sparsification. We study this problem through the perspective of RandNLA. Specifically, we utilize randomized matrix multiplication to give a clean and simple analysis of how sampling according to edge weights gives a spectral approximation to graph Laplacians, without requiring spectral information. Through the CR−MM algorithm, we attain a simple and computationally efficient sparsifier whose resulting Laplacian estimate is unbiased and of minimum variance. Furthermore, we define a new notion of additive spectral sparsifiers, which has not been considered in the literature.

Talk Title

Student Name

Abstract