29 research outputs found
Correlation Decay and Tractability of CSPs
The algebraic dichotomy conjecture of Bulatov, Krokhin and Jeavons yields an elegant characterization of the complexity of constraint satisfaction problems. Roughly speaking, the characterization asserts that a CSP L is tractable if and only if there exist certain non-trivial operations known as polymorphisms to combine solutions to L to create new ones.
In this work, we study the dynamical system associated with repeated applications of a polymorphism to a distribution over assignments. Specifically, we exhibit a correlation decay phenomenon that makes two variables or groups of variables that are not perfectly correlated become independent after repeated applications of a polymorphism.
We show that this correlation decay phenomenon can be utilized in designing algorithms for CSPs by exhibiting two applications:
1. A simple randomized algorithm to solve linear equations over a prime field, whose analysis crucially relies on correlation decay.
2. A sufficient condition for the simple linear programming relaxation for a 2-CSP to be sound (have no integrality gap) on a given instance
Detecting Adversarial Directions in Deep Reinforcement Learning to Make Robust Decisions
Learning in MDPs with highly complex state representations is currently
possible due to multiple advancements in reinforcement learning algorithm
design. However, this incline in complexity, and furthermore the increase in
the dimensions of the observation came at the cost of volatility that can be
taken advantage of via adversarial attacks (i.e. moving along worst-case
directions in the observation space). To solve this policy instability problem
we propose a novel method to detect the presence of these non-robust directions
via local quadratic approximation of the deep neural policy loss. Our method
provides a theoretical basis for the fundamental cut-off between safe
observations and adversarial observations. Furthermore, our technique is
computationally efficient, and does not depend on the methods used to produce
the worst-case directions. We conduct extensive experiments in the Arcade
Learning Environment with several different adversarial attack techniques. Most
significantly, we demonstrate the effectiveness of our approach even in the
setting where non-robust directions are explicitly optimized to circumvent our
proposed method.Comment: Published in ICML 202
Skill-Mix: a Flexible and Expandable Family of Evaluations for AI models
With LLMs shifting their role from statistical modeling of language to
serving as general-purpose AI agents, how should LLM evaluations change?
Arguably, a key ability of an AI agent is to flexibly combine, as needed, the
basic skills it has learned. The capability to combine skills plays an
important role in (human) pedagogy and also in a paper on emergence phenomena
(Arora & Goyal, 2023).
This work introduces Skill-Mix, a new evaluation to measure ability to
combine skills. Using a list of skills the evaluator repeatedly picks
random subsets of skills and asks the LLM to produce text combining that
subset of skills. Since the number of subsets grows like , for even modest
this evaluation will, with high probability, require the LLM to produce
text significantly different from any text in the training set. The paper
develops a methodology for (a) designing and administering such an evaluation,
and (b) automatic grading (plus spot-checking by humans) of the results using
GPT-4 as well as the open LLaMA-2 70B model.
Administering a version of to popular chatbots gave results that, while
generally in line with prior expectations, contained surprises. Sizeable
differences exist among model capabilities that are not captured by their
ranking on popular LLM leaderboards ("cramming for the leaderboard").
Furthermore, simple probability calculations indicate that GPT-4's reasonable
performance on is suggestive of going beyond "stochastic parrot" behavior
(Bender et al., 2021), i.e., it combines skills in ways that it had not seen
during training.
We sketch how the methodology can lead to a Skill-Mix based eco-system of
open evaluations for AI capabilities of future models
Extended Formulation Lower Bounds for Combinatorial Optimization
Linear and semidefinite programs are fundamental algorithmic tools, often providing conjecturallyoptimal results for a variety of combinatorial optimization problems. Thus, a naturalquestion is to understand the limitations of linear and semidefinite programming relaxations.In particular, the goal is to prove unconditional lower bounds on the size of any linear orsemidefinite programming relaxation for a given problem.In this dissertation, I will give two results of this flavor. First, I will show that any linearprogramming relaxation for refuting random instances of constraint satisfaction problems(e.g. k-SAT) requires super-polynomial size. This theorem can be understood as evidencethat refuting CSPs is hard, since it rules out a broad class of algorithms. Second, I willshow that any symmetric semidefinite programming relaxation for the matching problemin general graphs requires exponential size. Since there is a polynomial time algorithm forthe matching problem, this result provides an example of the limitations of semidefiniteprogramming relaxations
Faster Algorithms and Constant Lower Bounds for the Worst-Case Expected Error
The study of statistical estimation without distributional assumptions on data values, but with knowledge of data collection methods was recently introduced by Chen, Valiant and Valiant (NeurIPS 2020). In this framework, the goal is to design estimators that minimize the worst-case expected error. Here the expectation is over a known, randomized data collection process from some population, and the data values corresponding to each element of the population are assumed to be worst-case. Chen, Valiant and Valiant show that, when data values are ℓ∞-normalized, there is a polynomial time algorithm to compute an estimator for the mean with worst-case expected error that is within a factor π2 of the optimum within the natural class of semilinear estimators. However, their algorithm is based on optimizing a somewhat complex concave objective function over a constrained set of positive semidefinite matrices, and thus does not come with explicit runtime guarantees beyond being polynomial time in the input. In this paper we design provably efficient algorithms for approximating the optimal semilinear estimator based on online convex optimization. In the setting where data values are ℓ∞-normalized, our algorithm achieves a π2 -approximation by iteratively solving a sequence of standard SDPs. When data values are ℓ2-normalized, our algorithm iteratively computes the top eigenvector of a sequence of matrices, and does not lose any multiplicative approximation factor. Further, using experiments in settings where sample membership is correlated with data values (e.g. "importance sampling" and "snowball sampling"), we show that our ℓ2-normalized algorithm gives a similar advantage over standard estimators as the original ℓ∞-normalized algorithm of Chen, Valiant and Valiant, but with much lower computational complexity. We complement these positive results by stating a simple combinatorial condition which, if satisfied by a data collection process, implies that any (not necessarily semilinear) estimator for the mean has constant worst-case expected error