17,456 research outputs found
1-Bit Matrix Completion under Exact Low-Rank Constraint
We consider the problem of noisy 1-bit matrix completion under an exact rank
constraint on the true underlying matrix . Instead of observing a subset
of the noisy continuous-valued entries of a matrix , we observe a subset
of noisy 1-bit (or binary) measurements generated according to a probabilistic
model. We consider constrained maximum likelihood estimation of , under a
constraint on the entry-wise infinity-norm of and an exact rank
constraint. This is in contrast to previous work which has used convex
relaxations for the rank. We provide an upper bound on the matrix estimation
error under this model. Compared to the existing results, our bound has faster
convergence rate with matrix dimensions when the fraction of revealed 1-bit
observations is fixed, independent of the matrix dimensions. We also propose an
iterative algorithm for solving our nonconvex optimization with a certificate
of global optimality of the limiting point. This algorithm is based on low rank
factorization of . We validate the method on synthetic and real data with
improved performance over existing methods.Comment: 6 pages, 3 figures, to appear in CISS 201
1-Bit Matrix Completion
In this paper we develop a theory of matrix completion for the extreme case
of noisy 1-bit observations. Instead of observing a subset of the real-valued
entries of a matrix M, we obtain a small number of binary (1-bit) measurements
generated according to a probability distribution determined by the real-valued
entries of M. The central question we ask is whether or not it is possible to
obtain an accurate estimate of M from this data. In general this would seem
impossible, but we show that the maximum likelihood estimate under a suitable
constraint returns an accurate estimate of M when ||M||_{\infty} <= \alpha, and
rank(M) <= r. If the log-likelihood is a concave function (e.g., the logistic
or probit observation models), then we can obtain this maximum likelihood
estimate by optimizing a convex program. In addition, we also show that if
instead of recovering M we simply wish to obtain an estimate of the
distribution generating the 1-bit measurements, then we can eliminate the
requirement that ||M||_{\infty} <= \alpha. For both cases, we provide lower
bounds showing that these estimates are near-optimal. We conclude with a suite
of experiments that both verify the implications of our theorems as well as
illustrate some of the practical applications of 1-bit matrix completion. In
particular, we compare our program to standard matrix completion methods on
movie rating data in which users submit ratings from 1 to 5. In order to use
our program, we quantize this data to a single bit, but we allow the standard
matrix completion program to have access to the original ratings (from 1 to 5).
Surprisingly, the approach based on binary data performs significantly better
Matrix Completion via Max-Norm Constrained Optimization
Matrix completion has been well studied under the uniform sampling model and
the trace-norm regularized methods perform well both theoretically and
numerically in such a setting. However, the uniform sampling model is
unrealistic for a range of applications and the standard trace-norm relaxation
can behave very poorly when the underlying sampling scheme is non-uniform.
In this paper we propose and analyze a max-norm constrained empirical risk
minimization method for noisy matrix completion under a general sampling model.
The optimal rate of convergence is established under the Frobenius norm loss in
the context of approximately low-rank matrix reconstruction. It is shown that
the max-norm constrained method is minimax rate-optimal and yields a unified
and robust approximate recovery guarantee, with respect to the sampling
distributions. The computational effectiveness of this method is also
discussed, based on first-order algorithms for solving convex optimizations
involving max-norm regularization.Comment: 33 page
Estimation in high dimensions: a geometric perspective
This tutorial provides an exposition of a flexible geometric framework for
high dimensional estimation problems with constraints. The tutorial develops
geometric intuition about high dimensional sets, justifies it with some results
of asymptotic convex geometry, and demonstrates connections between geometric
results and estimation problems. The theory is illustrated with applications to
sparse recovery, matrix completion, quantization, linear and logistic
regression and generalized linear models.Comment: 56 pages, 9 figures. Multiple minor change
A Max-Norm Constrained Minimization Approach to 1-Bit Matrix Completion
We consider in this paper the problem of noisy 1-bit matrix completion under
a general non-uniform sampling distribution using the max-norm as a convex
relaxation for the rank. A max-norm constrained maximum likelihood estimate is
introduced and studied. The rate of convergence for the estimate is obtained.
Information-theoretical methods are used to establish a minimax lower bound
under the general sampling model. The minimax upper and lower bounds together
yield the optimal rate of convergence for the Frobenius norm loss.
Computational algorithms and numerical performance are also discussed.Comment: 33 pages, 3 figure
- …