1,537 research outputs found
A Heuristic Approach to the Consecutive Ones Submatrix Problem
أعطيت مصفوفة (0،1)، تم اقتراح مسألة المصفوفة الجزئية ذات الواحدات المتعاقبة والتي تهدف إلى إيجاد تبديل للأعمدة التي تزيد من عدد الأعمدة التي تحتوي معًا على قالب واحد فقط من الواحدات المتعاقبة في كل صف. سيتم اقتراح اسلوب الاستدلال لحل المسألة. كما سيتم دراسة مسألة تقليل القوالب المتتالية ذات الصلة بمسألة المصفوفة الجزئية ذات الواحدات المتعاقبة. تم اقتراح اجراء جديد لتحسين طريقة إدراج العمود. يتم بعد ذلك تقييم مصفوفات العالم الحقيقي ومصفوفات متولدة عشوائيًا من مسألة غطاء المجموعة و تعرض النتائج الحسابية.Given a matrix, the Consecutive Ones Submatrix (C1S) problem which aims to find the permutation of columns that maximizes the number of columns having together only one block of consecutive ones in each row is considered here. A heuristic approach will be suggested to solve the problem. Also, the Consecutive Blocks Minimization (CBM) problem which is related to the consecutive ones submatrix will be considered. The new procedure is proposed to improve the column insertion approach. Then real world and random matrices from the set covering problem will be evaluated and computational results will be highlighted
Linear-Time Algorithms for Finding Tucker Submatrices and Lekkerkerker-Boland Subgraphs
Lekkerkerker and Boland characterized the minimal forbidden induced subgraphs
for the class of interval graphs. We give a linear-time algorithm to find one
in any graph that is not an interval graph. Tucker characterized the minimal
forbidden submatrices of binary matrices that do not have the consecutive-ones
property. We give a linear-time algorithm to find one in any binary matrix that
does not have the consecutive-ones property.Comment: A preliminary version of this work appeared in WG13: 39th
International Workshop on Graph-Theoretic Concepts in Computer Scienc
Computational barriers in minimax submatrix detection
This paper studies the minimax detection of a small submatrix of elevated
mean in a large matrix contaminated by additive Gaussian noise. To investigate
the tradeoff between statistical performance and computational cost from a
complexity-theoretic perspective, we consider a sequence of discretized models
which are asymptotically equivalent to the Gaussian model. Under the hypothesis
that the planted clique detection problem cannot be solved in randomized
polynomial time when the clique size is of smaller order than the square root
of the graph size, the following phase transition phenomenon is established:
when the size of the large matrix , if the submatrix size
for any , computational complexity
constraints can incur a severe penalty on the statistical performance in the
sense that any randomized polynomial-time test is minimax suboptimal by a
polynomial factor in ; if for any
, minimax optimal detection can be attained within
constant factors in linear time. Using Schatten norm loss as a representative
example, we show that the hardness of attaining the minimax estimation rate can
crucially depend on the loss function. Implications on the hardness of support
recovery are also obtained.Comment: Published at http://dx.doi.org/10.1214/14-AOS1300 in the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Augmented L1 and Nuclear-Norm Models with a Globally Linearly Convergent Algorithm
This paper studies the long-existing idea of adding a nice smooth function to
"smooth" a non-differentiable objective function in the context of sparse
optimization, in particular, the minimization of
, where is a vector, as well as the
minimization of , where is a matrix and
and are the nuclear and Frobenius norms of ,
respectively. We show that they can efficiently recover sparse vectors and
low-rank matrices. In particular, they enjoy exact and stable recovery
guarantees similar to those known for minimizing and under
the conditions on the sensing operator such as its null-space property,
restricted isometry property, spherical section property, or RIPless property.
To recover a (nearly) sparse vector , minimizing
returns (nearly) the same solution as minimizing
almost whenever . The same relation also
holds between minimizing and minimizing
for recovering a (nearly) low-rank matrix , if . Furthermore, we show that the linearized Bregman algorithm for
minimizing subject to enjoys global
linear convergence as long as a nonzero solution exists, and we give an
explicit rate of convergence. The convergence property does not require a
solution solution or any properties on . To our knowledge, this is the best
known global convergence result for first-order sparse optimization algorithms.Comment: arXiv admin note: text overlap with arXiv:1207.5326 by other author
Minimising the number of gap-zeros in binary matrices
We study a problem of minimising the total number of zeros in the gaps between blocks of consecutive ones in the columns of a binary matrix by permuting its rows. The problem is referred to as the Consecutive Ones Matrix Augmentation Problem, and is known to be NP-hard. An analysis of the structure of an optimal solution allows us to focus on a restricted solution space, and to use an implicit representation for searching the space. We develop an exact solution algorithm, which is linear-time in the number of rows if the number of columns is constant, and two constructive heuristics to tackle instances with an arbitrary number of columns. The heuristics use a novel solution representation based upon row sequencing. In our computational study, all heuristic solutions are either optimal or close to an optimum. One of the heuristics is particularly effective, especially for problems with a large number of rows
- …