47 research outputs found
Algorithms for Matrix Multiplication via Sampling and Opportunistic Matrix Multiplication
Karppa & Kaski (2019) proposed a novel type of ``broken" or ``opportunistic"
multiplication algorithm, based on a variant of Strassen's algorithm, and used
this to develop new algorithms for Boolean matrix multiplication, among other
tasks. For instance, their algorithm can compute Boolean matrix multiplication
in time. While faster matrix
multiplication algorithms exist asymptotically, in practice most such
algorithms are infeasible for practical problems.
In this note, we describe an alternate way to use the broken matrix
multiplication algorithm to approximately compute matrix multiplication, either
for real-valued matrices or Boolean matrices. In brief, instead of running
multiple iterations of the broken algorithm on the original input matrix, we
form a new larger matrix by sampling and run a single iteration of the broken
algorithm. Asymptotically, the resulting algorithm has runtime , a slight improvement of
Karppa-Kaski's algorithm.
Since the goal is to obtain new practical matrix-multiplication algorithms,
these asymptotic runtime bounds are not directly useful. We estimate the
runtime for our algorithm for some sample problems which are at the upper
limits of practical algorithms; unfortunately, for these parameters, the new
algorithm does not appear to be beneficial
Limits on the Universal Method for Matrix Multiplication
In this work, we prove limitations on the known methods for designing matrix multiplication algorithms. Alman and Vassilevska Williams [Alman and Williams, 2018] recently defined the Universal Method, which substantially generalizes all the known approaches including Strassen\u27s Laser Method [V. Strassen, 1987] and Cohn and Umans\u27 Group Theoretic Method [Cohn and Umans, 2003]. We prove concrete lower bounds on the algorithms one can design by applying the Universal Method to many different tensors. Our proofs use new tools for upper bounding the asymptotic slice rank of a wide range of tensors. Our main result is that the Universal method applied to any Coppersmith-Winograd tensor CW_q cannot yield a bound on omega, the exponent of matrix multiplication, better than 2.16805. By comparison, it was previously only known that the weaker "Galactic Method" applied to CW_q could not achieve an exponent of 2.
We also study the Laser Method (which is, in principle, a highly special case of the Universal Method) and prove that it is "complete" for matrix multiplication algorithms: when it applies to a tensor T, it achieves omega = 2 if and only if it is possible for the Universal method applied to T to achieve omega = 2. Hence, the Laser Method, which was originally used as an algorithmic tool, can also be seen as a lower bounding tool. For example, in their landmark paper, Coppersmith and Winograd [Coppersmith and Winograd, 1990] achieved a bound of omega <= 2.376, by applying the Laser Method to CW_q. By our result, the fact that they did not achieve omega=2 implies a lower bound on the Universal Method applied to CW_q. Indeed, if it were possible for the Universal Method applied to CW_q to achieve omega=2, then Coppersmith and Winograd\u27s application of the Laser Method would have achieved omega=2
The Asymptotic Rank Conjecture and the Set Cover Conjecture are not Both True
Strassen's asymptotic rank conjecture [Progr. Math. 120 (1994)] claims a
strong submultiplicative upper bound on the rank of a three-tensor obtained as
an iterated Kronecker product of a constant-size base tensor. The conjecture,
if true, most notably would put square matrix multiplication in quadratic time.
We note here that some more-or-less unexpected algorithmic results in the area
of exponential-time algorithms would also follow. Specifically, we study the
so-called set cover conjecture, which states that for any there
exists a positive integer constant such that no algorithm solves the
-Set Cover problem in worst-case time . The -Set Cover problem asks, given as input an
-element universe , a family of size-at-most- subsets of
, and a positive integer , whether there is a subfamily of at most
sets in whose union is . The conjecture was formulated by Cygan
et al. in the monograph Parameterized Algorithms [Springer, 2015] but was
implicit as a hypothesis already in Cygan et al. [CCC 2012, ACM Trans.
Algorithms 2016], there conjectured to follow from the Strong Exponential Time
Hypothesis. We prove that if the asymptotic rank conjecture is true, then the
set cover conjecture is false. Using a reduction by Krauthgamer and Trabelsi
[STACS 2019], in this scenario we would also get a
-time randomized algorithm for some constant
for another well-studied problem for which no such algorithm is
known, namely that of deciding whether a given -vertex directed graph has a
Hamiltonian cycle
Faster Matrix Multiplication via Asymmetric Hashing
Fast matrix multiplication is one of the most fundamental problems in
algorithm research. The exponent of the optimal time complexity of matrix
multiplication is usually denoted by . This paper discusses new ideas
for improving the laser method for fast matrix multiplication. We observe that
the analysis of higher powers of the Coppersmith-Winograd tensor [Coppersmith &
Winograd 1990] incurs a "combination loss", and we partially compensate for it
using an asymmetric version of CW's hashing method. By analyzing the eighth
power of the CW tensor, we give a new bound of , which improves
the previous best bound of [Alman & Vassilevska Williams
2020]. Our result breaks the lower bound of in [Ambainis, Filmus & Le
Gall 2015] because of the new method for analyzing component (constituent)
tensors.Comment: 67 page
Fundamentals
Volume 1 establishes the foundations of this new field. It goes through all the steps from data collection, their summary and clustering, to different aspects of resource-aware learning, i.e., hardware, memory, energy, and communication awareness. Machine learning methods are inspected with respect to resource requirements and how to enhance scalability on diverse computing architectures ranging from embedded systems to large computing clusters
Recommended from our members
Learning and validating clinically meaningful phenotypes from electronic health data
The ever-growing adoption of electronic health records (EHR) to record patients' health journeys has resulted in vast amounts of heterogeneous, complex, and unwieldy information [Hripcsak and Albers, 2013]. Distilling this raw data into clinical insights presents great opportunities and challenges for the research and medical communities. One approach to this distillation is called computational phenotyping. Computational phenotyping is the process of extracting clinically relevant and interesting characteristics from a set of clinical documentation, such as that which is recorded in electronic health records (EHRs). Clinicians can use computational phenotyping, which can be viewed as a form of dimensionality reduction where a set of phenotypes form a latent space, to reason about populations, identify patients for randomized case-control studies, and extrapolate patient disease trajectories. In recent years, high-throughput computational approaches have made strides in extracting potentially clinically interesting phenotypes from data contained in EHR systems.
Tensor factorization methods have shown particular promise in deriving phenotypes. However, phenotyping methods via tensor factorization have the following weaknesses: 1) the extracted phenotypes can lack diversity, which makes them more difficult for clinicians to reason about and utilize in practice, 2) many of the tensor factorization methods are unsupervised and do not utilize side information that may be available about the population or about the relationships between the clinical characteristics in the data (e.g., diagnoses and medications), and 3) validating the clinical relevance of the extracted phenotypes requires domain training and expertise. This dissertation addresses all three of these limitations. First, we present tensor factorization methods that discover sparse and concise phenotypes in unsupervised, supervised, and semi-supervised settings. Second, via two tools we built, we show how to leverage domain expertise in the form of publicly available medical articles to evaluate the clinical validity of the discovered phenotypes. Third, we combine tensor factorization and the phenotype validation tools to guide the discovery process to more clinically relevant phenotypes.Computational Science, Engineering, and Mathematic
Fundamentals
Volume 1 establishes the foundations of this new field. It goes through all the steps from data collection, their summary and clustering, to different aspects of resource-aware learning, i.e., hardware, memory, energy, and communication awareness. Machine learning methods are inspected with respect to resource requirements and how to enhance scalability on diverse computing architectures ranging from embedded systems to large computing clusters
Programming Languages and Systems
This open access book constitutes the proceedings of the 31st European Symposium on Programming, ESOP 2022, which was held during April 5-7, 2022, in Munich, Germany, as part of the European Joint Conferences on Theory and Practice of Software, ETAPS 2022. The 21 regular papers presented in this volume were carefully reviewed and selected from 64 submissions. They deal with fundamental issues in the specification, design, analysis, and implementation of programming languages and systems
Programming Languages and Systems
This open access book constitutes the proceedings of the 31st European Symposium on Programming, ESOP 2022, which was held during April 5-7, 2022, in Munich, Germany, as part of the European Joint Conferences on Theory and Practice of Software, ETAPS 2022. The 21 regular papers presented in this volume were carefully reviewed and selected from 64 submissions. They deal with fundamental issues in the specification, design, analysis, and implementation of programming languages and systems
Applied Metaheuristic Computing
For decades, Applied Metaheuristic Computing (AMC) has been a prevailing optimization technique for tackling perplexing engineering and business problems, such as scheduling, routing, ordering, bin packing, assignment, facility layout planning, among others. This is partly because the classic exact methods are constrained with prior assumptions, and partly due to the heuristics being problem-dependent and lacking generalization. AMC, on the contrary, guides the course of low-level heuristics to search beyond the local optimality, which impairs the capability of traditional computation methods. This topic series has collected quality papers proposing cutting-edge methodology and innovative applications which drive the advances of AMC