70,242 research outputs found
Cluster Computing and the Power of Edge Recognition
We study the robustness--the invariance under definition changes--of the
cluster class CL#P [HHKW05]. This class contains each #P function that is
computed by a balanced Turing machine whose accepting paths always form a
cluster with respect to some length-respecting total order with efficient
adjacency checks. The definition of CL#P is heavily influenced by the defining
paper's focus on (global) orders. In contrast, we define a cluster class,
CLU#P, to capture what seems to us a more natural model of cluster computing.
We prove that the naturalness is costless: CL#P = CLU#P. Then we exploit the
more natural, flexible features of CLU#P to prove new robustness results for
CL#P and to expand what is known about the closure properties of CL#P.
The complexity of recognizing edges--of an ordered collection of computation
paths or of a cluster of accepting computation paths--is central to this study.
Most particularly, our proofs exploit the power of unique discovery of
edges--the ability of nondeterministic functions to, in certain settings,
discover on exactly one (in some cases, on at most one) computation path a
critical piece of information regarding edges of orderings or clusters
Knowledge is at the Edge! How to Search in Distributed Machine Learning Models
With the advent of the Internet of Things and Industry 4.0 an enormous amount
of data is produced at the edge of the network. Due to a lack of computing
power, this data is currently send to the cloud where centralized machine
learning models are trained to derive higher level knowledge. With the recent
development of specialized machine learning hardware for mobile devices, a new
era of distributed learning is about to begin that raises a new research
question: How can we search in distributed machine learning models? Machine
learning at the edge of the network has many benefits, such as low-latency
inference and increased privacy. Such distributed machine learning models can
also learn personalized for a human user, a specific context, or application
scenario. As training data stays on the devices, control over possibly
sensitive data is preserved as it is not shared with a third party. This new
form of distributed learning leads to the partitioning of knowledge between
many devices which makes access difficult. In this paper we tackle the problem
of finding specific knowledge by forwarding a search request (query) to a
device that can answer it best. To that end, we use a entropy based quality
metric that takes the context of a query and the learning quality of a device
into account. We show that our forwarding strategy can achieve over 95%
accuracy in a urban mobility scenario where we use data from 30 000 people
commuting in the city of Trento, Italy.Comment: Published in CoopIS 201
Pattern vectors from algebraic graph theory
Graphstructures have proven computationally cumbersome for pattern analysis. The reason for this is that, before graphs can be converted to pattern vectors, correspondences must be established between the nodes of structures which are potentially of different size. To overcome this problem, in this paper, we turn to the spectral decomposition of the Laplacian matrix. We show how the elements of the spectral matrix for the Laplacian can be used to construct symmetric polynomials that are permutation invariants. The coefficients of these polynomials can be used as graph features which can be encoded in a vectorial manner. We extend this representation to graphs in which there are unary attributes on the nodes and binary attributes on the edges by using the spectral decomposition of a Hermitian property matrix that can be viewed as a complex analogue of the Laplacian. To embed the graphs in a pattern space, we explore whether the vectors of invariants can be embedded in a low- dimensional space using a number of alternative strategies, including principal components analysis ( PCA), multidimensional scaling ( MDS), and locality preserving projection ( LPP). Experimentally, we demonstrate that the embeddings result in well- defined graph clusters. Our experiments with the spectral representation involve both synthetic and real- world data. The experiments with synthetic data demonstrate that the distances between spectral feature vectors can be used to discriminate between graphs on the basis of their structure. The real- world experiments show that the method can be used to locate clusters of graphs
- …