95,424 research outputs found
Fair Evaluation of Global Network Aligners
Biological network alignment identifies topologically and functionally
conserved regions between networks of different species. It encompasses two
algorithmic steps: node cost function (NCF), which measures similarities
between nodes in different networks, and alignment strategy (AS), which uses
these similarities to rapidly identify high-scoring alignments. Different
methods use both different NCFs and different ASs. Thus, it is unclear whether
the superiority of a method comes from its NCF, its AS, or both. We already
showed on MI-GRAAL and IsoRankN that combining NCF of one method and AS of
another method can lead to a new superior method. Here, we evaluate MI-GRAAL
against newer GHOST to potentially further improve alignment quality. Also, we
approach several important questions that have not been asked systematically
thus far. First, we ask how much of the node similarity information in NCF
should come from sequence data compared to topology data. Existing methods
determine this more-less arbitrarily, which could affect the resulting
alignment(s). Second, when topology is used in NCF, we ask how large the size
of the neighborhoods of the compared nodes should be. Existing methods assume
that larger neighborhood sizes are better.
We find that MI-GRAAL's NCF is superior to GHOST's NCF, while the performance
of the methods' ASs is data-dependent. Thus, the combination of MI-GRAAL's NCF
and GHOST's AS could be a new superior method for certain data. Also, which
amount of sequence information is used within NCF does not affect alignment
quality, while the inclusion of topological information is crucial. Finally,
larger neighborhood sizes are preferred, but often, it is the second largest
size that is superior, and using this size would decrease computational
complexity.
Together, our results give several general recommendations for a fair
evaluation of network alignment methods.Comment: 19 pages. 10 figures. Presented at the 2014 ISMB Conference, July
13-15, Boston, M
Data-driven network alignment
Biological network alignment (NA) aims to find a node mapping between
species' molecular networks that uncovers similar network regions, thus
allowing for transfer of functional knowledge between the aligned nodes.
However, current NA methods do not end up aligning functionally related nodes.
A likely reason is that they assume it is topologically similar nodes that are
functionally related. However, we show that this assumption does not hold well.
So, a paradigm shift is needed with how the NA problem is approached. We
redefine NA as a data-driven framework, TARA (daTA-dRiven network Alignment),
which attempts to learn the relationship between topological relatedness and
functional relatedness without assuming that topological relatedness
corresponds to topological similarity, like traditional NA methods do. TARA
trains a classifier to predict whether two nodes from different networks are
functionally related based on their network topological patterns. We find that
TARA is able to make accurate predictions. TARA then takes each pair of nodes
that are predicted as related to be part of an alignment. Like traditional NA
methods, TARA uses this alignment for the across-species transfer of functional
knowledge. Clearly, TARA as currently implemented uses topological but not
protein sequence information for this task. We find that TARA outperforms
existing state-of-the-art NA methods that also use topological information,
WAVE and SANA, and even outperforms or complements a state-of-the-art NA method
that uses both topological and sequence information, PrimAlign. Hence, adding
sequence information to TARA, which is our future work, is likely to further
improve its performance
Direct Feedback Alignment with Sparse Connections for Local Learning
Recent advances in deep neural networks (DNNs) owe their success to training
algorithms that use backpropagation and gradient-descent. Backpropagation,
while highly effective on von Neumann architectures, becomes inefficient when
scaling to large networks. Commonly referred to as the weight transport
problem, each neuron's dependence on the weights and errors located deeper in
the network require exhaustive data movement which presents a key problem in
enhancing the performance and energy-efficiency of machine-learning hardware.
In this work, we propose a bio-plausible alternative to backpropagation drawing
from advances in feedback alignment algorithms in which the error computation
at a single synapse reduces to the product of three scalar values. Using a
sparse feedback matrix, we show that a neuron needs only a fraction of the
information previously used by the feedback alignment algorithms. Consequently,
memory and compute can be partitioned and distributed whichever way produces
the most efficient forward pass so long as a single error can be delivered to
each neuron. Our results show orders of magnitude improvement in data movement
and improvement in multiply-and-accumulate operations over
backpropagation. Like previous work, we observe that any variant of feedback
alignment suffers significant losses in classification accuracy on deep
convolutional neural networks. By transferring trained convolutional layers and
training the fully connected layers using direct feedback alignment, we
demonstrate that direct feedback alignment can obtain results competitive with
backpropagation. Furthermore, we observe that using an extremely sparse
feedback matrix, rather than a dense one, results in a small accuracy drop
while yielding hardware advantages. All the code and results are available
under https://github.com/bcrafton/ssdfa.Comment: 15 pages, 8 figure
- …