3,549 research outputs found
Sequence embedding for fast construction of guide trees for multiple sequence alignment
<p>Abstract</p> <p>Background</p> <p>The most widely used multiple sequence alignment methods require sequences to be clustered as an initial step. Most sequence clustering methods require a full distance matrix to be computed between all pairs of sequences. This requires memory and time proportional to <it>N</it><sup>2 </sup>for <it>N </it>sequences. When <it>N </it>grows larger than 10,000 or so, this becomes increasingly prohibitive and can form a significant barrier to carrying out very large multiple alignments.</p> <p>Results</p> <p>In this paper, we have tested variations on a class of embedding methods that have been designed for clustering large numbers of complex objects where the individual distance calculations are expensive. These methods involve embedding the sequences in a space where the similarities within a set of sequences can be closely approximated without having to compute all pair-wise distances.</p> <p>Conclusions</p> <p>We show how this approach greatly reduces computation time and memory requirements for clustering large numbers of sequences and demonstrate the quality of the clusterings by benchmarking them as guide trees for multiple alignment. Source code is available for download from <url>http://www.clustal.org/mbed.tgz</url>.</p
One-class classifiers based on entropic spanning graphs
One-class classifiers offer valuable tools to assess the presence of outliers
in data. In this paper, we propose a design methodology for one-class
classifiers based on entropic spanning graphs. Our approach takes into account
the possibility to process also non-numeric data by means of an embedding
procedure. The spanning graph is learned on the embedded input data and the
outcoming partition of vertices defines the classifier. The final partition is
derived by exploiting a criterion based on mutual information minimization.
Here, we compute the mutual information by using a convenient formulation
provided in terms of the -Jensen difference. Once training is
completed, in order to associate a confidence level with the classifier
decision, a graph-based fuzzy model is constructed. The fuzzification process
is based only on topological information of the vertices of the entropic
spanning graph. As such, the proposed one-class classifier is suitable also for
data characterized by complex geometric structures. We provide experiments on
well-known benchmarks containing both feature vectors and labeled graphs. In
addition, we apply the method to the protein solubility recognition problem by
considering several representations for the input samples. Experimental results
demonstrate the effectiveness and versatility of the proposed method with
respect to other state-of-the-art approaches.Comment: Extended and revised version of the paper "One-Class Classification
Through Mutual Information Minimization" presented at the 2016 IEEE IJCNN,
Vancouver, Canad
Comparison of Distances for Supervised Segmentation of White Matter Tractography
Tractograms are mathematical representations of the main paths of axons
within the white matter of the brain, from diffusion MRI data. Such
representations are in the form of polylines, called streamlines, and one
streamline approximates the common path of tens of thousands of axons. The
analysis of tractograms is a task of interest in multiple fields, like
neurosurgery and neurology. A basic building block of many pipelines of
analysis is the definition of a distance function between streamlines. Multiple
distance functions have been proposed in the literature, and different authors
use different distances, usually without a specific reason other than invoking
the "common practice". To this end, in this work we want to test such common
practices, in order to obtain factual reasons for choosing one distance over
another. For these reasons, in this work we compare many streamline distance
functions available in the literature. We focus on the common task of automatic
bundle segmentation and we adopt the recent approach of supervised segmentation
from expert-based examples. Using the HCP dataset, we compare several distances
obtaining guidelines on the choice of which distance function one should use
for supervised bundle segmentation
Kernel methods in genomics and computational biology
Support vector machines and kernel methods are increasingly popular in
genomics and computational biology, due to their good performance in real-world
applications and strong modularity that makes them suitable to a wide range of
problems, from the classification of tumors to the automatic annotation of
proteins. Their ability to work in high dimension, to process non-vectorial
data, and the natural framework they provide to integrate heterogeneous data
are particularly relevant to various problems arising in computational biology.
In this chapter we survey some of the most prominent applications published so
far, highlighting the particular developments in kernel methods triggered by
problems in biology, and mention a few promising research directions likely to
expand in the future
- …