137 research outputs found
The Application of Hybridized Genetic Algorithms to the Protein Folding Problem
The protein folding problem consists of attempting to determine the native conformation of a protein given its primary structure. This study examines various methods of hybridizing a genetic algorithm implementation in order to minimize an energy function and predict the conformation (structure) of Met-enkephalin. Genetic Algorithms are semi-optimal algorithms designed to explore and exploit a search space. The genetic algorithm uses selection, recombination, and mutation operators on populations of strings which represent possible solutions to the given problem. One step in solving the protein folding problem is the design of efficient energy minimization techniques. A conjugate gradient minimization technique is described and tested with different replacement frequencies. Baidwinian, Lamarckian, and probabilistic Lamarckian evolution are all tested. Another extension of simple genetic algorithms can be accomplished with niching. Niching works by de-emphasizing solutions based on their proximity to other solutions in the space. Several variations of niching are tested. Experiments are conducted to determine the benefits of each hybridization technique versus each other and versus the genetic algorithm by itself. The experiments are geared toward trying to find the lowest possible energy and hence the minimum conformation of Met-enkephalin. In the experiments, probabilistic Lamarckian strategies were successful in achieving energies below that of the published minimum in QUANTA
Weisfeiler and Leman go Hyperbolic: Learning Distance Preserving Node Representations
In recent years, graph neural networks (GNNs) have emerged as a promising
tool for solving machine learning problems on graphs. Most GNNs are members of
the family of message passing neural networks (MPNNs). There is a close
connection between these models and the Weisfeiler-Leman (WL) test of
isomorphism, an algorithm that can successfully test isomorphism for a broad
class of graphs. Recently, much research has focused on measuring the
expressive power of GNNs. For instance, it has been shown that standard MPNNs
are at most as powerful as WL in terms of distinguishing non-isomorphic graphs.
However, these studies have largely ignored the distances between the
representations of nodes/graphs which are of paramount importance for learning
tasks. In this paper, we define a distance function between nodes which is
based on the hierarchy produced by the WL algorithm, and propose a model that
learns representations which preserve those distances between nodes. Since the
emerging hierarchy corresponds to a tree, to learn these representations, we
capitalize on recent advances in the field of hyperbolic neural networks. We
empirically evaluate the proposed model on standard node and graph
classification datasets where it achieves competitive performance with
state-of-the-art models
Directional naive Bayes classifiers
Directional data are ubiquitous in science.
These data have some special properties that rule out the
use of classical statistics. Therefore, different distributions
and statistics, such as the univariate von Mises and the
multivariate von Mises–Fisher distributions, should be
used to deal with this kind of information. We extend the
naive Bayes classifier to the case where the conditional
probability distributions of the predictive variables follow
either of these distributions. We consider the simple scenario,
where only directional predictive variables are used,
and the hybrid case, where discrete, Gaussian and directional
distributions are mixed. The classifier decision
functions and their decision surfaces are studied at length.
Artificial examples are used to illustrate the behavior of the
classifiers. The proposed classifiers are then evaluated over
eight datasets, showing competitive performances against
other naive Bayes classifiers that use Gaussian distributions
or discretization to manage directional data
Constructive Approximation and Learning by Greedy Algorithms
This thesis develops several kernel-based greedy algorithms for different machine learning problems and analyzes their theoretical and empirical properties. Greedy approaches have been extensively used in the past for tackling problems in combinatorial optimization where finding even a feasible solution can be a computationally hard problem (i.e., not solvable in polynomial time). A key feature of greedy algorithms is that a solution is constructed recursively from the smallest constituent parts. In each step of the constructive process a component is added to the partial solution from the previous step and, thus, the size of the optimization problem is reduced. The selected components are given by optimization problems that are simpler and easier to solve than the original problem. As such schemes are typically fast at constructing a solution they can be very effective on complex optimization problems where finding an optimal/good solution has a high computational cost. Moreover, greedy solutions are rather intuitive and the schemes themselves are simple to design and easy to implement. There is a large class of problems for which greedy schemes generate an optimal solution or a good approximation of the optimum. In the first part of the thesis, we develop two deterministic greedy algorithms for optimization problems in which a solution is given by a set of functions mapping an instance space to the space of reals. The first of the two approaches facilitates data understanding through interactive visualization by providing means for experts to incorporate their domain knowledge into otherwise static kernel principal component analysis. This is achieved by greedily constructing embedding directions that maximize the variance at data points (unexplained by the previously constructed embedding directions) while adhering to specified domain knowledge constraints. The second deterministic greedy approach is a supervised feature construction method capable of addressing the problem of kernel choice. The goal of the approach is to construct a feature representation for which a set of linear hypotheses is of sufficient capacity — large enough to contain a satisfactory solution to the considered problem and small enough to allow good generalization from a small number of training examples. The approach mimics functional gradient descent and constructs features by fitting squared error residuals. We show that the constructive process is consistent and provide conditions under which it converges to the optimal solution. In the second part of the thesis, we investigate two problems for which deterministic greedy schemes can fail to find an optimal solution or a good approximation of the optimum. This happens as a result of making a sequence of choices which take into account only the immediate reward without considering the consequences onto future decisions. To address this shortcoming of deterministic greedy schemes, we propose two efficient randomized greedy algorithms which are guaranteed to find effective solutions to the corresponding problems. In the first of the two approaches, we provide a mean to scale kernel methods to problems with millions of instances. An approach, frequently used in practice, for this type of problems is the Nyström method for low-rank approximation of kernel matrices. A crucial step in this method is the choice of landmarks which determine the quality of the approximation. We tackle this problem with a randomized greedy algorithm based on the K-means++ cluster seeding scheme and provide a theoretical and empirical study of its effectiveness. In the second problem for which a deterministic strategy can fail to find a good solution, the goal is to find a set of objects from a structured space that are likely to exhibit an unknown target property. This discrete optimization problem is of significant interest to cyclic discovery processes such as de novo drug design. We propose to address it with an adaptive Metropolis–Hastings approach that samples candidates from the posterior distribution of structures conditioned on them having the target property. The proposed constructive scheme defines a consistent random process and our empirical evaluation demonstrates its effectiveness across several different application domains
RNAG: a new Gibbs sampler for predicting RNA secondary structure for unaligned sequences
Motivation: RNA secondary structure plays an important role in the function of many RNAs, and structural features are often key to their interaction with other cellular components. Thus, there has been considerable interest in the prediction of secondary structures for RNA families. In this article, we present a new global structural alignment algorithm, RNAG, to predict consensus secondary structures for unaligned sequences. It uses a blocked Gibbs sampling algorithm, which has a theoretical advantage in convergence time. This algorithm iteratively samples from the conditional probability distributions P(Structure | Alignment) and P(Alignment | Structure). Not surprisingly, there is considerable uncertainly in the high-dimensional space of this difficult problem, which has so far received limited attention in this field. We show how the samples drawn from this algorithm can be used to more fully characterize the posterior space and to assess the uncertainty of predictions
Development and application of deep learning and spatial statistics within 3D bone marrow imaging
The bone marrow is a highly specialised organ, responsible for the formation of blood cells. Despite 50 years of research, the spatial organisation of the bone marrow remains an area full of controversy and contradiction. One reason for this is that imaging of bone marrow tissue is notoriously difficult. Secondly, efficient methodologies to fully extract and analyse large datasets remain the Achilles heels of imaging-based research. In this thesis I present a pipeline for generating 3D bone marrow images followed by the large-scale data extraction and spatial statistical analysis of the resulting data. Using these techniques, in the context of 3D imaging, I am able to identify and classify the location of hundreds of thousands of cells within various bone marrow samples. I then introduce a series of statistical techniques tailored to work with spatial data, resulting in a 3D statistical map of the tissue from which multi-cellular interactions can be clearly understood. As an illustration of the power of this new approach, I apply this pipeline to diseased samples of bone marrow with a particular focus on leukaemia and its interactions with CD8+ T cells. In so doing I show that this novel pipeline can be used to unravel complex multi-cellular interactions and assist researchers in understanding the processes taking place within the bone marrow.Open Acces
Pattern recognition and machine learning for magnetic resonance images with kernel methods
The aim of this thesis is to apply a particular category of machine learning and
pattern recognition algorithms, namely the kernel methods, to both functional and
anatomical magnetic resonance images (MRI). This work specifically focused on
supervised learning methods. Both methodological and practical aspects are described
in this thesis.
Kernel methods have the computational advantage for high dimensional data,
therefore they are idea for imaging data. The procedures can be broadly divided into
two components: the construction of the kernels and the actual kernel algorithms
themselves. Pre-processed functional or anatomical images can be computed into a
linear kernel or a non-linear kernel. We introduce both kernel regression and kernel
classification algorithms in two main categories: probabilistic methods and
non-probabilistic methods. For practical applications, kernel classification methods
were applied to decode the cognitive or sensory states of the subject from the fMRI
signal and were also applied to discriminate patients with neurological diseases from
normal people using anatomical MRI. Kernel regression methods were used to predict
the regressors in the design of fMRI experiments, and clinical ratings from the
anatomical scans
- …