96 research outputs found

    Two essays in computational optimization: computing the clar number in fullerene graphs and distributing the errors in iterative interior point methods

    Get PDF
    Fullerene are cage-like hollow carbon molecules graph of pseudospherical sym- metry consisting of only pentagons and hexagons faces. It has been the object of interest for chemists and mathematicians due to its widespread application in various fields, namely including electronic and optic engineering, medical sci- ence and biotechnology. A Fullerene molecular, Γ n of n atoms has a multiplicity of isomers which increases as N iso ∼ O(n 9 ). For instance, Γ 180 has 79,538,751 isomers. The Fries and Clar numbers are stability predictors of a Fullerene molecule. These number can be computed by solving a (possibly N P -hard) combinatorial optimization problem. We propose several ILP formulation of such a problem each yielding a solution algorithm that provides the exact value of the Fries and Clar numbers. We compare the performances of the algorithm derived from the proposed ILP formulations. One of this algorithm is used to find the Clar isomers, i.e., those for which the Clar number is maximum among all isomers having a given size. We repeated this computational experiment for all sizes up to 204 atoms. In the course of the study a total of 2 649 413 774 isomers were analyzed.The second essay concerns developing an iterative primal dual infeasible path following (PDIPF) interior point (IP) algorithm for separable convex quadratic minimum cost flow network problem. In each iteration of PDIPF algorithm, the main computational effort is solving the underlying Newton search direction system. We concentrated on finding the solution of the corresponding linear system iteratively and inexactly. We assumed that all the involved inequalities can be solved inexactly and to this purpose, we focused on different approaches for distributing the error generated by iterative linear solvers such that the convergences of the PDIPF algorithm are guaranteed. As a result, we achieved theoretical bases that open the path to further interesting practical investiga- tion

    Embedding Graphs under Centrality Constraints for Network Visualization

    Full text link
    Visual rendering of graphs is a key task in the mapping of complex network data. Although most graph drawing algorithms emphasize aesthetic appeal, certain applications such as travel-time maps place more importance on visualization of structural network properties. The present paper advocates two graph embedding approaches with centrality considerations to comply with node hierarchy. The problem is formulated first as one of constrained multi-dimensional scaling (MDS), and it is solved via block coordinate descent iterations with successive approximations and guaranteed convergence to a KKT point. In addition, a regularization term enforcing graph smoothness is incorporated with the goal of reducing edge crossings. A second approach leverages the locally-linear embedding (LLE) algorithm which assumes that the graph encodes data sampled from a low-dimensional manifold. Closed-form solutions to the resulting centrality-constrained optimization problems are determined yielding meaningful embeddings. Experimental results demonstrate the efficacy of both approaches, especially for visualizing large networks on the order of thousands of nodes.Comment: Submitted to IEEE Transactions on Visualization and Computer Graphic

    Kernel-Based Feature Selection Techniques for Transport Proteins Based on Star Graph Topological Indices

    Get PDF
    [Abstract] The transport of the molecules inside cells is a very important topic, especially in Drug Metabolism. The experimental testing of the new proteins for the transporter molecular function is expensive and inefficient due to the large amount of new peptides. Therefore, there is a need for cheap and fast theoretical models to predict the transporter proteins. In the current work, the primary structure of a protein is represented as a molecular Star graph, characterized by a series of topological indices. The dataset was made up of 2,503 protein chains, out of which 413 have transporter molecular function and 2,090 have no transporter function. These indices were used as input to several classification techniques to find the best Quantitative Structure Activity Relationship (QSAR) model that can evaluate the transporter function of a new protein chain. Among several feature selection techniques, the Support Vector Machine Recursive Feature Elimination allows us to obtain a classification model based on 20 attributes with a true positive rate of 83% and a false positive rate of 16.7%.Xunta de Galicia; 1OSIN105004P

    Graph theory-based sequence descriptors as remote homology predictors

    Get PDF
    Indexación: Scopus.Alignment-free (AF) methodologies have increased in popularity in the last decades as alternative tools to alignment-based (AB) algorithms for performing comparative sequence analyses. They have been especially useful to detect remote homologs within the twilight zone of highly diverse gene/protein families and superfamilies. The most popular alignment-free methodologies, as well as their applications to classification problems, have been described in previous reviews. Despite a new set of graph theory-derived sequence/structural descriptors that have been gaining relevance in the detection of remote homology, they have been omitted as AF predictors when the topic is addressed. Here, we first go over the most popular AF approaches used for detecting homology signals within the twilight zone and then bring out the state-of-the-art tools encoding graph theory-derived sequence/structure descriptors and their success for identifying remote homologs. We also highlight the tendency of integrating AF features/measures with the AB ones, either into the same prediction model or by assembling the predictions from different algorithms using voting/weighting strategies, for improving the detection of remote signals. Lastly, we briefly discuss the efforts made to scale up AB and AF features/measures for the comparison of multiple genomes and proteomes. Alongside the achieved experiences in remote homology detection by both the most popular AF tools and other less known ones, we provide our own using the graphical–numerical methodologies, MARCH-INSIDE, TI2BioP, and ProtDCal. We also present a new Python-based tool (SeqDivA) with a friendly graphical user interface (GUI) for delimiting the twilight zone by using several similar criteria.https://www.mdpi.com/2218-273X/10/1/2

    Optimization Approaches for Open-Locating Dominating Sets

    Get PDF
    An Open Locating-Dominating Set (OLD set) is a subset of vertices in a graph such that every vertex in the graph has a neighbor in the OLD set and every vertex has a unique set of neighbors in the OLD set. This can also represent where sensors, capable of detecting an event occurrence at an adjacent vertex, could be placed such that one could always identify the location of an event by the specific vertices that indicated an event occurred in their neighborhood. By the open neighborhood construct, which differentiates OLD sets from identifying codes, a vertex is not able to report if it is the location of the event. This construct provides a robustness over identifying codes and opens new applications such as disease carrier and dark actor identification in networks. This work explores various aspects of OLD sets, beginning with an Integer Linear Program for quickly identifying the optimal OLD set on a graph. As many graphs do not admit OLD sets, or there may be times when the total size of the set is limited by an external factor, a concept called maximum covering OLD sets is developed and explored. The coverage radius of the sensors is then expanded in a presentation of Mixed-Weight OLD sets where sensors can cover more than just adjacent vertices. Finally, an application is presented to optimally monitor criminal and terrorist networks using OLD sets and related concepts to identify the optimal set of surveillance targets
    • …
    corecore