27,474 research outputs found

    Space- and Time-Efficient Algorithm for Maintaining Dense Subgraphs on One-Pass Dynamic Streams

    Get PDF
    While in many graph mining applications it is crucial to handle a stream of updates efficiently in terms of {\em both} time and space, not much was known about achieving such type of algorithm. In this paper we study this issue for a problem which lies at the core of many graph mining applications called {\em densest subgraph problem}. We develop an algorithm that achieves time- and space-efficiency for this problem simultaneously. It is one of the first of its kind for graph problems to the best of our knowledge. In a graph G=(V,E)G = (V, E), the "density" of a subgraph induced by a subset of nodes SVS \subseteq V is defined as E(S)/S|E(S)|/|S|, where E(S)E(S) is the set of edges in EE with both endpoints in SS. In the densest subgraph problem, the goal is to find a subset of nodes that maximizes the density of the corresponding induced subgraph. For any ϵ>0\epsilon>0, we present a dynamic algorithm that, with high probability, maintains a (4+ϵ)(4+\epsilon)-approximation to the densest subgraph problem under a sequence of edge insertions and deletions in a graph with nn nodes. It uses O~(n)\tilde O(n) space, and has an amortized update time of O~(1)\tilde O(1) and a query time of O~(1)\tilde O(1). Here, O~\tilde O hides a O(\poly\log_{1+\epsilon} n) term. The approximation ratio can be improved to (2+ϵ)(2+\epsilon) at the cost of increasing the query time to O~(n)\tilde O(n). It can be extended to a (2+ϵ)(2+\epsilon)-approximation sublinear-time algorithm and a distributed-streaming algorithm. Our algorithm is the first streaming algorithm that can maintain the densest subgraph in {\em one pass}. The previously best algorithm in this setting required O(logn)O(\log n) passes [Bahmani, Kumar and Vassilvitskii, VLDB'12]. The space required by our algorithm is tight up to a polylogarithmic factor.Comment: A preliminary version of this paper appeared in STOC 201

    Properties of Nucleon Resonances by means of a Genetic Algorithm

    Get PDF
    We present an optimization scheme that employs a Genetic Algorithm (GA) to determine the properties of low-lying nucleon excitations within a realistic photo-pion production model based upon an effective Lagrangian. We show that with this modern optimization technique it is possible to reliably assess the parameters of the resonances and the associated error bars as well as to identify weaknesses in the models. To illustrate the problems the optimization process may encounter, we provide results obtained for the nucleon resonances Δ\Delta(1230) and Δ\Delta(1700). The former can be easily isolated and thus has been studied in depth, while the latter is not as well known experimentally.Comment: 12 pages, 10 figures, 3 tables. Minor correction

    Sub-structural Niching in Estimation of Distribution Algorithms

    Full text link
    We propose a sub-structural niching method that fully exploits the problem decomposition capability of linkage-learning methods such as the estimation of distribution algorithms and concentrate on maintaining diversity at the sub-structural level. The proposed method consists of three key components: (1) Problem decomposition and sub-structure identification, (2) sub-structure fitness estimation, and (3) sub-structural niche preservation. The sub-structural niching method is compared to restricted tournament selection (RTS)--a niching method used in hierarchical Bayesian optimization algorithm--with special emphasis on sustained preservation of multiple global solutions of a class of boundedly-difficult, additively-separable multimodal problems. The results show that sub-structural niching successfully maintains multiple global optima over large number of generations and does so with significantly less population than RTS. Additionally, the market share of each of the niche is much closer to the expected level in sub-structural niching when compared to RTS

    An Empirical Biomarker-based Calculator for Autosomal Recessive Polycystic Kidney Disease - The Nieto-Narayan Formula

    Full text link
    Autosomal polycystic kidney disease (ARPKD) is associated with progressive enlargement of the kidneys fuelled by the formation and expansion of fluid-filled cysts. The disease is congenital and children that do not succumb to it during the neonatal period will, by age 10 years, more often than not, require nephrectomy+renal replacement therapy for management of both pain and renal insufficiency. Since increasing cystic index (CI; percent of kidney occupied by cysts) drives both renal expansion and organ dysfunction, management of these patients, including decisions such as elective nephrectomy and prioritization on the transplant waitlist, could clearly benefit from serial determination of CI. So also, clinical trials in ARPKD evaluating the efficacy of novel drug candidates could benefit from serial determination of CI. Although ultrasound is currently the imaging modality of choice for diagnosis of ARPKD, its utilization for assessing disease progression is highly limited. Magnetic resonance imaging or computed tomography, although more reliable for determination of CI, are expensive, time-consuming and somewhat impractical in the pediatric population. Using a well-established mammalian model of ARPKD, we undertook a big data-like analysis of minimally- or non-invasive serum and urine biomarkers of renal injury/dysfunction to derive a family of equations for estimating CI. We then applied a signal averaging protocol to distill these equations to a single empirical formula for calculation of CI. Such a formula will eventually find use in identifying and monitoring patients at high risk for progressing to end-stage renal disease and aid in the conduct of clinical trials.Comment: 3 tables and 8 figure

    Single temperature for Monte Carlo optimization on complex landscapes

    Full text link
    We propose a new strategy for Monte Carlo (MC) optimization on rugged multidimensional landscapes. The strategy is based on querying the statistical properties of the landscape in order to find the temperature at which the mean first passage time across the current region of the landscape is minimized. Thus, in contrast to other algorithms such as simulated annealing (SA), we explicitly match the temperature schedule to the statistics of landscape irregularities. In cases where this statistics is approximately the same over the entire landscape, or where non-local moves couple distant parts of the landscape, single-temperature MC will outperform any other MC algorithm with the same move set. We also find that in strongly anisotropic Coulomb spin glass and traveling salesman problems, the only relevant statistics (which we use to assign a single MC temperature) is that of irregularities in low-energy funnels. Our results may explain why protein folding in nature is efficient at room temperatures.Comment: 5 pages, 3 figure

    Hydrogen and fluorine in the surfaces of lunar samples

    Get PDF
    The resonant nuclear reaction F-19 (p, alpha gamma)0-16 has been used to perform depth sensitive analyses for both fluorine and hydrogen in lunar samples. The resonance at 0.83 MeV (center-of-mass) in this reaction has been applied to the measurement of the distribution of trapped solar protons in lunar samples to depths of about 1/2 micrometer. These results are interpreted in terms of terrestrial H2O surface contamination and a redistribution of the implanted solar H which has been influenced by heavy radiation damage in the surface region. Results are also presented for an experiment to test the penetration of H2O into laboratory glass samples which have been irradiated with 0-16 to simulate the radiation damaged surfaces of lunar glasses. Fluorine determinations have been performed in a 1 pm surface layer on lunar samples using the same F-19 alpha gamma)0-16 resonance. The data are discussed from the standpoint of lunar fluorine and Teflon contamination
    corecore