9,805 research outputs found

    Experimental study of the role of physicochemical surface processing on the IN ability of mineral dust particles

    Get PDF
    During the measurement campaign FROST 2 (FReezing Of duST 2), the Leipzig Aerosol Cloud Interaction Simulator (LACIS) was used to investigate the influence of various surface modifications on the ice nucleating ability of Arizona Test Dust (ATD) particles in the immersion freezing mode. The dust particles were exposed to sulfuric acid vapor, to water vapor with and without the addition of ammonia gas, and heat using a thermodenuder operating at 250 °C. Size selected, quasi monodisperse particles with a mobility diameter of 300 nm were fed into LACIS and droplets grew on these particles such that each droplet contained a single particle. Temperature dependent frozen fractions of these droplets were determined in a temperature range between −40 °C ≤T≤−28 °C. The pure ATD particles nucleated ice over a broad temperature range with their freezing behavior being separated into two freezing branches characterized through different slopes in the frozen fraction vs. temperature curves. Coating the ATD particles with sulfuric acid resulted in the particles' IN potential significantly decreasing in the first freezing branch (T>−35 °C) and a slight increase in the second branch (T≤−35 °C). The addition of water vapor after the sulfuric acid coating caused the disappearance of the first freezing branch and a strong reduction of the IN ability in the second freezing branch. The presence of ammonia gas during water vapor exposure had a negligible effect on the particles' IN ability compared to the effect of water vapor. Heating in the thermodenuder led to a decreased IN ability of the sulfuric acid coated particles for both branches but the additional heat did not or only slightly change the IN ability of the pure ATD and the water vapor exposed sulfuric acid coated particles. In other words, the combination of both sulfuric acid and water vapor being present is a main cause for the ice active surface features of the ATD particles being destroyed. A possible explanation could be the chemical transformation of ice active metal silicates to metal sulfates. The strongly enhanced reaction between sulfuric acid and dust in the presence of water vapor and the resulting significant reductions in IN potential are of importance for atmospheric ice cloud formation. Our findings suggest that the IN concentration can decrease by up to one order of magnitude for the conditions investigated

    Polynomial fixed-parameter algorithms : a case study for longest path on interval graphs.

    Get PDF
    We study the design of fixed-parameter algorithms for problems already known to be solvable in polynomial time. The main motivation is to get more efficient algorithms for problems with unattractive polynomial running times. Here, we focus on a fundamental graph problem: Longest Path; it is NP-hard in general but known to be solvable in O(n^4) time on n-vertex interval graphs. We show how to solve Longest Path on Interval Graphs, parameterized by vertex deletion number k to proper interval graphs, in O(k^9n) time. Notably, Longest Path is trivially solvable in linear time on proper interval graphs, and the parameter value k can be approximated up to a factor of 4 in linear time. From a more general perspective, we believe that using parameterized complexity analysis for polynomial-time solvable problems offers a very fertile ground for future studies for all sorts of algorithmic problems. It may enable a refined understanding of efficiency aspects for polynomial-time solvable problems, similarly to what classical parameterized complexity analysis does for NP-hard problems

    Parameterized Algorithmics for Computational Social Choice: Nine Research Challenges

    Full text link
    Computational Social Choice is an interdisciplinary research area involving Economics, Political Science, and Social Science on the one side, and Mathematics and Computer Science (including Artificial Intelligence and Multiagent Systems) on the other side. Typical computational problems studied in this field include the vulnerability of voting procedures against attacks, or preference aggregation in multi-agent systems. Parameterized Algorithmics is a subfield of Theoretical Computer Science seeking to exploit meaningful problem-specific parameters in order to identify tractable special cases of in general computationally hard problems. In this paper, we propose nine of our favorite research challenges concerning the parameterized complexity of problems appearing in this context

    Parameterized complexity of machine scheduling: 15 open problems

    Full text link
    Machine scheduling problems are a long-time key domain of algorithms and complexity research. A novel approach to machine scheduling problems are fixed-parameter algorithms. To stimulate this thriving research direction, we propose 15 open questions in this area whose resolution we expect to lead to the discovery of new approaches and techniques both in scheduling and parameterized complexity theory.Comment: Version accepted to Computers & Operations Researc

    Polynomial kernels for 3-leaf power graph modification problems

    Full text link
    A graph G=(V,E) is a 3-leaf power iff there exists a tree T whose leaves are V and such that (u,v) is an edge iff u and v are at distance at most 3 in T. The 3-leaf power graph edge modification problems, i.e. edition (also known as the closest 3-leaf power), completion and edge-deletion, are FTP when parameterized by the size of the edge set modification. However polynomial kernel was known for none of these three problems. For each of them, we provide cubic kernels that can be computed in linear time for each of these problems. We thereby answer an open problem first mentioned by Dom, Guo, Huffner and Niedermeier (2005).Comment: Submitte

    Towards optimal kernel for connected vertex cover in planar graphs

    Full text link
    We study the parameterized complexity of the connected version of the vertex cover problem, where the solution set has to induce a connected subgraph. Although this problem does not admit a polynomial kernel for general graphs (unless NP is a subset of coNP/poly), for planar graphs Guo and Niedermeier [ICALP'08] showed a kernel with at most 14k vertices, subsequently improved by Wang et al. [MFCS'11] to 4k. The constant 4 here is so small that a natural question arises: could it be already an optimal value for this problem? In this paper we answer this quesion in negative: we show a (11/3)k-vertex kernel for Connected Vertex Cover in planar graphs. We believe that this result will motivate further study in search for an optimal kernel

    Compression via Matroids: A Randomized Polynomial Kernel for Odd Cycle Transversal

    Full text link
    The Odd Cycle Transversal problem (OCT) asks whether a given graph can be made bipartite by deleting at most kk of its vertices. In a breakthrough result Reed, Smith, and Vetta (Operations Research Letters, 2004) gave a \BigOh(4^kkmn) time algorithm for it, the first algorithm with polynomial runtime of uniform degree for every fixed kk. It is known that this implies a polynomial-time compression algorithm that turns OCT instances into equivalent instances of size at most \BigOh(4^k), a so-called kernelization. Since then the existence of a polynomial kernel for OCT, i.e., a kernelization with size bounded polynomially in kk, has turned into one of the main open questions in the study of kernelization. This work provides the first (randomized) polynomial kernelization for OCT. We introduce a novel kernelization approach based on matroid theory, where we encode all relevant information about a problem instance into a matroid with a representation of size polynomial in kk. For OCT, the matroid is built to allow us to simulate the computation of the iterative compression step of the algorithm of Reed, Smith, and Vetta, applied (for only one round) to an approximate odd cycle transversal which it is aiming to shrink to size kk. The process is randomized with one-sided error exponentially small in kk, where the result can contain false positives but no false negatives, and the size guarantee is cubic in the size of the approximate solution. Combined with an \BigOh(\sqrt{\log n})-approximation (Agarwal et al., STOC 2005), we get a reduction of the instance to size \BigOh(k^{4.5}), implying a randomized polynomial kernelization.Comment: Minor changes to agree with SODA 2012 version of the pape
    corecore