2,256 research outputs found

    Linear Time Parameterized Algorithms via Skew-Symmetric Multicuts

    Full text link
    A skew-symmetric graph (D=(V,A),σ)(D=(V,A),\sigma) is a directed graph DD with an involution σ\sigma on the set of vertices and arcs. In this paper, we introduce a separation problem, dd-Skew-Symmetric Multicut, where we are given a skew-symmetric graph DD, a family of T\cal T of dd-sized subsets of vertices and an integer kk. The objective is to decide if there is a set XAX\subseteq A of kk arcs such that every set JJ in the family has a vertex vv such that vv and σ(v)\sigma(v) are in different connected components of D=(V,A(Xσ(X))D'=(V,A\setminus (X\cup \sigma(X)). In this paper, we give an algorithm for this problem which runs in time O((4d)k(m+n+))O((4d)^{k}(m+n+\ell)), where mm is the number of arcs in the graph, nn the number of vertices and \ell the length of the family given in the input. Using our algorithm, we show that Almost 2-SAT has an algorithm with running time O(4kk4)O(4^kk^4\ell) and we obtain algorithms for {\sc Odd Cycle Transversal} and {\sc Edge Bipartization} which run in time O(4kk4(m+n))O(4^kk^4(m+n)) and O(4kk5(m+n))O(4^kk^5(m+n)) respectively. This resolves an open problem posed by Reed, Smith and Vetta [Operations Research Letters, 2003] and improves upon the earlier almost linear time algorithm of Kawarabayashi and Reed [SODA, 2010]. We also show that Deletion q-Horn Backdoor Set Detection is a special case of 3-Skew-Symmetric Multicut, giving us an algorithm for Deletion q-Horn Backdoor Set Detection which runs in time O(12kk5)O(12^kk^5\ell). This gives the first fixed-parameter tractable algorithm for this problem answering a question posed in a paper by a superset of the authors [STACS, 2013]. Using this result, we get an algorithm for Satisfiability which runs in time O(12kk5)O(12^kk^5\ell) where kk is the size of the smallest q-Horn deletion backdoor set, with \ell being the length of the input formula

    Looking for bSM physics using top-quark polarization and decay-lepton kinematic asymmetries

    Full text link
    We explore beyond Standard Model (bSM) physics signatures in the l+jetsl+jets channel of ttt\overline{t} pair production process at the Tevatron and the LHC. We study the effects of bSM physics scenarios on the top quark polarization and on the kinematics of the decay leptons. To this end, we construct asymmetries using the lepton energy and angular distributions. Further, we find their correlations with the top polarization, net charge asymmetry and top forward backward asymmetry. We show that used together, these observables can help discriminate effectively between SM and different bSM scenarios which can lead to varying degrees of top polarization at the Tevatron as well as the LHC. We use two types of coloured mediator models to demonstrate the effectiveness of proposed observables, an ss-channel axigluon and a uu-channel diquark.Comment: 39 pages, 10 figures and 4 tables. To appear in Phys. Rev.

    A Linear Time Parameterized Algorithm for Node Unique Label Cover

    Get PDF
    The optimization version of the Unique Label Cover problem is at the heart of the Unique Games Conjecture which has played an important role in the proof of several tight inapproximability results. In recent years, this problem has been also studied extensively from the point of view of parameterized complexity. Cygan et al. [FOCS 2012] proved that this problem is fixed-parameter tractable (FPT) and Wahlstr\"om [SODA 2014] gave an FPT algorithm with an improved parameter dependence. Subsequently, Iwata, Wahlstr\"om and Yoshida [2014] proved that the edge version of Unique Label Cover can be solved in linear FPT-time. That is, there is an FPT algorithm whose dependence on the input-size is linear. However, such an algorithm for the node version of the problem was left as an open problem. In this paper, we resolve this question by presenting the first linear-time FPT algorithm for Node Unique Label Cover

    Longitudinal top polarisation measurement and anomalous WtbWtb coupling

    Full text link
    Kinematical distributions of decay products of the top quark carry information on the polarisation of the top as well as on any possible new physics in the decay of the top quark. We construct observables in the form of asymmetries in the kinematical distributions to probe their effects. Charged-lepton angular distributions in the decay are insensitive to anomalous couplings to leading order. Hence these can be a robust probe of top polarisation. However, these are difficult to measure in the case of highly boosted top quarks as compared to energy distributions of decay products. These are then sensitive, in general, to both top polarisation and top anomalous couplings. We compare various asymmetries for their sensitivities to the longitudinal polarisation of the top quark as well as to possible new physics in the WtbWtb vertex, paying special attention to the case of highly boosted top quarks. We perform a χ2\chi ^2- analysis to determine the regions in the longitudinal polarisation of the top quark and the couplings of the WtbWtb vertex constrained by different combinations of the asymmetries. Moreover, we find that use of observables sensitive to the longitudinal top polarisation can add to the sensitivity to which the WtbWtb vertex can be probed.Comment: significantly revised version, clarifications on the term 'polarisation' added, new references added, the title modified, 41 figures. Accepted for publication in EPJ

    Lossy Kernelization

    Get PDF
    In this paper we propose a new framework for analyzing the performance of preprocessing algorithms. Our framework builds on the notion of kernelization from parameterized complexity. However, as opposed to the original notion of kernelization, our definitions combine well with approximation algorithms and heuristics. The key new definition is that of a polynomial size α\alpha-approximate kernel. Loosely speaking, a polynomial size α\alpha-approximate kernel is a polynomial time pre-processing algorithm that takes as input an instance (I,k)(I,k) to a parameterized problem, and outputs another instance (I,k)(I',k') to the same problem, such that I+kkO(1)|I'|+k' \leq k^{O(1)}. Additionally, for every c1c \geq 1, a cc-approximate solution ss' to the pre-processed instance (I,k)(I',k') can be turned in polynomial time into a (cα)(c \cdot \alpha)-approximate solution ss to the original instance (I,k)(I,k). Our main technical contribution are α\alpha-approximate kernels of polynomial size for three problems, namely Connected Vertex Cover, Disjoint Cycle Packing and Disjoint Factors. These problems are known not to admit any polynomial size kernels unless NPcoNP/polyNP \subseteq coNP/poly. Our approximate kernels simultaneously beat both the lower bounds on the (normal) kernel size, and the hardness of approximation lower bounds for all three problems. On the negative side we prove that Longest Path parameterized by the length of the path and Set Cover parameterized by the universe size do not admit even an α\alpha-approximate kernel of polynomial size, for any α1\alpha \geq 1, unless NPcoNP/polyNP \subseteq coNP/poly. In order to prove this lower bound we need to combine in a non-trivial way the techniques used for showing kernelization lower bounds with the methods for showing hardness of approximationComment: 58 pages. Version 2 contain new results: PSAKS for Cycle Packing and approximate kernel lower bounds for Set Cover and Hitting Set parameterized by universe siz
    corecore