191,683 research outputs found

    Construct redundancy in process modelling grammars: Improving the explanatory power of ontological analysis

    Get PDF
    Conceptual modelling supports developers and users of information systems in areas of documentation, analysis or system redesign. The ongoing interest in the modelling of business processes has led to a variety of different grammars, raising the question of the quality of these grammars for modelling. An established way of evaluating the quality of a modelling grammar is by means of an ontological analysis, which can determine the extent to which grammars contain construct deficit, overload, excess or redundancy. While several studies have shown the relevance of most of these criteria, predictions about construct redundancy have yielded inconsistent results in the past, with some studies suggesting that redundancy may even be beneficial for modelling in practice. In this paper we seek to contribute to clarifying the concept of construct redundancy by introducing a revision to the ontological analysis method. Based on the concept of inheritance we propose an approach that distinguishes between specialized and distinct construct redundancy. We demonstrate the potential explanatory power of the revised method by reviewing and clarifying previous results found in the literature

    Random Linear Fountain Code with Improved Decoding Success Probability

    Full text link
    In this paper we study the problem of increasing the decoding success probability of random linear fountain code over GF(2) for small packet lengths used in delay-intolerant applications such as multimedia streaming. Such code over GF(2) are attractive as they have lower decoding complexity than codes over larger field size, but suffer from high transmission redundancy. In our proposed coding scheme we construct a codeword which is not a linear combination of any codewords previously transmitted to mitigate such transmission redundancy. We then note the observation that the probability of receiving a linearly dependent codeword is highest when the receiver has received k-1 linearly independent codewords. We propose using the BlockACK frame so that the codeword received after k-1 linearly independent codeword is always linearly independent, this reduces the expected redundancy by a factor of three.Comment: This paper appears in: Communications (APCC), 2016 22nd Asia-Pacific Conference o

    Redundancy, Deduction Schemes, and Minimum-Size Bases for Association Rules

    Full text link
    Association rules are among the most widely employed data analysis methods in the field of Data Mining. An association rule is a form of partial implication between two sets of binary variables. In the most common approach, association rules are parameterized by a lower bound on their confidence, which is the empirical conditional probability of their consequent given the antecedent, and/or by some other parameter bounds such as "support" or deviation from independence. We study here notions of redundancy among association rules from a fundamental perspective. We see each transaction in a dataset as an interpretation (or model) in the propositional logic sense, and consider existing notions of redundancy, that is, of logical entailment, among association rules, of the form "any dataset in which this first rule holds must obey also that second rule, therefore the second is redundant". We discuss several existing alternative definitions of redundancy between association rules and provide new characterizations and relationships among them. We show that the main alternatives we discuss correspond actually to just two variants, which differ in the treatment of full-confidence implications. For each of these two notions of redundancy, we provide a sound and complete deduction calculus, and we show how to construct complete bases (that is, axiomatizations) of absolutely minimum size in terms of the number of rules. We explore finally an approach to redundancy with respect to several association rules, and fully characterize its simplest case of two partial premises.Comment: LMCS accepted pape

    Tight Bounds for Gomory-Hu-like Cut Counting

    Full text link
    By a classical result of Gomory and Hu (1961), in every edge-weighted graph G=(V,E,w)G=(V,E,w), the minimum stst-cut values, when ranging over all s,tVs,t\in V, take at most V1|V|-1 distinct values. That is, these (V2)\binom{|V|}{2} instances exhibit redundancy factor Ω(V)\Omega(|V|). They further showed how to construct from GG a tree (V,E,w)(V,E',w') that stores all minimum stst-cut values. Motivated by this result, we obtain tight bounds for the redundancy factor of several generalizations of the minimum stst-cut problem. 1. Group-Cut: Consider the minimum (A,B)(A,B)-cut, ranging over all subsets A,BVA,B\subseteq V of given sizes A=α|A|=\alpha and B=β|B|=\beta. The redundancy factor is Ωα,β(V)\Omega_{\alpha,\beta}(|V|). 2. Multiway-Cut: Consider the minimum cut separating every two vertices of SVS\subseteq V, ranging over all subsets of a given size S=k|S|=k. The redundancy factor is Ωk(V)\Omega_{k}(|V|). 3. Multicut: Consider the minimum cut separating every demand-pair in DV×VD\subseteq V\times V, ranging over collections of D=k|D|=k demand pairs. The redundancy factor is Ωk(Vk)\Omega_{k}(|V|^k). This result is a bit surprising, as the redundancy factor is much larger than in the first two problems. A natural application of these bounds is to construct small data structures that stores all relevant cut values, like the Gomory-Hu tree. We initiate this direction by giving some upper and lower bounds.Comment: This version contains additional references to previous work (which have some overlap with our results), see Bibliographic Update 1.

    Codes Correcting Two Deletions

    Full text link
    In this work, we investigate the problem of constructing codes capable of correcting two deletions. In particular, we construct a code that requires redundancy approximately 8 log n + O(log log n) bits of redundancy, where n is the length of the code. To the best of the author's knowledge, this represents the best known construction in that it requires the lowest number of redundant bits for a code correcting two deletions

    Systematic Codes for Rank Modulation

    Get PDF
    The goal of this paper is to construct systematic error-correcting codes for permutations and multi-permutations in the Kendall's τ\tau-metric. These codes are important in new applications such as rank modulation for flash memories. The construction is based on error-correcting codes for multi-permutations and a partition of the set of permutations into error-correcting codes. For a given large enough number of information symbols kk, and for any integer tt, we present a construction for (k+r,k){(k+r,k)} systematic tt-error-correcting codes, for permutations from Sk+rS_{k+r}, with less redundancy symbols than the number of redundancy symbols in the codes of the known constructions. In particular, for a given tt and for sufficiently large kk we can obtain r=t+1r=t+1. The same construction is also applied to obtain related systematic error-correcting codes for multi-permutations.Comment: to be presented ISIT201

    Universal Quantum Computation using Exchange Interactions and Teleportation of Single-Qubit Operations

    Get PDF
    We show how to construct a universal set of quantum logic gates using control over exchange interactions and single- and two-spin measurements only. Single-spin unitary operations are teleported instead of being executed directly, thus eliminating a major difficulty in the construction of several of the most promising proposals for solid-state quantum computation, such as spin-coupled quantum dots, donor-atom nuclear spins in silicon, and electrons on helium. Contrary to previous proposals dealing with this difficulty, our scheme requires no encoding redundancy. We also discuss an application to superconducting phase qubits.Comment: 4.5 pages, including 2 figure
    corecore