306 research outputs found

    Finite Automata Algorithms in Map-Reduce

    Get PDF
    In this thesis the intersection of several large nondeterministic finite automata (NFA's) as well as minimization of a large deterministic finite automaton (DFA) in map-reduce are studied. We have derived a lower bound on replication rate for computing NFA intersections and provided three concrete algorithms for the problem. Our investigation of the replication rate for each of all three algorithms shows where each algorithm could be applied through detailed experiments on large datasets of finite automata. Denoting n the number of states in DFA A, we propose an algorithm to minimize A in n map-reduce rounds in the worst-case. Our experiments, however, indicate that the number of rounds, in practice, is much smaller than n for all DFA's we examined. In other words, this algorithm converges in d iterations by computing the equivalence classes of each state, where d is the diameter of the input DFA

    Pre-Reduction Graph Products: Hardnesses of Properly Learning DFAs and Approximating EDP on DAGs

    Full text link
    The study of graph products is a major research topic and typically concerns the term f(G∗H)f(G*H), e.g., to show that f(G∗H)=f(G)f(H)f(G*H)=f(G)f(H). In this paper, we study graph products in a non-standard form f(R[G∗H]f(R[G*H] where RR is a "reduction", a transformation of any graph into an instance of an intended optimization problem. We resolve some open problems as applications. (1) A tight n1−ϵn^{1-\epsilon}-approximation hardness for the minimum consistent deterministic finite automaton (DFA) problem, where nn is the sample size. Due to Board and Pitt [Theoretical Computer Science 1992], this implies the hardness of properly learning DFAs assuming NP≠RPNP\neq RP (the weakest possible assumption). (2) A tight n1/2−ϵn^{1/2-\epsilon} hardness for the edge-disjoint paths (EDP) problem on directed acyclic graphs (DAGs), where nn denotes the number of vertices. (3) A tight hardness of packing vertex-disjoint kk-cycles for large kk. (4) An alternative (and perhaps simpler) proof for the hardness of properly learning DNF, CNF and intersection of halfspaces [Alekhnovich et al., FOCS 2004 and J. Comput.Syst.Sci. 2008]

    The Covering Problem

    Full text link
    An important endeavor in computer science is to understand the expressive power of logical formalisms over discrete structures, such as words. Naturally, "understanding" is not a mathematical notion. This investigation requires therefore a concrete objective to capture this understanding. In the literature, the standard choice for this objective is the membership problem, whose aim is to find a procedure deciding whether an input regular language can be defined in the logic under investigation. This approach was cemented as the right one by the seminal work of Sch\"utzenberger, McNaughton and Papert on first-order logic and has been in use since then. However, membership questions are hard: for several important fragments, researchers have failed in this endeavor despite decades of investigation. In view of recent results on one of the most famous open questions, namely the quantifier alternation hierarchy of first-order logic, an explanation may be that membership is too restrictive as a setting. These new results were indeed obtained by considering more general problems than membership, taking advantage of the increased flexibility of the enriched mathematical setting. This opens a promising research avenue and efforts have been devoted at identifying and solving such problems for natural fragments. Until now however, these problems have been ad hoc, most fragments relying on a specific one. A unique new problem replacing membership as the right one is still missing. The main contribution of this paper is a suitable candidate to play this role: the Covering Problem. We motivate this problem with 3 arguments. First, it admits an elementary set theoretic formulation, similar to membership. Second, we are able to reexplain or generalize all known results with this problem. Third, we develop a mathematical framework and a methodology tailored to the investigation of this problem

    Completeness for Identity-free Kleene Lattices

    Get PDF
    We provide a finite set of axioms for identity-free Kleene lattices, which we prove sound and complete for the equational theory of their relational models. Our proof builds on the completeness theorem for Kleene algebra, and on a novel automata construction that makes it possible to extract axiomatic proofs using a Kleene-like algorithm

    Dense urban elevation models from stereo images by an affine region merging approach

    Get PDF
    The main subject of this thesis is the computation of Dense Disparity Maps from a pair of satelite or aerial stereo images from an urban scene, taken from two different viewpoints. Several steps are needed to obtain the ?nal disparity map from the pair of images. We focus here on one of these steps: how to match the points in one image with the points in the other one. This matching process is closely related to the computation of the altitudes of the objects present in the scene. Indeed, the precision we can obtain in these altitude values is directly proportional to the precision in the matching process. This precision in the altitude is also inversely proportional to the distance between both viewpoints where the images are taken(baseline). The matching process is a widely studied ?eld in the Computer Vision Community and several methods and algorithms have been developed so far ([31, 27, 49]). Most of them consider a big base- line con?guration, which increases the performance in the altitude and also simpli?es the matching process. However, this assumption presents a major drawback with objects that are occluded in one image but appear in the other one. The bigger the baseline is, the more objects are occluded in one image and are not occluded in the other one. Recently, a different approach in which the images are taken with a very small baseline started to be analyzed ([19, 20]). This approach has the advantage of eliminating most of the ambiguities presented when one object occluded in one image is not occluded in the other one. Indeed, if we consider that we have a very small baseline, the occlusions presented in both images are almost the same. Now, this con?guration obviously decreases the precision in the ?nal altitude. In order to continue obtaining highly accurate altitude values, the precision in the matching process must be im- proved. The methods developed so far which consider the small baseline approach, compute altitude values with a high precision at some points, but leave the rest of them with no altitude values at all, generating a non-dense disparity map. Based on the fact that piecewise-a?ne models are reasonable for the elevation in urban areas, we propose a new method to interpolate and denoise those non-dense disparity maps. Under lambertian illumination hypothesis 1 , it is reasonable to assume that homogeneous regions in the graylevel image, correspond to the same a?ne elevation model. In other words, the borders between the piecewise a?ne elevation model are included to a large extent within contrasted graylevel borders. Hence, it is reasonable to look for an piecewise a?ne ?t to the elevation model where the borders between regions are taken from a graylevel segmenation of the image We present a region-merging algorithm that starts with an over-segmentation of the gray-level im- age. The disparity values at each region are approximated by an a?ne model, and a meaningfulness measure of the ?t is assigned to each of them. Using this meaningfulness as a merging order, the method iterates until no new merge is possible, according to a merging criterion which is also based on the meaningfulness of each pair of neighboring regions. In the last step, the algorithm performs a validation of the ?nal regions using again the meaningfulness of the ?t. The regions validated in this last step are those for which the a?ne model is a good approximation. The region-merging algorithm presented in this work can be seen as an attempt to incorporate a semantical meaning to real scenes: we have developed a validation method to determine whether the data within a region is well approximated by an a?ne model or not. Hence, we could analyze more complex models, de?ning a suitable validation criterion for each of them. In this way, we can search for the model that best explains a given data set in terms of its meaningfulness

    Regular Separability of Parikh Automata

    Get PDF
    We investigate a subclass of languages recognized by vector addition systems, namely languages of nondeterministic Parikh automata. While the regularity problem (is the language of a given automaton regular?) is undecidable for this model, we surprisingly show decidability of the regular separability problem: given two Parikh automata, is there a regular language that contains one of them and is disjoint from the other? We supplement this result by proving undecidability of the same problem already for languages of visibly one counter automata

    Quantum automata, braid group and link polynomials

    Full text link
    The spin--network quantum simulator model, which essentially encodes the (quantum deformed) SU(2) Racah--Wigner tensor algebra, is particularly suitable to address problems arising in low dimensional topology and group theory. In this combinatorial framework we implement families of finite--states and discrete--time quantum automata capable of accepting the language generated by the braid group, and whose transition amplitudes are colored Jones polynomials. The automaton calculation of the polynomial of (the plat closure of) a link L on 2N strands at any fixed root of unity is shown to be bounded from above by a linear function of the number of crossings of the link, on the one hand, and polynomially bounded in terms of the braid index 2N, on the other. The growth rate of the time complexity function in terms of the integer k appearing in the root of unity q can be estimated to be (polynomially) bounded by resorting to the field theoretical background given by the Chern-Simons theory.Comment: Latex, 36 pages, 11 figure

    Detection of traffic congestion and incidents from GPS trace analysis

    Get PDF
    This paper presents an expert system for detecting traffic congestion and incidents from real-time GPS data collected from GPS trackers or drivers’ smartphones. First, GPS traces are pre-processed and placed in the road map. Then, the system assigns to each road segment of the map a traffic state based on the speeds of the vehicles. Finally, it sends to the users traffic alerts based on a spatiotemporal analysis of the classified segments. Each traffic alert contains the affected area, a traffic state (e.g., incident, slowed traffic, blocked traffic), and the estimated velocity of vehicles in the area. The proposed system is intended to be a valuable support tool in traffic management for municipalities and citizens. The information produced by the system can be successfully employed to adopt actions for improving the city mobility, e.g., regulate vehicular traffic, or can be exploited by the users, who may spontaneously decide to modify their path in order to avoid the traffic jam. The elaboration performed by the expert system is independent of the context (urban o non-urban) and may be directly employed in several city road networks with almost no change of the system parameters, and without the need for a learning process or historical data. The experimental analysis was performed using a combination of simulated GPS data and real GPS data from the city of Pisa. The results on incidents show a detection rate of 91.6%, and an average detection time lower than 7 min. Regarding congestion, we show how the system is able to recognize different levels of congestion depending on different road use
    • …
    corecore