451 research outputs found

    Multi-Head Finite Automata: Characterizations, Concepts and Open Problems

    Full text link
    Multi-head finite automata were introduced in (Rabin, 1964) and (Rosenberg, 1966). Since that time, a vast literature on computational and descriptional complexity issues on multi-head finite automata documenting the importance of these devices has been developed. Although multi-head finite automata are a simple concept, their computational behavior can be already very complex and leads to undecidable or even non-semi-decidable problems on these devices such as, for example, emptiness, finiteness, universality, equivalence, etc. These strong negative results trigger the study of subclasses and alternative characterizations of multi-head finite automata for a better understanding of the nature of non-recursive trade-offs and, thus, the borderline between decidable and undecidable problems. In the present paper, we tour a fragment of this literature

    Tighter Connections between Derandomization and Circuit Lower Bounds

    Get PDF
    We tighten the connections between circuit lower bounds and derandomization for each of the following three types of derandomization: - general derandomization of promiseBPP (connected to Boolean circuits), - derandomization of Polynomial Identity Testing (PIT) over fixed finite fields (connected to arithmetic circuit lower bounds over the same field), and - derandomization of PIT over the integers (connected to arithmetic circuit lower bounds over the integers). We show how to make these connections uniform equivalences, although at the expense of using somewhat less common versions of complexity classes and for a less studied notion of inclusion. Our main results are as follows: 1. We give the first proof that a non-trivial (nondeterministic subexponential-time) algorithm for PIT over a fixed finite field yields arithmetic circuit lower bounds. 2. We get a similar result for the case of PIT over the integers, strengthening a result of Jansen and Santhanam [JS12] (by removing the need for advice). 3. We derive a Boolean circuit lower bound for NEXP intersect coNEXP from the assumption of sufficiently strong non-deterministic derandomization of promiseBPP (without advice), as well as from the assumed existence of an NP-computable non-empty property of Boolean functions useful for proving superpolynomial circuit lower bounds (in the sense of natural proofs of [RR97]); this strengthens the related results of [IKW02]. 4. Finally, we turn all of these implications into equivalences for appropriately defined promise classes and for a notion of robust inclusion/separation (inspired by [FS11]) that lies between the classical "almost everywhere" and "infinitely often" notions

    Dynamically optimal treatment allocation using Reinforcement Learning

    Full text link
    Devising guidance on how to assign individuals to treatment is an important goal in empirical research. In practice, individuals often arrive sequentially, and the planner faces various constraints such as limited budget/capacity, or borrowing constraints, or the need to place people in a queue. For instance, a governmental body may receive a budget outlay at the beginning of a year, and it may need to decide how best to allocate resources within the year to individuals who arrive sequentially. In this and other examples involving inter-temporal trade-offs, previous work on devising optimal policy rules in a static context is either not applicable, or sub-optimal. Here we show how one can use offline observational data to estimate an optimal policy rule that maximizes expected welfare in this dynamic context. We allow the class of policy rules to be restricted for legal, ethical or incentive compatibility reasons. The problem is equivalent to one of optimal control under a constrained policy class, and we exploit recent developments in Reinforcement Learning (RL) to propose an algorithm to solve this. The algorithm is easily implementable with speedups achieved through multiple RL agents learning in parallel processes. We also characterize the statistical regret from using our estimated policy rule by casting the evolution of the value function under each policy in a Partial Differential Equation (PDE) form and using the theory of viscosity solutions to PDEs. We find that the policy regret decays at a n−1/2n^{-1/2} rate in most examples; this is the same rate as in the static case.Comment: 67 page

    Risk-Aware Stability of Discrete-Time Systems

    Full text link
    We develop a generalized stability framework for stochastic discrete-time systems, where the generality pertains to the ways in which the distribution of the state energy can be characterized. We use tools from finance and operations research called risk functionals (i.e., risk measures) to facilitate diverse distributional characterizations. In contrast, classical stochastic stability notions characterize the state energy on average or in probability, which can obscure the variability of stochastic system behavior. After drawing connections between various risk-aware stability concepts for nonlinear systems, we specialize to linear systems and derive sufficient conditions for the satisfaction of some risk-aware stability properties. These results pertain to real-valued coherent risk functionals and a mean-conditional-variance functional. The results reveal novel noise-to-state stability properties, which assess disturbances in ways that reflect the chosen measure of risk. We illustrate the theory through examples about robustness, parameter choices, and state-feedback controllers

    Relaxing Fundamental Assumptions in Iterative Learning Control

    Full text link
    Iterative learning control (ILC) is perhaps best decribed as an open loop feedforward control technique where the feedforward signal is learned through repetition of a single task. As the name suggests, given a dynamic system operating on a finite time horizon with the same desired trajectory, ILC aims to iteratively construct the inverse image (or its approximation) of the desired trajectory to improve transient tracking. In the literature, ILC is often interpreted as feedback control in the iteration domain due to the fact that learning controllers use information from past trials to drive the tracking error towards zero. However, despite the significant body of literature and powerful features, ILC is yet to reach widespread adoption by the control community, due to several assumptions that restrict its generality when compared to feedback control. In this dissertation, we relax some of these assumptions, mainly the fundamental invariance assumption, and move from the idea of learning through repetition to two dimensional systems, specifically repetitive processes, that appear in the modeling of engineering applications such as additive manufacturing, and sketch out future research directions for increased practicality: We develop an L1 adaptive feedback control based ILC architecture for increased robustness, fast convergence, and high performance under time varying uncertainties and disturbances. Simulation studies of the behavior of this combined L1-ILC scheme under iteration varying uncertainties lead us to the robust stability analysis of iteration varying systems, where we show that these systems are guaranteed to be stable when the ILC update laws are designed to be robust, which can be done using existing methods from the literature. As a next step to the signal space approach adopted in the analysis of iteration varying systems, we shift the focus of our work to repetitive processes, and show that the exponential stability of a nonlinear repetitive system is equivalent to that of its linearization, and consequently uniform stability of the corresponding state space matrix.PhDElectrical Engineering: SystemsUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/133232/1/altin_1.pd

    Beyond the Knowledge-Based Theory of the Geographic Cluster

    Get PDF
    The knowledge-based theory of the geographic cluster represents a major attempt to re-conceptualize clusters, in essence arguing that the localization of firms in similar and related industries stimulates learning and innovation, giving a competitive advantage to clustered firms. This paper critically examines the knowledge-based theory the cluster, arguing that it has greatly overstated the advantages of co-location to firms and misidentified the mechanisms through which learning occurs in clusters. In particular, the theory is criticized on three points: the flexible, under-specified way that it defines its object of study; the focus on firms as an explanatory variable instead of more fundamental processes of resource accumulation; and the functionalist mode of theory that employs as an explanation. Ways to address of each of these issues are discussed. In a final section I suggest that the rather static notions of learning put forward in the knowledge-based theory of the cluster be replaced by a developmental theory of regional dynamics that focuses on both learning and structural transformation.geographic cluster, localization, relatedness, knowledge-based theory
    • 

    corecore