57 research outputs found

    Residual Nominal Automata

    Get PDF
    Nominal automata are models for accepting languages over infinite alphabets. In this paper we refine the hierarchy of nondeterministic nominal automata, by developing the theory of residual nominal automata. In particular, we show that they admit canonical minimal representatives, and that the universality problem becomes decidable. We also study exact learning of these automata, and settle questions that were left open about their learnability via observations

    The price of a perfect system: learnability and the distribution of errors in the speech of children learning English as a first language

    Get PDF
    This study reports on a strictly-cognitive and symptomatic approach to the treatment of phonological disorders, by an effect which can also be reproduced in most normally- developing children. To explain how this works, it is necessary to address certain asymmetries and singularities in the distribution of children's speech errors over the whole range of development. Particular words occasion particular errors. In early phonology there is 'fronting' with Coronal displacing Dorsal, and harmonies where Coronal is lost. In the middle of phonological acquisition, the harmonic pattern changes with coronal harmony coming to prevail over other forms. As well as these asymmetries, there is also the case of harmonic or migratory errors involving the property of affrication, but not the affricate as a whole, i.e. ignoring the property of voicing. Many of these asymmetries and singularities and the harmony or movement of affrication are described here for the first time. They are all difficult to explain in current theoretical models, especially in 'bottom-up' models. On the basis of the 'top-down' notion of 'parameters' from recent work in phonology, I shall assume that: A) finite learnability has to be ensured; B) there can be no privileged information about the learnability target; and C) phonological theory and the study of speech development (normal and otherwise) have an object in common. I shall propose: A) a Parameter Setting Function, as part of the human genome, possibly a defining part; B) Phonological Parapraxis', as a way of characterising the generalisations here about incompetent phonology by the general mechanisms of floating' and 'non-association'; C) a Stage (_n-1) as a necessary construct in the theory of acquisition, typically not reached before 8;6; D) a' Representability Inspection' relating normal competence to Chomsky's Articulatory/ Perceptual interface', sensitive to a relation between featural properties such as roundness or labiality and prosodic properties such as the foot and syllable; E) a syndrome. Specific Speech and Language Impairment, SSLI, extending the notion of Specific Language Impairment, SLI.I shall hypothesise that: A) segmental and suprasegmental representations interact; B) the phonological learnability space is uniform and consistent; C) it is the very minimality of the learnability system which makes it vulnerable to SSLI. This: A) side-steps the implausible inference that development proceeds by the loss of 'processes'; B) accounts for at least some of the asymmetries noted above; C) lets parameters set' a degree of abstract exponence; D) makes it possible to abolish 'processes' such as fronting, lisping, consonant harmony, in favour of successive degrees of imprecision in the parameterisation; E) provides a conceptual mechanism for the cognitive and symptomatic therapy, mentioned above: the therapy effects an increase in the set of phonological structures which are 'representable' by the child

    Substitutional quantification and set theory

    Full text link
    Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/43184/1/10992_2004_Article_BF00258434.pd

    Sentence alignment of Hungarian-English parallel corpora using a hybrid algorithm

    Get PDF
    We present an efficient hybrid method for aligning sentences with their translations in a parallel bilingual corpus. The new algorithm is composed of a length-based and anchor matching method that uses Named Entity recognition. This algorithm combines the speed of length-based models with the accuracy of anchor finding methods. The accuracy of finding cognates for Hungarian-English language pair is extremely low, hence we thought of using a novel approach that includes Named Entity recognition. Due to the well selected anchors it was found to outperform the best two sentence alignment algorithms so far published for the Hungarian-English language pair

    Topological Aspects of Epistemology and Metaphysics

    Get PDF
    The aim of this paper is to show that (elementary) topology may be useful for dealing with problems of epistemology and metaphysics. More precisely, I want to show that the introduction of topological structures may elucidate the role of the spatial structures (in a broad sense) that underly logic and cognition. In some detail I’ll deal with “Cassirer’s problem” that may be characterized as an early forrunner of Goodman’s “grue-bleen” problem. On a larger scale, topology turns out to be useful in elaborating the approach of conceptual spaces that in the last twenty years or so has found quite a few applications in cognitive science, psychology, and linguistics. In particular, topology may help distinguish “natural” from “not-so-natural” concepts. This classical problem that up to now has withstood all efforts to solve (or dissolve) it by purely logical methods. Finally, in order to show that a topological perspective may also offer a fresh look on classical metaphysical problems, it is shown that Leibniz’s famous principle of the identity of indiscernibles is closely related to some well-known topological separation axioms. More precisely, the topological perspective gives rise in a natural way to some novel variations of Leibniz’s principle

    Exponential separations between classical and quantum learners

    Full text link
    Despite significant effort, the quantum machine learning community has only demonstrated quantum learning advantages for artificial cryptography-inspired datasets when dealing with classical data. In this paper we address the challenge of finding learning problems where quantum learning algorithms can achieve a provable exponential speedup over classical learning algorithms. We reflect on computational learning theory concepts related to this question and discuss how subtle differences in definitions can result in significantly different requirements and tasks for the learner to meet and solve. We examine existing learning problems with provable quantum speedups and find that they largely rely on the classical hardness of evaluating the function that generates the data, rather than identifying it. To address this, we present two new learning separations where the classical difficulty primarily lies in identifying the function generating the data. Furthermore, we explore computational hardness assumptions that can be leveraged to prove quantum speedups in scenarios where data is quantum-generated, which implies likely quantum advantages in a plethora of more natural settings (e.g., in condensed matter and high energy physics). We also discuss the limitations of the classical shadow paradigm in the context of learning separations, and how physically-motivated settings such as characterizing phases of matter and Hamiltonian learning fit in the computational learning framework.Comment: this article supersedes arXiv:2208.0633

    Testing product states, quantum Merlin-Arthur games and tensor optimisation

    Full text link
    We give a test that can distinguish efficiently between product states of n quantum systems and states which are far from product. If applied to a state psi whose maximum overlap with a product state is 1-epsilon, the test passes with probability 1-Theta(epsilon), regardless of n or the local dimensions of the individual systems. The test uses two copies of psi. We prove correctness of this test as a special case of a more general result regarding stability of maximum output purity of the depolarising channel. A key application of the test is to quantum Merlin-Arthur games with multiple Merlins, where we obtain several structural results that had been previously conjectured, including the fact that efficient soundness amplification is possible and that two Merlins can simulate many Merlins: QMA(k)=QMA(2) for k>=2. Building on a previous result of Aaronson et al, this implies that there is an efficient quantum algorithm to verify 3-SAT with constant soundness, given two unentangled proofs of O(sqrt(n) polylog(n)) qubits. We also show how QMA(2) with log-sized proofs is equivalent to a large number of problems, some related to quantum information (such as testing separability of mixed states) as well as problems without any apparent connection to quantum mechanics (such as computing injective tensor norms of 3-index tensors). As a consequence, we obtain many hardness-of-approximation results, as well as potential algorithmic applications of methods for approximating QMA(2) acceptance probabilities. Finally, our test can also be used to construct an efficient test for determining whether a unitary operator is a tensor product, which is a generalisation of classical linearity testing.Comment: 44 pages, 1 figure, 7 appendices; v6: added references, rearranged sections, added discussion of connections to classical CS. Final version to appear in J of the AC

    A Theory of Emergent In-Context Learning as Implicit Structure Induction

    Full text link
    Scaling large language models (LLMs) leads to an emergent capacity to learn in-context from example demonstrations. Despite progress, theoretical understanding of this phenomenon remains limited. We argue that in-context learning relies on recombination of compositional operations found in natural language data. We derive an information-theoretic bound showing how in-context learning abilities arise from generic next-token prediction when the pretraining distribution has sufficient amounts of compositional structure, under linguistically motivated assumptions. A second bound provides a theoretical justification for the empirical success of prompting LLMs to output intermediate steps towards an answer. To validate theoretical predictions, we introduce a controlled setup for inducing in-context learning; unlike previous approaches, it accounts for the compositional nature of language. Trained transformers can perform in-context learning for a range of tasks, in a manner consistent with the theoretical results. Mirroring real-world LLMs in a miniature setup, in-context learning emerges when scaling parameters and data, and models perform better when prompted to output intermediate steps. Probing shows that in-context learning is supported by a representation of the input's compositional structure. Taken together, these results provide a step towards theoretical understanding of emergent behavior in large language models
    • …
    corecore