34,938 research outputs found
Exterminator: Automatically Correcting Memory Errors with High Probability
Programs written in C and C++ are susceptible to memory errors, including buffer overflows and dangling pointers. These errors, which can lead to crashes, erroneous execution, and security vulnerabilities, are notoriously costly to repair. Tracking down their location in the source code is difficult, even when the full memory state of the program is available. Once the errors are finally found, fixing them remains challenging: even for critical security-sensitive bugs, the average time between initial reports and the issuance of a patch is nearly one month. We present Exterminator, a system that automatically corrects heap-based memory errors without programmer intervention. Exterminator exploits randomization to pinpoint errors with high precision. From this information, Exterminator derives runtime patches that fix these errors both in current and subsequent executions. In addition, Exterminator enables collaborative bug correction by merging patches generated by multiple users. We present analytical and empirical results that demonstrate Exterminator’s effectiveness at detecting and correcting both injected and real faults
New Technique to Enhance the Performance of Spoken Dialogue Systems by Means of Implicit Recovery of ASR Errors
This paper proposes a new technique to implicitly correct some ASR
errors made by spoken dialogue systems, which is implemented at two levels:
statistical and linguistic. The goal of the former level is to employ for the correction
knowledge extracted from the analysis of a training corpus comprised of
utterances and their corresponding ASR results. The outcome of the analysis is
a set of syntactic-semantic models and a set of lexical models, which are optimally
selected during the correction. The goal of the correction at the linguistic
level is to repair errors not detected during the statistical level which affects the
semantics of the sentences. Experiments carried out with a previouslydeveloped
spoken dialogue system for the fast food domain indicate that the
technique allows enhancing word accuracy, spoken language understanding and
task completion by 8.5%, 16.54% and 44.17% absolute, respectively.Ministerio de Ciencia y Tecnología TIN2007-64718 HAD
Context-sensitive Spelling Correction Using Google Web 1T 5-Gram Information
In computing, spell checking is the process of detecting and sometimes
providing spelling suggestions for incorrectly spelled words in a text.
Basically, a spell checker is a computer program that uses a dictionary of
words to perform spell checking. The bigger the dictionary is, the higher is
the error detection rate. The fact that spell checkers are based on regular
dictionaries, they suffer from data sparseness problem as they cannot capture
large vocabulary of words including proper names, domain-specific terms,
technical jargons, special acronyms, and terminologies. As a result, they
exhibit low error detection rate and often fail to catch major errors in the
text. This paper proposes a new context-sensitive spelling correction method
for detecting and correcting non-word and real-word errors in digital text
documents. The approach hinges around data statistics from Google Web 1T 5-gram
data set which consists of a big volume of n-gram word sequences, extracted
from the World Wide Web. Fundamentally, the proposed method comprises an error
detector that detects misspellings, a candidate spellings generator based on a
character 2-gram model that generates correction suggestions, and an error
corrector that performs contextual error correction. Experiments conducted on a
set of text documents from different domains and containing misspellings,
showed an outstanding spelling error correction rate and a drastic reduction of
both non-word and real-word errors. In a further study, the proposed algorithm
is to be parallelized so as to lower the computational cost of the error
detection and correction processes.Comment: LACSC - Lebanese Association for Computational Sciences -
http://www.lacsc.or
Verifying the Steane code with Quantomatic
In this paper we give a partially mechanized proof of the correctness of
Steane's 7-qubit error correcting code, using the tool Quantomatic. To the best
of our knowledge, this represents the largest and most complicated verification
task yet carried out using Quantomatic.Comment: In Proceedings QPL 2013, arXiv:1412.791
Enhanced Integrated Scoring for Cleaning Dirty Texts
An increasing number of approaches for ontology engineering from text are
gearing towards the use of online sources such as company intranet and the
World Wide Web. Despite such rise, not much work can be found in aspects of
preprocessing and cleaning dirty texts from online sources. This paper presents
an enhancement of an Integrated Scoring for Spelling error correction,
Abbreviation expansion and Case restoration (ISSAC). ISSAC is implemented as
part of a text preprocessing phase in an ontology engineering system. New
evaluations performed on the enhanced ISSAC using 700 chat records reveal an
improved accuracy of 98% as compared to 96.5% and 71% based on the use of only
basic ISSAC and of Aspell, respectively.Comment: More information is available at
http://explorer.csse.uwa.edu.au/reference
Quantum Computing with Very Noisy Devices
In theory, quantum computers can efficiently simulate quantum physics, factor
large numbers and estimate integrals, thus solving otherwise intractable
computational problems. In practice, quantum computers must operate with noisy
devices called ``gates'' that tend to destroy the fragile quantum states needed
for computation. The goal of fault-tolerant quantum computing is to compute
accurately even when gates have a high probability of error each time they are
used. Here we give evidence that accurate quantum computing is possible with
error probabilities above 3% per gate, which is significantly higher than what
was previously thought possible. However, the resources required for computing
at such high error probabilities are excessive. Fortunately, they decrease
rapidly with decreasing error probabilities. If we had quantum resources
comparable to the considerable resources available in today's digital
computers, we could implement non-trivial quantum computations at error
probabilities as high as 1% per gate.Comment: 47 page
- …