3,868 research outputs found
Recommended from our members
Optimal Short-Circuit Resilient Formulas
We consider fault-tolerant boolean formulas in which the output of a faulty gate is short-circuited to one of the gate\u27s inputs. A recent result by Kalai et al. [FOCS 2012] converts any boolean formula into a resilient formula of polynomial size that works correctly if less than a fraction 1/6 of the gates (on every input-to-output path) are faulty. We improve the result of Kalai et al., and show how to efficiently fortify any boolean formula against a fraction 1/5 of short-circuit gates per path, with only a polynomial blowup in size. We additionally show that it is impossible to obtain formulas with higher resilience and sub-exponential growth in size.
Towards our results, we consider interactive coding schemes when noiseless feedback is present; these produce resilient boolean formulas via a Karchmer-Wigderson relation. We develop a coding scheme that resists up to a fraction 1/5 of corrupted transmissions in each direction of the interactive channel. We further show that such a level of noise is maximal for coding schemes with sub-exponential blowup in communication. Our coding scheme takes a surprising inspiration from Blockchain technology
An Atypical Survey of Typical-Case Heuristic Algorithms
Heuristic approaches often do so well that they seem to pretty much always
give the right answer. How close can heuristic algorithms get to always giving
the right answer, without inducing seismic complexity-theoretic consequences?
This article first discusses how a series of results by Berman, Buhrman,
Hartmanis, Homer, Longpr\'{e}, Ogiwara, Sch\"{o}ening, and Watanabe, from the
early 1970s through the early 1990s, explicitly or implicitly limited how well
heuristic algorithms can do on NP-hard problems. In particular, many desirable
levels of heuristic success cannot be obtained unless severe, highly unlikely
complexity class collapses occur. Second, we survey work initiated by Goldreich
and Wigderson, who showed how under plausible assumptions deterministic
heuristics for randomized computation can achieve a very high frequency of
correctness. Finally, we consider formal ways in which theory can help explain
the effectiveness of heuristics that solve NP-hard problems in practice.Comment: This article is currently scheduled to appear in the December 2012
issue of SIGACT New
Evaluation of A Resilience Embedded System Using Probabilistic Model-Checking
If a Micro Processor Unit (MPU) receives an external electric signal as
noise, the system function will freeze or malfunction easily. A new resilience
strategy is implemented in order to reset the MPU automatically and stop the
MPU from freezing or malfunctioning. The technique is useful for embedded
systems which work in non-human environments. However, evaluating resilience
strategies is difficult because their effectiveness depends on numerous,
complex, interacting factors.
In this paper, we use probabilistic model checking to evaluate the embedded
systems installed with the above mentioned new resilience strategy. Qualitative
evaluations are implemented with 6 PCTL formulas, and quantitative evaluations
use two kinds of evaluation. One is system failure reduction, and the other is
ADT (Average Down Time), the industry standard. Our work demonstrates the
benefits brought by the resilience strategy. Experimental results indicate that
our evaluation is cost-effective and reliable.Comment: In Proceedings ESSS 2014, arXiv:1405.055
Power efficient resilient microarchitectures for PVT variability mitigation
Nowadays, the high power density and the process, voltage, and temperature variations became the most critical issues that limit the performance of the digital integrated circuits because of the continuous scaling of the fabrication technology. Dynamic voltage and frequency scaling technique is used to reduce the power consumption while different time relaxation techniques and error recovery microarchitectures are used to tolerate the process, voltage, and temperature variations. These techniques reduce the throughput by scaling down the frequency or flushing and restarting the errant pipeline. This thesis presents a novel resilient microarchitecture which is called ERSUT-based resilient microarchitecture to tolerate the induced delays generated by the voltage scaling or the process, voltage, and temperature variations. The resilient microarchitecture detects and recovers the induced errors without flushing the pipeline and without scaling down the operating frequency. An ERSUT-based resilient 16 × 16 bit MAC unit, implemented using Global Foundries 65 nm technology and ARM standard cells library, is introduced as a case study with 18.26% area overhead and up to 1.5x speedup. At the typical conditions, the maximum frequency of the conventional MAC unit is about 375 MHz while the resilient MAC unit operates correctly at a frequency up to 565 MHz. In case of variations, the resilient MAC unit tolerates induced delays up to 50% of the clock period while keeping its throughput equal to the conventional MAC unit’s maximum throughput. At 375 MHz, the resilient MAC unit is able to scale down the supply voltage from 1.2 V to 1.0 V saving about 29% of the power consumed by the conventional MAC unit. A double-edge-triggered microarchitecture is also introduced to reduce the power consumption extremely by reducing the frequency of the clock tree to the half while preserving the same maximum throughput. This microarchitecture is applied to different ISCAS’89 benchmark circuits in addition to the 16x16 bit MAC unit and the average power reduction of all these circuits is 63.58% while the average area overhead is 31.02%. All these circuits are designed using Global Foundries 65nm technology and ARM standard cells library. Towards the full automation of the ERSUT-based resilient microarchitecture, an ERSUT-based algorithm is introduced in C++ to accelerate the design process of the ERSUT-based microarchitecture. The developed algorithm reduces the design-time efforts dramatically and allows the ERSUT-based microarchitecture to be adopted by larger industrial designs. Depending on the ERSUT-based algorithm, a validation study about applying the ERSUT-based microarchitecture on the MAC unit and different ISCAS’89 benchmark circuits with different complexity weights is introduced. This study shows that 72% of these circuits tolerates more than 14% of their clock periods and 54.5% of these circuits tolerates more than 20% while 27% of these circuits tolerates more than 30%. Consequently, the validation study proves that the ERSUT-based resilient microarchitecture is a valid applicable solution for different circuits with different complexity weights
Recommended from our members
Complexity Theory
Computational Complexity Theory is the mathematical study of the intrinsic power and limitations of computational resources like time, space, or randomness. The current workshop focused on recent developments in various sub-areas including arithmetic complexity, Boolean complexity, communication complexity, cryptography, probabilistic proof systems, pseudorandomness and randomness extraction. Many of the developments are related to diverse mathematical fields such as algebraic geometry, combinatorial number theory, probability theory, representation theory, and the theory of error-correcting codes
Фізика: лабораторний практикум
Present book contains theory ,description of experimental equipment ,procedure and analysis recommendations to the physics laboratory experiments for foreign students studying in Ukraine
Composite-pulse magnetometry with a solid-state quantum sensor
The sensitivity of quantum magnetometers is challenged by control errors and,
especially in the solid-state, by their short coherence times. Refocusing
techniques can overcome these limitations and improve the sensitivity to
periodic fields, but they come at the cost of reduced bandwidth and cannot be
applied to sense static (DC) or aperiodic fields. Here we experimentally
demonstrate that continuous driving of the sensor spin by a composite pulse
known as rotary-echo (RE) yields a flexible magnetometry scheme, mitigating
both driving power imperfections and decoherence. A suitable choice of RE
parameters compensates for different scenarios of noise strength and origin.
The method can be applied to nanoscale sensing in variable environments or to
realize noise spectroscopy. In a room-temperature implementation based on a
single electronic spin in diamond, composite-pulse magnetometry provides a
tunable trade-off between sensitivities in the microT/sqrt(Hz) range,
comparable to those obtained with Ramsey spectroscopy, and coherence times
approaching T1
- …