1,311 research outputs found
Advancing Hardware Security Using Polymorphic and Stochastic Spin-Hall Effect Devices
Protecting intellectual property (IP) in electronic circuits has become a
serious challenge in recent years. Logic locking/encryption and layout
camouflaging are two prominent techniques for IP protection. Most existing
approaches, however, particularly those focused on CMOS integration, incur
excessive design overheads resulting from their need for additional circuit
structures or device-level modifications. This work leverages the innate
polymorphism of an emerging spin-based device, called the giant spin-Hall
effect (GSHE) switch, to simultaneously enable locking and camouflaging within
a single instance. Using the GSHE switch, we propose a powerful primitive that
enables cloaking all the 16 Boolean functions possible for two inputs. We
conduct a comprehensive study using state-of-the-art Boolean satisfiability
(SAT) attacks to demonstrate the superior resilience of the proposed primitive
in comparison to several others in the literature. While we tailor the
primitive for deterministic computation, it can readily support stochastic
computation; we argue that stochastic behavior can break most, if not all,
existing SAT attacks. Finally, we discuss the resilience of the primitive
against various side-channel attacks as well as invasive monitoring at runtime,
which are arguably even more concerning threats than SAT attacks.Comment: Published in Proc. Design, Automation and Test in Europe (DATE) 201
X-Rel: Energy-Efficient and Low-Overhead Approximate Reliability Framework for Error-Tolerant Applications Deployed in Critical Systems
Triple Modular Redundancy (TMR) is one of the most common techniques in
fault-tolerant systems, in which the output is determined by a majority voter.
However, the design diversity of replicated modules and/or soft errors that are
more likely to happen in the nanoscale era may affect the majority voting
scheme. Besides, the significant overheads of the TMR scheme may limit its
usage in energy consumption and area-constrained critical systems. However, for
most inherently error-resilient applications such as image processing and
vision deployed in critical systems (like autonomous vehicles and robotics),
achieving a given level of reliability has more priority than precise results.
Therefore, these applications can benefit from the approximate computing
paradigm to achieve higher energy efficiency and a lower area. This paper
proposes an energy-efficient approximate reliability (X-Rel) framework to
overcome the aforementioned challenges of the TMR systems and get the full
potential of approximate computing without sacrificing the desired reliability
constraint and output quality. The X-Rel framework relies on relaxing the
precision of the voter based on a systematical error bounding method that
leverages user-defined quality and reliability constraints. Afterward, the size
of the achieved voter is used to approximate the TMR modules such that the
overall area and energy consumption are minimized. The effectiveness of
employing the proposed X-Rel technique in a TMR structure, for different
quality constraints as well as with various reliability bounds are evaluated in
a 15-nm FinFET technology. The results of the X-Rel voter show delay, area, and
energy consumption reductions of up to 86%, 87%, and 98%, respectively, when
compared to those of the state-of-the-art approximate TMR voters.Comment: This paper has been published in IEEE Transactions on Very Large
Scale Integration (VLSI) System
Design Disjunction for Resilient Reconfigurable Hardware
Contemporary reconfigurable hardware devices have the capability to achieve high performance, power efficiency, and adaptability required to meet a wide range of design goals. With scaling challenges facing current complementary metal oxide semiconductor (CMOS), new concepts and methodologies supporting efficient adaptation to handle reliability issues are becoming increasingly prominent. Reconfigurable hardware and their ability to realize self-organization features are expected to play a key role in designing future dependable hardware architectures. However, the exponential increase in density and complexity of current commercial SRAM-based field-programmable gate arrays (FPGAs) has escalated the overhead associated with dynamic runtime design adaptation. Traditionally, static modular redundancy techniques are considered to surmount this limitation; however, they can incur substantial overheads in both area and power requirements. To achieve a better trade-off among performance, area, power, and reliability, this research proposes design-time approaches that enable fine selection of redundancy level based on target reliability goals and autonomous adaptation to runtime demands. To achieve this goal, three studies were conducted: First, a graph and set theoretic approach, named Hypergraph-Cover Diversity (HCD), is introduced as a preemptive design technique to shift the dominant costs of resiliency to design-time. In particular, union-free hypergraphs are exploited to partition the reconfigurable resources pool into highly separable subsets of resources, each of which can be utilized by the same synthesized application netlist. The diverse implementations provide reconfiguration-based resilience throughout the system lifetime while avoiding the significant overheads associated with runtime placement and routing phases. Evaluation on a Motion-JPEG image compression core using a Xilinx 7-series-based FPGA hardware platform has demonstrated the potential of the proposed FT method to achieve 37.5% area saving and up to 66% reduction in power consumption compared to the frequently-used TMR scheme while providing superior fault tolerance. Second, Design Disjunction based on non-adaptive group testing is developed to realize a low-overhead fault tolerant system capable of handling self-testing and self-recovery using runtime partial reconfiguration. Reconfiguration is guided by resource grouping procedures which employ non-linear measurements given by the constructive property of f-disjunctness to extend runtime resilience to a large fault space and realize a favorable range of tradeoffs. Disjunct designs are created using the mosaic convergence algorithm developed such that at least one configuration in the library evades any occurrence of up to d resource faults, where d is lower-bounded by f. Experimental results for a set of MCNC and ISCAS benchmarks have demonstrated f-diagnosability at the individual slice level with average isolation resolution of 96.4% (94.4%) for f=1 (f=2) while incurring an average critical path delay impact of only 1.49% and area cost roughly comparable to conventional 2-MR approaches. Finally, the proposed Design Disjunction method is evaluated as a design-time method to improve timing yield in the presence of large random within-die (WID) process variations for application with a moderately high production capacity
High Performance and Optimal Configuration of Accurate Heterogeneous Block-Based Approximate Adder
Approximate computing is an emerging paradigm to improve power and
performance efficiency for error-resilient application. Recent approximate
adders have significantly extended the design space of accuracy-power
configurable approximate adders, and find optimal designs by exploring the
design space. In this paper, a new energy-efficient heterogeneous block-based
approximate adder (HBBA) is proposed; which is a generic/configurable model
that can be transformed to a particular adder by defining some configurations.
An HBBA, in general, is composed of heterogeneous sub-adders, where each
sub-adder can have a different configuration. A set of configurations of all
the sub-adders in an HBBA defines its configuration. The block-based adders are
approximated through inexact logic configuration and truncated carry chains.
HBBA increases design space providing additional design points that fall on the
Pareto-front and offer better power-accuracy trade-off compared to other
configurations. Furthermore, to avoid Mont-Carlo simulations, we propose an
analytical modelling technique to evaluate the probability of error and
Probability Mass Function (PMF) of error value. Moreover, the estimation method
estimates delay, area and power of heterogeneous block-based approximate
adders. Thus, based on the analytical model and estimation method, the optimal
configuration under a given error constraint can be selected from the whole
design space of the proposed adder model by exhaustive search. The simulation
results show that our HBBA provides improved accuracy in terms of error metrics
compared to some state-of-the-art approximate adders. HBBA with 32 bits length
serves about 15% reduction in area and up to 17% reduction in energy compared
to state-of-the-art approximate adders.Comment: Submitted to the IEEE-TCAD journal, 16 pages, 16 figure
Engineering Photon Sources for Practical Quantum Information Processing:If you liked it then you should have put a ring on it
Integrated quantum photonics offers a promising route to the realisation of universal fault-tolerant quantum computers. Much progress has been made on the theoretical aspects of a future quantum information processor, reducing both error thresholds and circuit complexity. Currently, engineering efforts are focused on integrating the most valuable technologies for a photonic quantum computer; pure single-photon sources, low-loss phase shifters and passivecircuit components, as well as efficient single-photon detectors and corresponding electronics.Here, we present efforts to target the former under the constraints imposed by the latter. We engineer the spectral correlations of photons produced by a heralded single-photon source, such that they produce photons in pure quantum states (99.1±0.1 % purity), and enable additional optimisation using temporal shaping of the pump field. Our source also has a high intrinsicheralding efficiency (94.0 ± 2.9 %) and produces photon pairs at a rate (4.4 ± 0.1 MHz mW−2) which is an order of magnitude better than previously predicted by the literature for a resonant source of this purity. Additionally, we present tomographic methodologies that fully describe the photonic quantum states that we produce, without the use of analytical models, and as a means of verifying the quantum states we create, entitled – "Quantum-referenced SpontaneousEmission Tomography" (Q-SpET). We also design reconfigurable photonic circuits that can be operated at cryogenic temperatures, with zero static power consumption, entitled – "Cladding Layer Manipulation" (CLM). These devices function as on-chip phase shifters, enabling the local reconfiguration of circuit elements using established technologies but removing the need for active power consumption to maintain the reconfigured circuit. These devices are capable ofan Lπ = 12.3 ± 0.3 µm, a ∼7x reduction in length when compared to the thermo-optic phaseshifters used throughout this thesis. Finally, we investigate how pure photon sources operate as part of larger circuits within the typical design rules of photonic quantum circuits. Using this information to accurately model all of the spurious contributions to the final photonic quantumstate, which we call a form of nonlinear noise. This noise can decrease source purity to below 40 %, significantly affecting the fidelity of Hong-Ou-Mandel interference, and subsequently, our ability to reliably create fundamental resources for photonic quantum computers. All of this contributes to our design of a fundamental building block for integrated quantum photonic processors, the functionality of which can be predicted at scale, under the conditions imposed by the rest of the processor
Recommended from our members
MAGNETO-ELECTRIC APPROXIMATE COMPUTATIONAL FRAMEWORK FOR BAYESIAN INFERENCE
Probabilistic graphical models like Bayesian Networks (BNs) are powerful artificial-intelligence formalisms, with similarities to cognition and higher order reasoning in the human brain. These models have been, to great success, applied to several challenging real-world applications. Use of these formalisms to a greater set of applications is impeded by the limitations of the currently used software-based implementations. New emerging-technology based circuit paradigms which leverage physical equivalence, i.e., operating directly on probabilities vs. introducing layers of abstraction, promise orders of magnitude increase in performance and efficiency of BN implementations, enabling networks with millions of random variables. While majority of applications with small network size (100s of nodes) require only single digit precision for accurate results, applications with larger size (1000s to millions of nodes) require higher precision computation. We introduce a new BN integrated circuit fabric based on mixed-signal magneto-electric circuits which perform probabilistic computations based on the principle of approximate computation. Precision scaling in this fabric is logarithmic in area vs. linear in prior directions. Results show 33x area benefit for a 0.001 precision compared to prior direction, while maintaining three orders of magnitude performance benefits vs. 100-core processor implementations
- …