4,679 research outputs found

    Lessons learned from the design of a mobile multimedia system in the Moby Dick project

    Get PDF
    Recent advances in wireless networking technology and the exponential development of semiconductor technology have engendered a new paradigm of computing, called personal mobile computing or ubiquitous computing. This offers a vision of the future with a much richer and more exciting set of architecture research challenges than extrapolations of the current desktop architectures. In particular, these devices will have limited battery resources, will handle diverse data types, and will operate in environments that are insecure, dynamic and which vary significantly in time and location. The research performed in the MOBY DICK project is about designing such a mobile multimedia system. This paper discusses the approach made in the MOBY DICK project to solve some of these problems, discusses its contributions, and accesses what was learned from the project

    A CRC usefulness assessment for adaptation layers in satellite systems

    Get PDF
    This paper assesses the real usefulness of CRCs in today's satellite network-to-link adaptation layers under the lights of enhanced error control and framing techniques, focusing on the DVB-S and DVB-S2 standards. Indeed, the outer block codes of their FEC schemes (Reed-Solomon and BCH, respectively) can provide very accurate error-detection information to the receiver in addition to their correction capabilities, at virtually no cost. This handy feature could be used to manage on a frame-by-frame basis what CRCs do locally, on the frames' contents, saving the bandwidth and processing load associated with them, and paving the way for enhanced transport of IP over DVB-S2. Mathematical and experimental results clearly show that if FEC has been properly congured for combined error correction and detection, having an uncorrected event after FEC decoding is likely to be an extremely improbable event. Under such conditions, it seems possible and attractive to optimize the way global error-control is done over satellite links by reducing the role of CRCs, or even by removing them from the overall encapsulation process

    Stochastic macromodeling for hierarchical uncertainty quantification of nonlinear electronic systems

    Get PDF
    A hierarchical stochastic macromodeling approach is proposed for the efficient variability analysis of complex nonlinear electronic systems. A combination of the Transfer Function Trajectory and Polynomial Chaos methods is used to generate stochastic macromodels. In order to reduce the computational complexity of the model generation when the number of stochastic variables increases, a hierarchical system decomposition is used. Pertinent numerical results validate the proposed methodology

    Reliable and Energy Efficient MLC STT-RAM Buffer for CNN Accelerators

    Get PDF
    We propose a lightweight scheme where the formation of a data block is changed in such a way that it can tolerate soft errors significantly better than the baseline. The key insight behind our work is that CNN weights are normalized between -1 and 1 after each convolutional layer, and this leaves one bit unused in half-precision floating-point representation. By taking advantage of the unused bit, we create a backup for the most significant bit to protect it against the soft errors. Also, considering the fact that in MLC STT-RAMs the cost of memory operations (read and write), and reliability of a cell are content-dependent (some patterns take larger current and longer time, while they are more susceptible to soft error), we rearrange the data block to minimize the number of costly bit patterns. Combining these two techniques provides the same level of accuracy compared to an error-free baseline while improving the read and write energy by 9% and 6%, respectively

    Autonomic computing architecture for SCADA cyber security

    Get PDF
    Cognitive computing relates to intelligent computing platforms that are based on the disciplines of artificial intelligence, machine learning, and other innovative technologies. These technologies can be used to design systems that mimic the human brain to learn about their environment and can autonomously predict an impending anomalous situation. IBM first used the term ‘Autonomic Computing’ in 2001 to combat the looming complexity crisis (Ganek and Corbi, 2003). The concept has been inspired by the human biological autonomic system. An autonomic system is self-healing, self-regulating, self-optimising and self-protecting (Ganek and Corbi, 2003). Therefore, the system should be able to protect itself against both malicious attacks and unintended mistakes by the operator
    • 

    corecore