5,379 research outputs found
Critique of Architectures for Long-Term Digital Preservation
Evolving technology and fading human memory threaten the long-term intelligibility of many kinds of documents. Furthermore, some records are susceptible to improper alterations that make them untrustworthy. Trusted Digital Repositories (TDRs) and Trustworthy Digital Objects (TDOs) seem to be the only broadly applicable digital preservation methodologies proposed. We argue that the TDR approach has shortfalls as a method for long-term digital preservation of sensitive information. Comparison of TDR and TDO methodologies suggests differentiating near-term preservation measures from what is needed for the long term.
TDO methodology addresses these needs, providing for making digital documents durably intelligible. It uses EDP standards for a few file formats and XML structures for text documents. For other information formats, intelligibility is assured by using a virtual computer. To protect sensitive information—content whose inappropriate alteration might mislead its readers, the integrity and authenticity of each TDO is made testable by embedded public-key cryptographic message digests and signatures. Key authenticity is protected recursively in a social hierarchy. The proper focus for long-term preservation technology is signed packages that each combine a record collection with its metadata and that also bind context—Trustworthy Digital Objects.
Recommended from our members
Improving virtual memory performance in virtualized environments
Virtual Memory is a major system performance bottleneck in virtualized environments. In addition to expensive address translations, frequent virtual machine context switches are common in virtualized environments, resulting in increased TLB miss rates, subsequent expensive page walks and data cache contention due to incoming page table entries evicting useful data. Orthogonally, translation coherence, which is currently an expensive operation implemented in software, can consume up to 50% of the runtime of an application executing on the guest. To improve the performance of virtual memory in virtualized environments, two solutions have been proposed in this thesis - namely, (1) Context Switch Aware Large TLB (CSALT), an architecture which addresses the problem of increased TLB miss rates and their adverse impact on data caches. CSALT copes with the increased demand of context switches by storing a large number TLB entries. It mitigates data cache contention by employing a novel TLB-aware cache partitioning scheme. On 8-core systems that switch between two virtual machine contexts executing multi-threaded workloads, CSALT achieves an average performance improvement of 85% over a baseline with conventional L1-L2 TLBs and 25% over a baseline which has a large L3 TLB (2) Translation Coherence using Addressable TLBs (TCAT), a hardware translation coherence scheme which eliminates almost all of the overheads associated with address translation coherence. TCAT overlays translation coherence atop cache coherence to accurately identify slave cores. It then leverages the addressable Part-Of-Memory TLB (POM-TLB) to eliminate expensive Inter Processor Interrupts (IPI) and achieve precise invalidations on the slave core. On 8-core systems with one virtual machine context executing multi-threaded workloads, TCAT achieves an average performance improvement of 13% over the kvmtlb baselineElectrical and Computer Engineerin
Defense in Depth of Resource-Constrained Devices
The emergent next generation of computing, the so-called Internet of Things (IoT), presents significant challenges to security, privacy, and trust. The devices commonly used in IoT scenarios are often resource-constrained with reduced computational strength, limited power consumption, and stringent availability requirements. Additionally, at least in the consumer arena, time-to-market is often prioritized at the expense of quality assurance and security. An initial lack of standards has compounded the problems arising from this rapid development. However, the explosive growth in the number and types of IoT devices has now created a multitude of competing standards and technology silos resulting in a highly fragmented threat model. Tens of billions of these devices have been deployed in consumers\u27 homes and industrial settings. From smart toasters and personal health monitors to industrial controls in energy delivery networks, these devices wield significant influence on our daily lives. They are privy to highly sensitive, often personal data and responsible for real-world, security-critical, physical processes. As such, these internet-connected things are highly valuable and vulnerable targets for exploitation. Current security measures, such as reactionary policies and ad hoc patching, are not adequate at this scale. This thesis presents a multi-layered, defense in depth, approach to preventing and mitigating a myriad of vulnerabilities associated with the above challenges. To secure the pre-boot environment, we demonstrate a hardware-based secure boot process for devices lacking secure memory. We introduce a novel implementation of remote attestation backed by blockchain technologies to address hardware and software integrity concerns for the long-running, unsupervised, and rarely patched systems found in industrial IoT settings. Moving into the software layer, we present a unique method of intraprocess memory isolation as a barrier to several prevalent classes of software vulnerabilities. Finally, we exhibit work on network analysis and intrusion detection for the low-power, low-latency, and low-bandwidth wireless networks common to IoT applications. By targeting these areas of the hardware-software stack, we seek to establish a trustworthy system that extends from power-on through application runtime
Improving address translation performance in virtualized multi-tenant systems
With the explosive growth in dataset sizes, application memory footprints are commonly reaching hundreds of GBs. Such huge datasets pressure the TLBs, resulting
in frequent misses that must be resolved through a page walk – a long-latency pointer
chase through multiple levels of the in-memory radix-tree-based page table. Page walk
latency is particularly high under virtualization where address translation mandates traversing two radix-tree page tables in a process called a nested page walk, performing
up to 24 memory accesses. Page walk latency can be also amplified by the effects
caused by the colocation of applications on the same server used in an attempt to increase utilization. Under colocation, cache contention makes cache misses during a
nested page walk more frequent, piling up page walk latency. Both virtualization and
colocation are widely adopted in cloud platforms, such as Amazon Web Services and
Google Cloud Engine. As a result, in cloud environments, page walk latency can
reach hundreds of cycles, significantly reducing the overall application’s performance.
This thesis addresses the problem of the high page walk latency by 1 identifying
the sources of the high page walk latency under virtualization and/or colocation, and
2 proposing hardware and software techniques that accelerate page walks by means
of new memory allocation strategies for the page table and data which can be easily
adopted by existing systems.
Firstly, we quantify how the dataset size growth, virtualization, and colocation affect page walk latency. We also study how a high page walk latency affects perform ance. Due to the lack of dedicated tools for evaluating address translation overhead
on modern processors, we design a methodology to vary the page walk latency experienced by an application running on real hardware. To quantify the performance impact
of address translation, we measure the application’s execution time while varying the
page walk latency. We find that under virtualization, address translation considerably
limits performance: an application can waste up to 68% of execution time due to stalls
originating from page walks. In addition, we investigate which accesses from a nested
page walk are most significant for the overall page walk latency by examining from
where in the memory hierarchy these accesses are served. We find that accesses to the
deeper levels of the page table radix tree are responsible for most of the overall page
walk latency.
Based on these observations, we introduce two address translation acceleration
techniques that can be applied to any ISA that employs radix-tree page tables and
nested page walks. The first of these techniques is Prefetched Address Translation
(ASAP), a new software-hardware approach for mitigating the high page walk latency
caused by virtualization and/or application colocation. At the heart of ASAP is a
lightweight technique for directly indexing individual levels of the page table radix
tree. Direct indexing enables ASAP to fetch nodes from deeper levels of the page
table without first accessing the preceding levels, thus lowering the page walk latency.
ASAP is fully compatible with the existing radix-tree-based page table and requires
only incremental and isolated changes to the memory subsystem.
The second technique is PTEMagnet, a new software-only approach for reducing
address translation latency under virtualization and application colocation. Initially,
we identify a new address translation bottleneck caused by memory fragmentation
stemming from the interaction of virtualization, application colocation, and the Linux
memory allocator. The fragmentation results in the effective cache footprint of the
host page table being larger than that of the guest page table. The bloated footprint
of the host page table leads to frequent cache misses during nested page walks, increasing page walk latency. In response to these observations, we propose PTEMag net. PTEMagnet prevents memory fragmentation by fine-grained reservation-based
memory allocation in the guest OS. PTEMagnet is fully legacy-preserving, requiring
no modifications to either user code or mechanisms for address translation and virtualization.
In summary, this thesis proposes non-disruptive upgrades to the virtual memory
subsystem for reducing page walk latency in virtualized deployments. In doing so,
this thesis evaluates the impact of page walk latency on the application’s performance, identifies the bottlenecks of the existing address translation mechanism caused
by virtualization, application colocation, and the Linux memory allocator, and proposes software-hardware and software-only solutions for eliminating the bottlenecks
Recommended from our members
Duplo: A framework for OCaml post-link optimisation
We present a novel framework,
Duplo
, for the low-level post-link optimisation of OCaml programs, achieving a speedup of 7% and a reduction of at least 15% of the code size of widely-used OCaml applications. Unlike existing post-link optimisers, which typically operate on target-specific machine code, our framework operates on a Low-Level Intermediate Representation (LLIR) capable of representing both the OCaml programs and any C dependencies they invoke through the foreign-function interface (FFI). LLIR is analysed, transformed and lowered to machine code by our post-link optimiser, LLIR-OPT. Most importantly, LLIR allows the optimiser to cross the OCaml-C language boundary, mitigating the overhead incurred by the FFI and enabling analyses and transformations in a previously unavailable context. The optimised IR is then lowered to amd64 machine code through the existing target-specific code generator of LLVM, modified to handle garbage collection just as effectively as the native OCaml backend. We equip our optimiser with a suite of SSA-based transformations and points-to analyses capable of capturing the semantics and representing the memory models of both languages, along with a cross-language inliner to embed C methods into OCaml callers. We evaluate the gains of our framework, which can be attributed to both our optimiser and the more sophisticated amd64 backend of LLVM, on a wide-range of widely-used OCaml applications, as well as an existing suite of micro- and macro-benchmarks used to track the performance of the OCaml compiler.
EPSRC EP/P020011/1, Cambridge Trust
Optimizing Epicardial Restraint and Reinforcement Following Myocardial Infarction: Moving Towards Localized, Biomimetic, and Multitherapeutic Options
The mechanical reinforcement of the ventricular wall after a myocardial infarction has been shown to modulate and attenuate negative remodeling that can lead to heart failure. Strategies include wraps, meshes, cardiac patches, or fluid-filled bladders. Here, we review the literature describing these strategies in the two broad categories of global restraint and local reinforcement. We further subdivide the global restraint category into biventricular and univentricular support. We discuss efforts to optimize devices in each of these categories, particularly in the last five years. These include adding functionality, biomimicry, and adjustability. We also discuss computational models of these strategies, and how they can be used to predict the reduction of stresses in the heart muscle wall. We discuss the range of timing of intervention that has been reported. Finally, we give a perspective on how novel fabrication technologies, imaging techniques, and computational models could potentially enhance these therapeutic strategies. Keywords: ventricular restraint; infarct reinforcement; biomimetic
- …