7,440 research outputs found

    A Similarity Measure for GPU Kernel Subgraph Matching

    Full text link
    Accelerator architectures specialize in executing SIMD (single instruction, multiple data) in lockstep. Because the majority of CUDA applications are parallelized loops, control flow information can provide an in-depth characterization of a kernel. CUDAflow is a tool that statically separates CUDA binaries into basic block regions and dynamically measures instruction and basic block frequencies. CUDAflow captures this information in a control flow graph (CFG) and performs subgraph matching across various kernel's CFGs to gain insights to an application's resource requirements, based on the shape and traversal of the graph, instruction operations executed and registers allocated, among other information. The utility of CUDAflow is demonstrated with SHOC and Rodinia application case studies on a variety of GPU architectures, revealing novel thread divergence characteristics that facilitates end users, autotuners and compilers in generating high performing code

    Unsupervised Anomaly-based Malware Detection using Hardware Features

    Get PDF
    Recent works have shown promise in using microarchitectural execution patterns to detect malware programs. These detectors belong to a class of detectors known as signature-based detectors as they catch malware by comparing a program's execution pattern (signature) to execution patterns of known malware programs. In this work, we propose a new class of detectors - anomaly-based hardware malware detectors - that do not require signatures for malware detection, and thus can catch a wider range of malware including potentially novel ones. We use unsupervised machine learning to build profiles of normal program execution based on data from performance counters, and use these profiles to detect significant deviations in program behavior that occur as a result of malware exploitation. We show that real-world exploitation of popular programs such as IE and Adobe PDF Reader on a Windows/x86 platform can be detected with nearly perfect certainty. We also examine the limits and challenges in implementing this approach in face of a sophisticated adversary attempting to evade anomaly-based detection. The proposed detector is complementary to previously proposed signature-based detectors and can be used together to improve security.Comment: 1 page, Latex; added description for feature selection in Section 4, results unchange

    VETA x ray data acquisition and control system

    Get PDF
    We describe the X-ray Data Acquisition and Control System (XDACS) used together with the X-ray Detection System (XDS) to characterize the X-ray image during testing of the AXAF P1/H1 mirror pair at the MSFC X-ray Calibration Facility. A variety of X-ray data were acquired, analyzed and archived during the testing including: mirror alignment, encircled energy, effective area, point spread function, system housekeeping and proportional counter window uniformity data. The system architecture is presented with emphasis placed on key features that include a layered UNIX tool approach, dedicated subsystem controllers, real-time X-window displays, flexibility in combining tools, network connectivity and system extensibility. The VETA test data archive is also described

    VETA-I x ray test analysis

    Get PDF
    This interim report presents some definitive results from our analysis of the VETA-I x-ray testing data. It also provides a description of the hardware and software used in the conduct of the VETA-I x-ray test program performed at the MSFC x-ray Calibration Facility (XRCF). These test results also serve to supply data and information to include in the TRW final report required by DPD 692, DR XC04. To provide an authoritative compendium of results, we have taken nine papers as published in the SPIE Symposium, 'Grazing Incidence X-ray/EUV Optics for Astronomy and Projection Lithography' and have reproduced them as the content of this report

    Fine-grained characterization of edge workloads

    Get PDF
    Edge computing is an emerging paradigm within the field of distributed computing, aimed at bringing data processing capabilities closer to the data-generating sources to enable real-time processing and reduce latency. However, the lack of representative data in the literature poses a significant challenge for evaluating the effectiveness of new algorithms and techniques developed for this paradigm. A part of the process towards alleviating this problem includes creating realistic and relevant workloads for the edge computing community. Research has already been conducted towards this goal, but resulting workload characterizations from these studies have been shown to not give an accurate representation of the workloads. This research gap highlights the need for developing new methodologies that can accurately characterize edge computing workloads. In this work we propose a novel methodology to characterize edge computing workloads, which leverages hardware performance counters to capture the behavior and characteristics of edge workloads in high detail. We explore the concept of representing workloads in a high-dimensional space, and develop a "proof-of-concept" classification model, that classifies workloads on a continuous "imprecise" data spectrum, to demonstrate the effectiveness and potential of the proposed characterization methodology. This research contributes to the field of edge computing by identifying and addressing the limitations of existing edge workload characterization techniques, and also opens up further avenues of research with regards to edge computing workload characterization
    corecore