214,685 research outputs found

    Application of advanced computational technology to propulsion CFD

    Get PDF
    The Internal Fluid Mechanics Division of the NASA Lewis Research Center is combining the key elements of computational fluid dynamics, aerothermodynamic experiments, and advanced computational technology to bring internal computational fluid dynamics (ICFM) to a state of practical application for aerospace propulsion system design. This paper presents an overview of efforts under way at NASA Lewis to advance and apply computational technology to ICFM. These efforts include the use of modern, software engineering principles for code development, the development of an AI-based user-interface for large codes, the establishment of a high-performance, data communications network to link ICFM researchers and facilities, and the application of parallel processing to speed up computationally intensive and/or time-critical ICFM problems. A multistage compressor flow physics program is cited as an example of efforts to use advanced computational technology to enhance a current NASA Lewis ICFM research program

    Analysis of Latency-Aware Network Slicing in 5G Packet xHaul Networks

    Get PDF
    Packet-switched xHaul networks are a scalable solution enabling convergent transport of diverse types of radio data flows, such as fronthaul/midhaul/backhaul (FH/MH/BH) flows, between remote sites and a central site (hub) in 5G radio access networks (RANs). Such networks can be realized using the cost-efficient Ethernet technology, which enhanced with time-sensitive networking (TSN) features allows for prioritized transmission of latency-sensitive fronthaul flows. Provisioning of multiple types of 5G services of different service requirements in a shared network, commonly referred to as network slicing, requires adequate handling of transported data flows in order to satisfy particular service/slice requirements. In this work, we investigate two traffic prioritization policies, namely, flow-aware (FA) and latency-aware (LA), in a packet-switched xHaul network supporting slices of different latency requirements. We evaluate the effectiveness of the policies in a network-planning case study, where virtualized radio processing resources allocated at the processing pool (PP) facilities, for two slices related to enhanced mobile broadband (eMBB) and ultra-reliable low latency communications (URLLC) services, are subject to optimization. Using numerical experiments, we analyze PP cost savings from applying the LA policy (vs. FA) in various network scenarios. The savings in active PPs reach up to 40%-60% in ring scenarios and 30% in a mesh network, whereas the gains in overall PP cost are up to 20% for the cost values assumed in the analysis

    A selected glossary of electronic data interchange and related terms

    No full text
    School of Managemen

    Architecture and Design of Medical Processor Units for Medical Networks

    Full text link
    This paper introduces analogical and deductive methodologies for the design medical processor units (MPUs). From the study of evolution of numerous earlier processors, we derive the basis for the architecture of MPUs. These specialized processors perform unique medical functions encoded as medical operational codes (mopcs). From a pragmatic perspective, MPUs function very close to CPUs. Both processors have unique operation codes that command the hardware to perform a distinct chain of subprocesses upon operands and generate a specific result unique to the opcode and the operand(s). In medical environments, MPU decodes the mopcs and executes a series of medical sub-processes and sends out secondary commands to the medical machine. Whereas operands in a typical computer system are numerical and logical entities, the operands in medical machine are objects such as such as patients, blood samples, tissues, operating rooms, medical staff, medical bills, patient payments, etc. We follow the functional overlap between the two processes and evolve the design of medical computer systems and networks.Comment: 17 page
    • …
    corecore