3,617 research outputs found
Energy Efficiency and Quality of Services in Virtualized Cloud Radio Access Network
Cloud Radio Access Network (C-RAN) is being widely studied for soft and green fifth generation of Long Term Evolution - Advanced (LTE-A). The recent technology advancement in network virtualization function (NFV) and software defined radio (SDR) has enabled virtualization of Baseband Units (BBU) and sharing of underlying general purpose processing (GPP) infrastructure. Also, new innovations in optical transport network (OTN) such as Dark Fiber provides low latency and high bandwidth channels that can support C-RAN for more than forty-kilometer radius. All these advancements make C-RAN feasible and practical. Several virtualization strategies and architectures are proposed for C-RAN and it has been established that C-RAN offers higher energy efficiency and better resource utilization than the current decentralized radio access network (D-RAN). This project studies proposed resource utilization strategy and device a method to calculate power utilization. Then proposes and analyzes a new resource management and virtual BBU placement strategy for C-RAN based on demand prediction and inter-BBU communication load. The new approach is compared with existing state of art strategies with same input scenarios and load. The trade-offs between energy efficiency and quality of services is discussed. The project concludes with comparison between different strategies based on complexity of the system, performance in terms of service availability and optimization efficiency in different scenarios
HEPCloud, a New Paradigm for HEP Facilities: CMS Amazon Web Services Investigation
Historically, high energy physics computing has been performed on large
purpose-built computing systems. These began as single-site compute facilities,
but have evolved into the distributed computing grids used today. Recently,
there has been an exponential increase in the capacity and capability of
commercial clouds. Cloud resources are highly virtualized and intended to be
able to be flexibly deployed for a variety of computing tasks. There is a
growing nterest among the cloud providers to demonstrate the capability to
perform large-scale scientific computing. In this paper, we discuss results
from the CMS experiment using the Fermilab HEPCloud facility, which utilized
both local Fermilab resources and virtual machines in the Amazon Web Services
Elastic Compute Cloud. We discuss the planning, technical challenges, and
lessons learned involved in performing physics workflows on a large-scale set
of virtualized resources. In addition, we will discuss the economics and
operational efficiencies when executing workflows both in the cloud and on
dedicated resources.Comment: 15 pages, 9 figure
Understanding the Computational Requirements of Virtualized Baseband Units using a Programmable Cloud Radio Access Network Testbed
Cloud Radio Access Network (C-RAN) is emerging as a transformative
architecture for the next generation of mobile cellular networks. In C-RAN, the
Baseband Unit (BBU) is decoupled from the Base Station (BS) and consolidated in
a centralized processing center. While the potential benefits of C-RAN have
been studied extensively from the theoretical perspective, there are only a few
works that address the system implementation issues and characterize the
computational requirements of the virtualized BBU. In this paper, a
programmable C-RAN testbed is presented where the BBU is virtualized using the
OpenAirInterface (OAI) software platform, and the eNodeB and User Equipment
(UEs) are implemented using USRP boards. Extensive experiments have been
performed in a FDD downlink LTE emulation system to characterize the
performance and computing resource consumption of the BBU under various
conditions. It is shown that the processing time and CPU utilization of the BBU
increase with the channel resources and with the Modulation and Coding Scheme
(MCS) index, and that the CPU utilization percentage can be well approximated
as a linear increasing function of the maximum downlink data rate. These
results provide real-world insights into the characteristics of the BBU in
terms of computing resource and power consumption, which may serve as inputs
for the design of efficient resource-provisioning and allocation strategies in
C-RAN systems.Comment: In Proceedings of the IEEE International Conference on Autonomic
Computing (ICAC), July 201
MorphoSys: efficient colocation of QoS-constrained workloads in the cloud
In hosting environments such as IaaS clouds, desirable application performance is usually guaranteed through the use of Service Level Agreements (SLAs), which specify minimal fractions of resource capacities that must be allocated for unencumbered use for proper operation. Arbitrary colocation of applications with different SLAs on a single host may result in inefficient utilization of the host’s resources. In this paper, we propose that periodic resource allocation and consumption models -- often used to characterize real-time workloads -- be used for a more granular expression of SLAs. Our proposed SLA model has the salient feature that it exposes flexibilities that enable the infrastructure provider to safely transform SLAs from one form to another for the purpose of achieving more efficient colocation. Towards that goal, we present MORPHOSYS: a framework for a service that allows the manipulation of SLAs to enable efficient colocation of arbitrary workloads in a dynamic setting. We present results from extensive trace-driven simulations of colocated Video-on-Demand servers in a cloud setting. These results show that potentially-significant reduction in wasted resources (by as much as 60%) are possible using MORPHOSYS.National Science Foundation (0720604, 0735974, 0820138, 0952145, 1012798
HPC Cloud for Scientific and Business Applications: Taxonomy, Vision, and Research Challenges
High Performance Computing (HPC) clouds are becoming an alternative to
on-premise clusters for executing scientific applications and business
analytics services. Most research efforts in HPC cloud aim to understand the
cost-benefit of moving resource-intensive applications from on-premise
environments to public cloud platforms. Industry trends show hybrid
environments are the natural path to get the best of the on-premise and cloud
resources---steady (and sensitive) workloads can run on on-premise resources
and peak demand can leverage remote resources in a pay-as-you-go manner.
Nevertheless, there are plenty of questions to be answered in HPC cloud, which
range from how to extract the best performance of an unknown underlying
platform to what services are essential to make its usage easier. Moreover, the
discussion on the right pricing and contractual models to fit small and large
users is relevant for the sustainability of HPC clouds. This paper brings a
survey and taxonomy of efforts in HPC cloud and a vision on what we believe is
ahead of us, including a set of research challenges that, once tackled, can
help advance businesses and scientific discoveries. This becomes particularly
relevant due to the fast increasing wave of new HPC applications coming from
big data and artificial intelligence.Comment: 29 pages, 5 figures, Published in ACM Computing Surveys (CSUR
- …