7,843 research outputs found
Enabling Micro-level Demand-Side Grid Flexiblity in Resource Constrained Environments
The increased penetration of uncertain and variable renewable energy presents
various resource and operational electric grid challenges. Micro-level
(household and small commercial) demand-side grid flexibility could be a
cost-effective strategy to integrate high penetrations of wind and solar
energy, but literature and field deployments exploring the necessary
information and communication technologies (ICTs) are scant. This paper
presents an exploratory framework for enabling information driven grid
flexibility through the Internet of Things (IoT), and a proof-of-concept
wireless sensor gateway (FlexBox) to collect the necessary parameters for
adequately monitoring and actuating the micro-level demand-side. In the summer
of 2015, thirty sensor gateways were deployed in the city of Managua
(Nicaragua) to develop a baseline for a near future small-scale demand response
pilot implementation. FlexBox field data has begun shedding light on
relationships between ambient temperature and load energy consumption, load and
building envelope energy efficiency challenges, latency communication network
challenges, and opportunities to engage existing demand-side user behavioral
patterns. Information driven grid flexibility strategies present great
opportunity to develop new technologies, system architectures, and
implementation approaches that can easily scale across regions, incomes, and
levels of development
Enabling stream processing for people-centric IoT based on the fog computing paradigm
The world of machine-to-machine (M2M) communication is gradually moving from vertical single purpose solutions to multi-purpose and collaborative applications interacting across industry verticals, organizations and people - A world of Internet of Things (IoT). The dominant approach for delivering IoT applications relies on the development of cloud-based IoT platforms that collect all the data generated by the sensing elements and centrally process the information to create real business value. In this paper, we present a system that follows the Fog Computing paradigm where the sensor resources, as well as the intermediate layers between embedded devices and cloud computing datacenters, participate by providing computational, storage, and control. We discuss the design aspects of our system and present a pilot deployment for the evaluating the performance in a real-world environment. Our findings indicate that Fog Computing can address the ever-increasing amount of data that is inherent in an IoT world by effective communication among all elements of the architecture
Cloud engineering is search based software engineering too
Many of the problems posed by the migration of computation to cloud platforms can be formulated and solved using techniques associated with Search Based Software Engineering (SBSE). Much of cloud software engineering involves problems of optimisation: performance, allocation, assignment and the dynamic balancing of resources to achieve pragmatic trade-offs between many competing technical and business objectives. SBSE is concerned with the application of computational search and optimisation to solve precisely these kinds of software engineering challenges. Interest in both cloud computing and SBSE has grown rapidly in the past five years, yet there has been little work on SBSE as a means of addressing cloud computing challenges. Like many computationally demanding activities, SBSE has the potential to benefit from the cloud; ‘SBSE in the cloud’. However, this paper focuses, instead, of the ways in which SBSE can benefit cloud computing. It thus develops the theme of ‘SBSE for the cloud’, formulating cloud computing challenges in ways that can be addressed using SBSE
Predicting Intermediate Storage Performance for Workflow Applications
Configuring a storage system to better serve an application is a challenging
task complicated by a multidimensional, discrete configuration space and the
high cost of space exploration (e.g., by running the application with different
storage configurations). To enable selecting the best configuration in a
reasonable time, we design an end-to-end performance prediction mechanism that
estimates the turn-around time of an application using storage system under a
given configuration. This approach focuses on a generic object-based storage
system design, supports exploring the impact of optimizations targeting
workflow applications (e.g., various data placement schemes) in addition to
other, more traditional, configuration knobs (e.g., stripe size or replication
level), and models the system operation at data-chunk and control message
level.
This paper presents our experience to date with designing and using this
prediction mechanism. We evaluate this mechanism using micro- as well as
synthetic benchmarks mimicking real workflow applications, and a real
application.. A preliminary evaluation shows that we are on a good track to
meet our objectives: it can scale to model a workflow application run on an
entire cluster while offering an over 200x speedup factor (normalized by
resource) compared to running the actual application, and can achieve, in the
limited number of scenarios we study, a prediction accuracy that enables
identifying the best storage system configuration
- …