5,079 research outputs found
Providing Long-Term Participation Incentive in Participatory Sensing
Providing an adequate long-term participation incentive is important for a
participatory sensing system to maintain enough number of active users
(sensors), so as to collect a sufficient number of data samples and support a
desired level of service quality. In this work, we consider the sensor
selection problem in a general time-dependent and location-aware participatory
sensing system, taking the long-term user participation incentive into explicit
consideration. We study the problem systematically under different information
scenarios, regarding both future information and current information
(realization). In particular, we propose a Lyapunov-based VCG auction policy
for the on-line sensor selection, which converges asymptotically to the optimal
off-line benchmark performance, even with no future information and under
(current) information asymmetry. Extensive numerical results show that our
proposed policy outperforms the state-of-art policies in the literature, in
terms of both user participation (e.g., reducing the user dropping probability
by 25% to 90%) and social performance (e.g., increasing the social welfare by
15% to 80%).Comment: This manuscript serves as the online technical report of the article
published in IEEE International Conference on Computer Communications
(INFOCOM), 201
Socially Trusted Collaborative Edge Computing in Ultra Dense Networks
Small cell base stations (SBSs) endowed with cloud-like computing
capabilities are considered as a key enabler of edge computing (EC), which
provides ultra-low latency and location-awareness for a variety of emerging
mobile applications and the Internet of Things. However, due to the limited
computation resources of an individual SBS, providing computation services of
high quality to its users faces significant challenges when it is overloaded
with an excessive amount of computation workload. In this paper, we propose
collaborative edge computing among SBSs by forming SBS coalitions to share
computation resources with each other, thereby accommodating more computation
workload in the edge system and reducing reliance on the remote cloud. A novel
SBS coalition formation algorithm is developed based on the coalitional game
theory to cope with various new challenges in small-cell-based edge systems,
including the co-provisioning of radio access and computing services,
cooperation incentives, and potential security risks. To address these
challenges, the proposed method (1) allows collaboration at both the user-SBS
association stage and the SBS peer offloading stage by exploiting the ultra
dense deployment of SBSs, (2) develops a payment-based incentive mechanism that
implements proportionally fair utility division to form stable SBS coalitions,
and (3) builds a social trust network for managing security risks among SBSs
due to collaboration. Systematic simulations in practical scenarios are carried
out to evaluate the efficacy and performance of the proposed method, which
shows that tremendous edge computing performance improvement can be achieved.Comment: arXiv admin note: text overlap with arXiv:1010.4501 by other author
Optimal Joint Routing and Scheduling in Millimeter-Wave Cellular Networks
Millimeter-wave (mmWave) communication is a promising technology to cope with
the expected exponential increase in data traffic in 5G networks. mmWave
networks typically require a very dense deployment of mmWave base stations
(mmBS). To reduce cost and increase flexibility, wireless backhauling is needed
to connect the mmBSs. The characteristics of mmWave communication, and
specifically its high directional- ity, imply new requirements for efficient
routing and scheduling paradigms. We propose an efficient scheduling method,
so-called schedule-oriented optimization, based on matching theory that
optimizes QoS metrics jointly with routing. It is capable of solving any
scheduling problem that can be formulated as a linear program whose variables
are link times and QoS metrics. As an example of the schedule-oriented
optimization, we show the optimal solution of the maximum throughput fair
scheduling (MTFS). Practically, the optimal scheduling can be obtained even for
networks with over 200 mmBSs. To further increase the runtime performance, we
propose an efficient edge-coloring based approximation algorithm with provable
performance bound. It achieves over 80% of the optimal max-min throughput and
runs 5 to 100 times faster than the optimal algorithm in practice. Finally, we
extend the optimal and approximation algorithms for the cases of multi-RF-chain
mmBSs and integrated backhaul and access networks.Comment: To appear in Proceedings of INFOCOM '1
Overview of Infrastructure Charging, part 4, IMPROVERAIL Project Deliverable 9, âImproved Data Background to Support Current and Future Infrastructure Charging Systemsâ
Improverail aims are to further support the establishment of railway infrastructure management in accordance with Directive 91/440, as well as the new railway infrastructure directives, by developing the necessary tools for modelling the management of railway infrastructure; by evaluating improved methods for capacity and resources management, which allow the improvement of the Life Cycle Costs (LCC) calculating methods, including elements related to vehicle - infrastructure interaction and external costs; and by improving data background in support of charging for use of railway infrastructure. To achieve these objectives, Improverail is organised along 8 workpackages, with specific objectives, responding to the requirements of the task 2.2.1/10 of the 2nd call made in the 5th RTD Framework Programme in December 1999.This part is the task 7.1 (Review of infrastructure charging systems) to the workpackage 7 (Analysis of the relation between infrastructure cost variation and diversity of infrastructure charging systems).Before explaining the economic characteristics of railway and his basic pricing principles, authors must specify the objectives of railways infrastructure charging.principle of pricing ; rail infrastructure charging ; public service obligation ; rail charging practice ; Europe ; Improverail
High-Throughput Computing on High-Performance Platforms: A Case Study
The computing systems used by LHC experiments has historically consisted of
the federation of hundreds to thousands of distributed resources, ranging from
small to mid-size resource. In spite of the impressive scale of the existing
distributed computing solutions, the federation of small to mid-size resources
will be insufficient to meet projected future demands. This paper is a case
study of how the ATLAS experiment has embraced Titan---a DOE leadership
facility in conjunction with traditional distributed high- throughput computing
to reach sustained production scales of approximately 52M core-hours a years.
The three main contributions of this paper are: (i) a critical evaluation of
design and operational considerations to support the sustained, scalable and
production usage of Titan; (ii) a preliminary characterization of a next
generation executor for PanDA to support new workloads and advanced execution
modes; and (iii) early lessons for how current and future experimental and
observational systems can be integrated with production supercomputers and
other platforms in a general and extensible manner
Cloud-based charging management of heterogeneous electric vehicles in a network of charging stations : price incentive vs. capacity expansion
This paper presents a novel cloud-based charging management system for electric vehicles (EVs). Two levels of cloud computing, i.e., local and remote cloud, are employed to meet the different latency requirements of the heterogeneous EVs while exploiting the lower-cost computing in remote clouds. Specifically, we consider time-sensitive EVs at highway exit charging stations and EVs with relaxed timing constraints at parking lot charging stations. We propose algorithms for the interplay among EVs, charging stations, system operator, and clouds. Considering the contention-based random access for EVs to a 4G Long-Term Evolution network, and the quality of service metrics (average waiting time and blocking probability), the model is composed of: queuing-based cloud server planning, capacity planning in charging stations, delay analysis, and profit maximization. We propose and analyze a price-incentive method that shifts heavy load from peak to off-peak hours, a capacity expansion method that accommodates the peak demand by purchasing additional electricity, and a hybrid method of prince-incentive and capacity expansion that balances the immediate charging needs of customers with the alleviation of the peak power grid load through price-incentive based demand control. Numerical results demonstrate the effectiveness of the proposed methods and elucidate the tradeoffs between the methods
Distributed Monte Carlo Simulation
Monte Carlo simulation is an effective way to analyze models of sophisticated problems, but often suffers from high computational complexity. Distributed computing is an effective technology that can be used for compute-intensive applications, such as Monte Carlo simulation. The goal of this thesis is to combine the concepts of Monte Carlo simulation and distributed computing in an effort to develop an efficient system capable of rapidly executing computationally-demanding simulations.;When distributed computing is used to support the simulations of multiple users, a scheduling algorithm is required to allocate resources among the users\u27 jobs. In this thesis, a scheduling algorithm is developed that is suitable for Monte Carlo simulation and utilizes the available distributed-computing resources. The unified framework for scheduling is capable of accommodating classic scheduling algorithms such as equal job share, first-in first-out (FIFO), and proportional fair scheduling. The behavior of the scheduler can be controlled by just three parameters. By choosing appropriate parameter values, individual users and their jobs can be assigned different priorities. By introducing an appropriate analytical model, the role of these parameters on system behavior is thoroughly investigated. Using insights obtained by studying the analytical model, a complete distributed Monte Carlo system is designed and presented as a case study
Technical Report for Trend Prediction Based Intelligent UAV Trajectory Planning for Large-scale Dynamic Scenarios
The unmanned aerial vehicle (UAV)-enabled communication technology is
regarded as an efficient and effective solution for some special application
scenarios where existing terrestrial infrastructures are overloaded to provide
reliable services. To maximize the utility of the UAV-enabled system while
meeting the QoS and energy constraints, the UAV needs to plan its trajectory
considering the dynamic characteristics of scenarios, which is formulated as
the Markov Decision Process (MDP). To solve the above problem, a deep
reinforcement learning (DRL)-based scheme is proposed here, which predicts the
trend of the dynamic scenarios to provide a long-term view for the UAV
trajectory planning. Simulation results validate that our proposed scheme
converges more quickly and achieves the better performance in dynamic
scenarios
- âŠ