11,568 research outputs found
Economic Efficiency Requires Interaction
We study the necessity of interaction between individuals for obtaining
approximately efficient allocations. The role of interaction in markets has
received significant attention in economic thinking, e.g. in Hayek's 1945
classic paper.
We consider this problem in the framework of simultaneous communication
complexity. We analyze the amount of simultaneous communication required for
achieving an approximately efficient allocation. In particular, we consider two
settings: combinatorial auctions with unit demand bidders (bipartite matching)
and combinatorial auctions with subadditive bidders. For both settings we first
show that non-interactive systems have enormous communication costs relative to
interactive ones. On the other hand, we show that limited interaction enables
us to find approximately efficient allocations
Energy Efficient Ant Colony Algorithms for Data Aggregation in Wireless Sensor Networks
In this paper, a family of ant colony algorithms called DAACA for data
aggregation has been presented which contains three phases: the initialization,
packet transmission and operations on pheromones. After initialization, each
node estimates the remaining energy and the amount of pheromones to compute the
probabilities used for dynamically selecting the next hop. After certain rounds
of transmissions, the pheromones adjustment is performed periodically, which
combines the advantages of both global and local pheromones adjustment for
evaporating or depositing pheromones. Four different pheromones adjustment
strategies are designed to achieve the global optimal network lifetime, namely
Basic-DAACA, ES-DAACA, MM-DAACA and ACS-DAACA. Compared with some other data
aggregation algorithms, DAACA shows higher superiority on average degree of
nodes, energy efficiency, prolonging the network lifetime, computation
complexity and success ratio of one hop transmission. At last we analyze the
characteristic of DAACA in the aspects of robustness, fault tolerance and
scalability.Comment: To appear in Journal of Computer and System Science
Randomized Algorithms for Tracking Distributed Count, Frequencies, and Ranks
We show that randomization can lead to significant improvements for a few
fundamental problems in distributed tracking. Our basis is the {\em
count-tracking} problem, where there are players, each holding a counter
that gets incremented over time, and the goal is to track an
\eps-approximation of their sum continuously at all times,
using minimum communication. While the deterministic communication complexity
of the problem is \Theta(k/\eps \cdot \log N), where is the final value
of when the tracking finishes, we show that with randomization, the
communication cost can be reduced to \Theta(\sqrt{k}/\eps \cdot \log N). Our
algorithm is simple and uses only O(1) space at each player, while the lower
bound holds even assuming each player has infinite computing power. Then, we
extend our techniques to two related distributed tracking problems: {\em
frequency-tracking} and {\em rank-tracking}, and obtain similar improvements
over previous deterministic algorithms. Both problems are of central importance
in large data monitoring and analysis, and have been extensively studied in the
literature.Comment: 19 pages, 1 figur
Recommended from our members
Parallel data compression
Data compression schemes remove data redundancy in communicated and stored data and increase the effective capacities of communication and storage devices. Parallel algorithms and implementations for textual data compression are surveyed. Related concepts from parallel computation and information theory are briefly discussed. Static and dynamic methods for codeword construction and transmission on various models of parallel computation are described. Included are parallel methods which boost system speed by coding data concurrently, and approaches which employ multiple compression techniques to improve compression ratios. Theoretical and empirical comparisons are reported and areas for future research are suggested
A low complexity resource allocation algorithm for multicast service delivery in OFDMA networks
Allocating and managing radio resources to multicast transmissions in Orthogonal Frequency-Division Multiple Access (OFDMA) systems is the challenging research issue addressed by this paper. A subgrouping technique, which divides the subscribers into subgroups according to the experienced channel quality, is considered to overcome the throughput limitations of conventional multicast data delivery schemes. A low complexity algorithm, designed to work with different resource allocation strategies, is also proposed to reduce the computational complexity of the subgroup formation problem. Simulation results, carried out by considering the Long Term Evolution (LTE) system based on OFDMA, testify the effectiveness of the proposed solution, which achieves a near-optimal performance with a limited computational load for the system
Recommended from our members
Multimedia delivery in the future internet
The term âNetworked Mediaâ implies that all kinds of media including text, image, 3D graphics, audio
and video are produced, distributed, shared, managed and consumed on-line through various networks,
like the Internet, Fiber, WiFi, WiMAX, GPRS, 3G and so on, in a convergent manner [1]. This white
paper is the contribution of the Media Delivery Platform (MDP) cluster and aims to cover the Networked
challenges of the Networked Media in the transition to the Future of the Internet.
Internet has evolved and changed the way we work and live. End users of the Internet have been confronted
with a bewildering range of media, services and applications and of technological innovations concerning
media formats, wireless networks, terminal types and capabilities. And there is little evidence that the pace
of this innovation is slowing. Today, over one billion of users access the Internet on regular basis, more
than 100 million users have downloaded at least one (multi)media file and over 47 millions of them do so
regularly, searching in more than 160 Exabytes1 of content. In the near future these numbers are expected
to exponentially rise. It is expected that the Internet content will be increased by at least a factor of 6, rising
to more than 990 Exabytes before 2012, fuelled mainly by the users themselves. Moreover, it is envisaged
that in a near- to mid-term future, the Internet will provide the means to share and distribute (new)
multimedia content and services with superior quality and striking flexibility, in a trusted and personalized
way, improving citizensâ quality of life, working conditions, edutainment and safety.
In this evolving environment, new transport protocols, new multimedia encoding schemes, cross-layer inthe
network adaptation, machine-to-machine communication (including RFIDs), rich 3D content as well as
community networks and the use of peer-to-peer (P2P) overlays are expected to generate new models of
interaction and cooperation, and be able to support enhanced perceived quality-of-experience (PQoE) and
innovative applications âon the moveâ, like virtual collaboration environments, personalised services/
media, virtual sport groups, on-line gaming, edutainment. In this context, the interaction with content
combined with interactive/multimedia search capabilities across distributed repositories, opportunistic P2P
networks and the dynamic adaptation to the characteristics of diverse mobile terminals are expected to
contribute towards such a vision.
Based on work that has taken place in a number of EC co-funded projects, in Framework Program 6 (FP6)
and Framework Program 7 (FP7), a group of experts and technology visionaries have voluntarily
contributed in this white paper aiming to describe the status, the state-of-the art, the challenges and the way
ahead in the area of Content Aware media delivery platforms
- âŠ