33,903 research outputs found
Complexity-Aware Scheduling for an LDPC Encoded C-RAN Uplink
Centralized Radio Access Network (C-RAN) is a new paradigm for wireless
networks that centralizes the signal processing in a computing cloud, allowing
commodity computational resources to be pooled. While C-RAN improves
utilization and efficiency, the computational load occasionally exceeds the
available resources, creating a computational outage. This paper provides a
mathematical characterization of the computational outage probability for
low-density parity check (LDPC) codes, a common class of error-correcting
codes. For tractability, a binary erasures channel is assumed. Using the
concept of density evolution, the computational demand is determined for a
given ensemble of codes as a function of the erasure probability. The analysis
reveals a trade-off: aggressively signaling at a high rate stresses the
computing pool, while conservatively backing-off the rate can avoid
computational outages. Motivated by this trade-off, an effective
computationally aware scheduling algorithm is developed that balances demands
for high throughput and low outage rates.Comment: Conference on Information Sciences and Systems (CISS) 2017, to appea
The Partial Evaluation Approach to Information Personalization
Information personalization refers to the automatic adjustment of information
content, structure, and presentation tailored to an individual user. By
reducing information overload and customizing information access,
personalization systems have emerged as an important segment of the Internet
economy. This paper presents a systematic modeling methodology - PIPE
(`Personalization is Partial Evaluation') - for personalization.
Personalization systems are designed and implemented in PIPE by modeling an
information-seeking interaction in a programmatic representation. The
representation supports the description of information-seeking activities as
partial information and their subsequent realization by partial evaluation, a
technique for specializing programs. We describe the modeling methodology at a
conceptual level and outline representational choices. We present two
application case studies that use PIPE for personalizing web sites and describe
how PIPE suggests a novel evaluation criterion for information system designs.
Finally, we mention several fundamental implications of adopting the PIPE model
for personalization and when it is (and is not) applicable.Comment: Comprehensive overview of the PIPE model for personalizatio
Quid Pro Quo: A Mechanism for Fair Collaboration in Networked Systems
Collaboration may be understood as the execution of coordinated tasks (in the
most general sense) by groups of users, who cooperate for achieving a common
goal. Collaboration is a fundamental assumption and requirement for the correct
operation of many communication systems. The main challenge when creating
collaborative systems in a decentralized manner is dealing with the fact that
users may behave in selfish ways, trying to obtain the benefits of the tasks
but without participating in their execution. In this context, Game Theory has
been instrumental to model collaborative systems and the task allocation
problem, and to design mechanisms for optimal allocation of tasks. In this
paper, we revise the classical assumptions and propose a new approach to this
problem. First, we establish a system model based on heterogenous nodes (users,
players), and propose a basic distributed mechanism so that, when a new task
appears, it is assigned to the most suitable node. The classical technique for
compensating a node that executes a task is the use of payments (which in most
networks are hard or impossible to implement). Instead, we propose a
distributed mechanism for the optimal allocation of tasks without payments. We
prove this mechanism to be robust event in the presence of independent selfish
or rationally limited players. Additionally, our model is based on very weak
assumptions, which makes the proposed mechanisms susceptible to be implemented
in networked systems (e.g., the Internet).Comment: 23 pages, 5 figures, 3 algorithm
The organization and management of the Virtual Astronomical Observatory
The U.S. Virtual Astronomical Observatory (VAO; http://www.us-vao.org/) has
been in operation since May 2010. Its goal is to enable new science through
efficient integration of distributed multi-wavelength data. This paper
describes the management and organization of the VAO, and emphasizes the
techniques used to ensure efficiency in a distributed organization. Management
methods include using an annual program plan as the basis for establishing
contracts with member organizations, regular communication, and monitoring of
processes.Comment: 9 pages, 3 figures. SPIE Conference 8449: Modeling, Systems
Engineering, and Project Management for Astronomy
Prochlo: Strong Privacy for Analytics in the Crowd
The large-scale monitoring of computer users' software activities has become
commonplace, e.g., for application telemetry, error reporting, or demographic
profiling. This paper describes a principled systems architecture---Encode,
Shuffle, Analyze (ESA)---for performing such monitoring with high utility while
also protecting user privacy. The ESA design, and its Prochlo implementation,
are informed by our practical experiences with an existing, large deployment of
privacy-preserving software monitoring.
(cont.; see the paper
- …