3,727 research outputs found

    On the Efficiency-vs-Security Tradeoff in the Smart Grid

    Full text link
    The smart grid is envisioned to significantly enhance the efficiency of energy consumption, by utilizing two-way communication channels between consumers and operators. For example, operators can opportunistically leverage the delay tolerance of energy demands in order to balance the energy load over time, and hence, reduce the total operational cost. This opportunity, however, comes with security threats, as the grid becomes more vulnerable to cyber-attacks. In this paper, we study the impact of such malicious cyber-attacks on the energy efficiency of the grid in a simplified setup. More precisely, we consider a simple model where the energy demands of the smart grid consumers are intercepted and altered by an active attacker before they arrive at the operator, who is equipped with limited intrusion detection capabilities. We formulate the resulting optimization problems faced by the operator and the attacker and propose several scheduling and attack strategies for both parties. Interestingly, our results show that, as opposed to facilitating cost reduction in the smart grid, increasing the delay tolerance of the energy demands potentially allows the attacker to force increased costs on the system. This highlights the need for carefully constructed and robust intrusion detection mechanisms at the operator.Comment: A shorter version appears in IEEE CDC 201

    Proactive Customer Service: Operational Benefits and Economic Frictions

    Get PDF
    Problem Definition: We study a service setting where the provider has information about some customers' future service needs and may initiate service for such customers proactively, if they agree to be flexible with respect to the timing of service delivery. Academic / Practical Relevance: Information about future customer service needs is becoming increasingly available through remote monitoring systems and data analytics. However, the literature has not systematically examined proactive service as a tool that can be used to better match demand to service supply when customers are strategic. Methodology: We combine i) queueing theory, and in particular a diffusion approximation developed specifically for this problem that allows us to derive analytic approximations for customer waiting times, with ii) game theory, which captures customer incentives to adopt proactive service. Results: We show that proactive service can reduce customer waiting times, even if only a relatively small proportion of customers agree to be flexible, the information lead time is limited, and the system makes occasional errors in providing proactive service - in fact we show that the system's ability to tolerate errors increases with (nominal) utilization. Nevertheless, we show that these benefits may fail to materialize in equilibrium because of economic frictions: customers will under-adopt proactive service (due to free-riding) and over-join the system (due to negative congestion-based externalities). We also show that the service provider can incentivize optimal customer behavior through appropriate pricing. Managerial Implications: Our results suggest that proactive service may offer substantial operational benefits, but caution that it may fail to fulfill its potential due to customer self-interested behavior

    Load shedding in network monitoring applications

    Get PDF
    Monitoring and mining real-time network data streams are crucial operations for managing and operating data networks. The information that network operators desire to extract from the network traffic is of different size, granularity and accuracy depending on the measurement task (e.g., relevant data for capacity planning and intrusion detection are very different). To satisfy these different demands, a new class of monitoring systems is emerging to handle multiple and arbitrary monitoring applications. Such systems must inevitably cope with the effects of continuous overload situations due to the large volumes, high data rates and bursty nature of the network traffic. These overload situations can severely compromise the accuracy and effectiveness of monitoring systems, when their results are most valuable to network operators. In this thesis, we propose a technique called load shedding as an effective and low-cost alternative to over-provisioning in network monitoring systems. It allows these systems to handle efficiently overload situations in the presence of multiple, arbitrary and competing monitoring applications. We present the design and evaluation of a predictive load shedding scheme that can shed excess load in front of extreme traffic conditions and maintain the accuracy of the monitoring applications within bounds defined by end users, while assuring a fair allocation of computing resources to non-cooperative applications. The main novelty of our scheme is that it considers monitoring applications as black boxes, with arbitrary (and highly variable) input traffic and processing cost. Without any explicit knowledge of the application internals, the proposed scheme extracts a set of features from the traffic streams to build an on-line prediction model of the resource requirements of each monitoring application, which is used to anticipate overload situations and control the overall resource usage by sampling the input packet streams. This way, the monitoring system preserves a high degree of flexibility, increasing the range of applications and network scenarios where it can be used. Since not all monitoring applications are robust against sampling, we then extend our load shedding scheme to support custom load shedding methods defined by end users, in order to provide a generic solution for arbitrary monitoring applications. Our scheme allows the monitoring system to safely delegate the task of shedding excess load to the applications and still guarantee fairness of service with non-cooperative users. We implemented our load shedding scheme in an existing network monitoring system and deployed it in a research ISP network. We present experimental evidence of the performance and robustness of our system with several concurrent monitoring applications during long-lived executions and using real-world traffic traces.Postprint (published version

    Robust Optimization Framework to Operating Room Planning and Scheduling in Stochastic Environment

    Get PDF
    Arrangement of surgical activities can be classified as a three-level process that directly impacts the overall performance of a healthcare system. The goal of this dissertation is to study hierarchical planning and scheduling problems of operating room (OR) departments that arise in a publicly funded hospital. Uncertainty in surgery durations and patient arrivals, the existence of multiple resources and competing performance measures are among the important aspect of OR problems in practice. While planning can be viewed as the compromise of supply and demand within the strategic and tactical stages, scheduling is referred to the development of a detailed timetable that determines operational daily assignment of individual cases. Therefore, it is worthwhile to put effort in optimization of OR planning and surgical scheduling. We have considered several extensions of previous models and described several real-world applications. Firstly, we have developed a novel transformation framework for the robust optimization (RO) method to be used as a generalized approach to overcome the drawback of conventional RO approach owing to its difficulty in obtaining information regarding numerous control variable terms as well as added extra variables and constraints into the model in transforming deterministic models into the robust form. We have determined an optimal case mix planning for a given set of specialties for a single operating room department using the proposed standard RO framework. In this case-mix planning problem, demands for elective and emergency surgery are considered to be random variables realized over a set of probabilistic scenarios. A deterministic and a two-stage stochastic recourse programming model is also developed for the uncertain surgery case mix planning to demonstrate the applicability of the proposed RO models. The objective is to minimize the expected total loss incurred due to postponed and unmet demand as well as the underutilization costs. We have shown that the optimum solution can be found in polynomial time. Secondly, the tactical and operational level decision of OR block scheduling and advance scheduling problems are considered simultaneously to overcome the drawback of current literature in addressing these problems in isolation. We have focused on a hybrid master surgery scheduling (MSS) and surgical case assignment (SCA) problem under the assumption that both surgery durations and emergency arrivals follow probability distributions defined over a discrete set of scenarios. We have developed an integrated robust MSS and SCA model using the proposed standard transformation framework and determined the allocation of surgical specialties to the ORs as well as the assignment of surgeries within each specialty to the corresponding ORs in a coordinated way to minimize the costs associated with patients waiting time and hospital resource utilization. To demonstrate the usefulness and applicability of the two proposed models, a simulation study is carried utilizing data provided by Windsor Regional Hospital (WRH). The simulation results demonstrate that the two proposed models can mitigate the existing variability in parameter uncertainty. This provides a more reliable decision tool for the OR managers while limiting the negative impact of waiting time to the patients as well as welfare loss to the hospital

    From Cost Sharing Mechanisms to Online Selection Problems

    Get PDF
    We consider a general class of online optimization problems, called online selection problems, where customers arrive sequentially, and one has to decide upon arrival whether to accept or reject each customer. If a customer is rejected, then a rejection cost is incurred. The accepted customers are served with minimum possible cost, either online or after all customers have arrived. The goal is to minimize the total production costs for the accepted customers plus the rejection costs for the rejected customers. These selection problems are related to online variants of offline prize collecting combinatorial optimization problems that have been widely studied in the computer science literature. In this paper, we provide a general framework to develop online algorithms for this class of selection problems. In essence, the algorithmic framework leverages any cost sharing mechanism with certain properties into a poly-logarithmic competitive online algorithm for the respective problem; the competitive ratios are shown to be near-optimal. We believe that the general and transparent connection we establish between cost sharing mechanisms and online algorithms could lead to additional online algorithms for problems beyond the ones studied in this paper.National Science Foundation (U.S.) (CAREER Award CMMI-0846554)United States. Air Force Office of Scientific Research (FA9550-11-1-0150)United States. Air Force Office of Scientific Research (FA9550-08-1-0369)Solomon Buchsbaum AT&T Research Fun

    Efficient sharing mechanisms for virtualized multi-tenant heterogeneous networks

    Get PDF
    The explosion in data traffic, the physical resource constraints, and the insufficient financial incentives for deploying 5G networks, stress the need for a paradigm shift in network upgrades. Typically, operators are also the service providers, which charge the end users with low and flat tariffs, independently of the service enjoyed. A fine-scale management of the network resources is needed, both for optimizing costs and resource utilization, as well as for enabling new synergies among network owners and third-parties. In particular, operators could open their networks to third parties by means of fine-scale sharing agreements over customized networks for enhanced service provision, in exchange for an adequate return of investment for upgrading their infrastructures. The main objective of this thesis is to study the potential of fine-scale resource management and sharing mechanisms for enhancing service provision and for contributing to a sustainable road to 5G. More precisely, the state-of-the-art architectures and technologies for network programmability and scalability are studied, together with a novel paradigm for supporting service diversity and fine-scale sharing. We review the limits of conventional networks, we extend existing standardization efforts and define an enhanced architecture for enabling 5G networks' features (e.g., network-wide centralization and programmability). The potential of the proposed architecture is assessed in terms of flexible sharing and enhanced service provision, while the advantages of alternative business models are studied in terms of additional profits to the operators. We first study the data rate improvement achievable by means of spectrum and infrastructure sharing among operators and evaluate the profit increase justified by a better service provided. We present a scheme based on coalitional game theory for assessing the capability of accommodating more service requests when a cooperative approach is adopted, and for studying the conditions for beneficial sharing among coalitions of operators. Results show that: i) collaboration can be beneficial also in case of unbalanced cost redistribution within coalitions; ii) coalitions of equal-sized operators provide better profit opportunities and require lower tariffs. The second kind of sharing interaction that we consider is the one between operators and third-party service providers, in the form of fine-scale provision of customized portions of the network resources. We define a policy-based admission control mechanism, whose performance is compared with reference strategies. The proposed mechanism is based on auction theory and computes the optimal admission policy at a reduced complexity for different traffic loads and allocation frequencies. Because next-generation services include delay-critical services, we compare the admission control performances of conventional approaches with the proposed one, which proves to offer near real-time service provision and reduced complexity. Besides, it guarantees high revenues and low expenditures in exchange for negligible losses in terms of fairness towards service providers. To conclude, we study the case where adaptable timescales are adopted for the policy-based admission control, in order to promptly guarantee service requirements over traffic fluctuations. In order to reduce complexity, we consider the offline pre­computation of admission strategies with respect to reference network conditions, then we study the extension to unexplored conditions by means of computationally efficient methodologies. Performance is compared for different admission strategies by means of a proof of concept on real network traces. Results show that the proposed strategy provides a tradeoff in complexity and performance with respect to reference strategies, while reducing resource utilization and requirements on network awareness.La explosion del trafico de datos, los recursos limitados y la falta de incentivos para el desarrollo de 5G evidencian la necesidad de un cambio de paradigma en la gestion de las redes actuales. Los operadores de red suelen ser tambien proveedores de servicios, cobrando tarifas bajas y planas, independientemente del servicio ofrecido. Se necesita una gestion de recursos precisa para optimizar su utilizacion, y para permitir nuevas sinergias entre operadores y proveedores de servicios. Concretamente, los operadores podrian abrir sus redes a terceros compartiendolas de forma flexible y personalizada para mejorar la calidad de servicio a cambio de aumentar sus ganancias como incentivo para mejorar sus infraestructuras. El objetivo principal de esta tesis es estudiar el potencial de los mecanismos de gestion y comparticion de recursos a pequei\a escala para trazar un camino sostenible hacia el 5G. En concreto, se estudian las arquitecturas y tecnolog fas mas avanzadas de "programabilidad" y escalabilidad de las redes, junto a un nuevo paradigma para la diversificacion de servicios y la comparticion de recursos. Revisamos los limites de las redes convencionales, ampliamos los esfuerzos de estandarizacion existentes y definimos una arquitectura para habilitar la centralizacion y la programabilidad en toda la red. La arquitectura propuesta se evalua en terminos de flexibilidad en la comparticion de recursos, y de mejora en la prestacion de servicios, mientras que las ventajas de un modelo de negocio alternativo se estudian en terminos de ganancia para los operadores. En primer lugar, estudiamos el aumento en la tasa de datos gracias a un uso compartido del espectro y de las infraestructuras, y evaluamos la mejora en las ganancias de los operadores. Presentamos un esquema de admision basado en la teoria de juegos para acomodar mas solicitudes de servicio cuando se adopta un enfoque cooperativo, y para estudiar las condiciones para que la reparticion de recursos sea conveniente entre coaliciones de operadores. Los resultados ensei\an que: i) la colaboracion puede ser favorable tambien en caso de una redistribucion desigual de los costes en cada coalicion; ii) las coaliciones de operadores de igual tamai\o ofrecen mejores ganancias y requieren tarifas mas bajas. El segundo tipo de comparticion que consideramos se da entre operadores de red y proveedores de servicios, en forma de provision de recursos personalizada ya pequei\a escala. Definimos un mecanismo de control de trafico basado en polfticas de admision, cuyo rendimiento se compara con estrategias de referencia. El mecanismo propuesto se basa en la teoria de subastas y calcula la politica de admision optima con una complejidad reducida para diferentes cargas de trafico y tasa de asignacion. Con particular atencion a servicios 5G de baja latencia, comparamos las prestaciones de estrategias convencionales para el control de admision con las del metodo propuesto, que proporciona: i) un suministro de servicios casi en tiempo real; ii) una complejidad reducida; iii) unos ingresos elevados; y iv) unos gastos reducidos, a cambio de unas perdidas insignificantes en terminos de imparcialidad hacia los proveedores de servicios. Para concluir, estudiamos el caso en el que se adoptan escalas de tiempo adaptables para el control de admision, con el fin de garantizar puntualmente los requisitos de servicio bajo diferentes condiciones de trafico. Para reducir la complejidad, consideramos el calculo previo de las estrategias de admision con respecto a condiciones de red de referenda, adaptables a condiciones inexploradas por medio de metodologias computacionalmente eficientes. Se compara el rendimiento de diferentes estrategias de admision sobre trazas de trafico real. Los resultados muestran que la estrategia propuesta equilibra complejidad y ganancias, mientras se reduce la utilizacion de recursos y la necesidad de conocer el estado exacto de la red.Postprint (published version

    Behavioral Calibration and Analysis of a Large-Scale Travel Microsimulation

    Get PDF
    This article reports on the calibration and analysis of a fully disaggregate (agent-based) transport simulation for the metropolitan area of Zurich. The agent-based simulation goes beyond traditional transport models in that it equilibrates not only route choice but all-day travel behavior, including departure time choice and mode choice. Previous work has shown that the application of a novel calibration technique that adjusts all choice dimensions at once from traffic counts yields cross-validation results that are competitive with any state-of-the-art four-step model. While the previous study aims at a methodological illustration of the calibration method, this work focuses on the real-world scenario, and it elaborates on the usefulness of the obtained results for further demand analysis purpose
    • …
    corecore