18 research outputs found

    User-activity aware strategies for mobile information access

    Get PDF
    Information access suffers tremendously in wireless networks because of the low correlation between content transferred across low-bandwidth wireless links and actual data used to serve user requests. As a result, conventional content access mechanisms face such problems as unnecessary bandwidth consumption and large response times, and users experience significant performance degradation. In this dissertation, we analyze the cause of those problems and find that the major reason for inefficient information access in wireless networks is the absence of any user-activity awareness in current mechanisms. To solve these problems, we propose three user-activity aware strategies for mobile information access. Through simulations and implementations, we show that our strategies can outperform conventional information access schemes in terms of bandwidth consumption and user-perceived response times.Ph.D.Committee Chair: Raghupathy Sivakumar; Committee Member: Chuanyi Ji; Committee Member: George Riley; Committee Member: Magnus Egerstedt; Committee Member: Umakishore Ramachandra

    Contention management for distributed data replication

    Get PDF
    PhD ThesisOptimistic replication schemes provide distributed applications with access to shared data at lower latencies and greater availability. This is achieved by allowing clients to replicate shared data and execute actions locally. A consequence of this scheme raises issues regarding shared data consistency. Sometimes an action executed by a client may result in shared data that may conflict and, as a consequence, may conflict with subsequent actions that are caused by the conflicting action. This requires a client to rollback to the action that caused the conflicting data, and to execute some exception handling. This can be achieved by relying on the application layer to either ignore or handle shared data inconsistencies when they are discovered during the reconciliation phase of an optimistic protocol. Inconsistency of shared data has an impact on the causality relationship across client actions. In protocol design, it is desirable to preserve the property of causality between different actions occurring across a distributed application. Without application level knowledge, we assume an action causes all the subsequent actions at the same client. With application knowledge, we can significantly ease the protocol burden of provisioning causal ordering, as we can identify which actions do not cause other actions (even if they precede them). This, in turn, makes possible the client’s ability to rollback to past actions and to change them, without having to alter subsequent actions. Unfortunately, increased instances of application level causal relations between actions lead to a significant overhead in protocol. Therefore, minimizing the rollback associated with conflicting actions, while preserving causality, is seen as desirable for lower exception handling in the application layer. In this thesis, we present a framework that utilizes causality to create a scheduler that can inform a contention management scheme to reduce the rollback associated with the conflicting access of shared data. Our framework uses a backoff contention management scheme to provide causality preserving for those optimistic replication systems with high causality requirements, without the need for application layer knowledge. We present experiments which demonstrate that our framework reduces clients’ rollback and, more importantly, that the overall throughput of the system is improved when the contention management is used with applications that require causality to be preserved across all actions

    Actas da 10ª Conferência sobre Redes de Computadores

    Get PDF
    Universidade do MinhoCCTCCentro AlgoritmiCisco SystemsIEEE Portugal Sectio

    Public policy modeling and applications

    Full text link

    Towards self-optimizing frameworks for collaborative systems

    Get PDF
    Two important performance metrics in collaborative systems are local and remote response times. For certain classes of applications, it is possible to meet response time requirements better than existing systems through a new system without requiring hardware, network, or user-interface changes. This self-optimizing system improves response times by automatically making runtime adjustments to three aspects of a collaborative application. One of these aspects is the collaboration architecture. Previous work has shown that dynamically switching architectures at runtime can improve response times; however, no previous work performs the switch automatically. The thesis shows that (a) another important performance parameter is whether multicast or unicast is used to transmit commands, and (b) response times can be noticeably better with multicast than with unicast when transmission costs are high. Traditional architectures, however, support only unicast - a computer that processes input commands must also transmit commands to all other computers. To support multicast, a new bi-architecture model of collaborative systems is introduced in which two separate architectures govern the processing and transmission tasks that each computer must perform. The thesis also shows that another important performance aspect is the order in which a computer performs these tasks. These tasks can be scheduled sequentially or concurrently on a single-core, or in parallel on multiple cores. As the thesis shows, existing single-core policies trade-off noticeable improvements in local (remote) for noticeable degradations in remote (local) response times. A new lazy policy for scheduling these tasks on a single-core is introduced that trades-off an unnoticeable degradation in performance of some users for a much larger noticeable improvement in performance of others. The thesis also shows that on multi-core devices, the tasks should always be scheduled on separate cores. The self-optimizing system adjusts the processing architecture, communication architecture, and scheduling policy based on response time predictions given by a new analytical model. Both the analytical model and the self-optimizing system are validated through simulations and experiments in practical scenarios

    Decentralized data fusion and data harvesting framework for heterogeneous dynamic network systems

    Get PDF
    Diese Dissertation behandelt das Thema der dezentralisieren Sammlung und Fusion von Daten in heterogenen, ressourcenbeschraekten und dynamischen Netzwerkszenarien. Dazu wird ein generisches Framework vorgestellt, dass es erlaubt die Datensammlung, den Datenaustausch und auch die Datenfusion dynamisch zu konfigurieren. Im Zuge dessen wird auch eine Methode zur gerichteten Fusion von Daten auf graphentheoretischer Basis eingefrt, die es erlaubt eine logische Struktur fuer die Fusion von Informationen zu modellieren. Eine Markup-Sprache, die sowohl menschen- als auch maschinenlesbar ist, erlaubt es diese Struktur leicht zu editieren. Im Bereich der Protokolle zum Datenaustausch liegt der Fokus dieser Arbeit auf Energieeffizienz, um auch ressourcenbeschraenkte Geraete einzubinden. Ein weiterer Schwerpunkt liegt auf Robustheit fuer die betrachteten dynamischen Szenarien. Diese Dissertation schlaet zudem Design-Richtlinien vor, um verschiedene Ziele fuer unterschiedliche Applikationen umzusetzen. Diese lassen sich leicht in das vorgestellte Framework integrieren und darueber konfigurieren. Dadurch ergibt sich im Ganzen eine flexible Architektur, die sich leicht an dynamische Umgebungen anpassen laesst.With the increasing number of available smart phones, sensor nodes, and novel mobile smart devices such as Google glass, a large volume of data reflecting the environment is generated in the form of sensing data sources (such as GPS, received signal strength identification, accelerometer, microphone, images, videos and gyroscope, etc.). Some context-aware and data centric applications require the online processing of the data collected. The thesis researches on the decentralized data fusion and data harvesting framework for heterogeneous dynamic network system consisting of various devices with resource constraints. In order to achieve the flexible design, a general architecture is provided while the detailed data fusion and data exchange functions can be dynamically configured. A novel method to use directed fusion graph to model the logical structure of the distributed information fusion architecture is introduced. This directed fusion graph can accurately portray the interconnection among different data fusion components and the data exchange protocols, as well as the detailed data streams. The directed fusion graph is then transformed into a format with marked language, so that both human and machine can easily understand and edit. In the field of data exchange protocols, this thesis targets energy-efficiency considering the resource constraints of the devices and robustness, as the dynamic environment might cause failures to the system. It proposes a refined gossip strategy to reduce retransmission of redundant data. The thesis also suggests a design guideline to achieve different design aims for different applications. These results in this field can be integrated into the framework effortlessly. The configuration mechanism is another feature of this framework. Different from other research work which consider configuration as a post-design work separated from the main design of any middle-ware. This thesis considers the configuration part as another dimension of the framework. The whole strategy in configuration sets up the foundation for the flexible architecture, and makes it easy to adapt to the dynamic environment. The contributions in the above fields lead to a light-weight data fusion and data harvesting framework which can be deployed easily above wireless based, heterogeneous, dynamic network systems, even in extreme conditions, to handle data-centric applications

    Workload Modeling for Computer Systems Performance Evaluation

    Full text link

    Systematic Approaches for Telemedicine and Data Coordination for COVID-19 in Baja California, Mexico

    Get PDF
    Conference proceedings info: ICICT 2023: 2023 The 6th International Conference on Information and Computer Technologies Raleigh, HI, United States, March 24-26, 2023 Pages 529-542We provide a model for systematic implementation of telemedicine within a large evaluation center for COVID-19 in the area of Baja California, Mexico. Our model is based on human-centric design factors and cross disciplinary collaborations for scalable data-driven enablement of smartphone, cellular, and video Teleconsul-tation technologies to link hospitals, clinics, and emergency medical services for point-of-care assessments of COVID testing, and for subsequent treatment and quar-antine decisions. A multidisciplinary team was rapidly created, in cooperation with different institutions, including: the Autonomous University of Baja California, the Ministry of Health, the Command, Communication and Computer Control Center of the Ministry of the State of Baja California (C4), Colleges of Medicine, and the College of Psychologists. Our objective is to provide information to the public and to evaluate COVID-19 in real time and to track, regional, municipal, and state-wide data in real time that informs supply chains and resource allocation with the anticipation of a surge in COVID-19 cases. RESUMEN Proporcionamos un modelo para la implementación sistemática de la telemedicina dentro de un gran centro de evaluación de COVID-19 en el área de Baja California, México. Nuestro modelo se basa en factores de diseño centrados en el ser humano y colaboraciones interdisciplinarias para la habilitación escalable basada en datos de tecnologías de teleconsulta de teléfonos inteligentes, celulares y video para vincular hospitales, clínicas y servicios médicos de emergencia para evaluaciones de COVID en el punto de atención. pruebas, y para el tratamiento posterior y decisiones de cuarentena. Rápidamente se creó un equipo multidisciplinario, en cooperación con diferentes instituciones, entre ellas: la Universidad Autónoma de Baja California, la Secretaría de Salud, el Centro de Comando, Comunicaciones y Control Informático. de la Secretaría del Estado de Baja California (C4), Facultades de Medicina y Colegio de Psicólogos. Nuestro objetivo es proporcionar información al público y evaluar COVID-19 en tiempo real y rastrear datos regionales, municipales y estatales en tiempo real que informan las cadenas de suministro y la asignación de recursos con la anticipación de un aumento de COVID-19. 19 casos.ICICT 2023: 2023 The 6th International Conference on Information and Computer Technologieshttps://doi.org/10.1007/978-981-99-3236-
    corecore