28 research outputs found

    Flexible architecture for the future internet scalability of SDN control plane

    Get PDF
    Software-Defined Networking (SDN) separates the control plane from the data plane. The initial SDN approach involves a single centralized controller, which may not scale properly as a network grows in size. Distributed controllers have emerged to address the disadvantages of a single centralized controller. The control architecture needs to be distributed with traffic control between switches and controllers and among the controllers in order to allow SDNs for several thousand switches. One of the most significant research challenges for distributed controller architectures is to effectively manage controllers, which includes allocating enough controllers to appropriate network locations. To address these daunting issues, we make the following major contributions: This thesis expands the method of solving the Control Placement Problem (CPP) based on the K-means and K-center algorithms to include a Hierarchical Controller Placement Problem (HCPP), located at a high level of Super Controller (SC), a middle level of Master Controllers (MCs), and the lowest level of domain controllers (DCs). The optimization metric addresses latency between the controller and the switches assigned to it.. The proposed architecture and methodology are implemented using the topology of Western European NRENs from the Internet Topology Zoo. The entire network topology is divided into clusters, and the optimal number of controllers (DCs) and their placement are determined for each cluster. MC placement optimization determines the optimal number of MCs and their optimal placement. As a second contribution, an accumulated latency is defined to solve CPP, which takes into account both the latency between the controller and its associated switches and the latency between controllers. Under the constraint of latency, an optimization problem is formulated as per mixed-integer linear programming (MILP). The goal of the research is to reduce accumulated latency while also reducing the number of network controllers and optimizing their placement to achieve an optimal balance. The performance of the developed method is evaluated on Internet2 OS3E real network topology. To achieve the third objective, a metric was developed that includes reliability. The communication latency between controllers should also be considered because a low controller-switch delay does not always imply a short controller-controller delay for a particular controller placement. As the third contribution, we propose a novel metric for CPP to improve the reliability of controllers that takes into account both communication latency and communication reliability between switches and controllers, as well as between controllers. When a single link fails, reliability is taken into account. This aspect concluded by identifying the optimal controller placement to achieve low latencies in control plane traffic. The goal of this project is to reduce the average latency. As the fourth contribution, this study evaluates the Joint Latency and Reliability-aware Controller Placement (LRCP) optimization model. As the evaluation metric, control plane latency (CPL) is defined as the sum of the average switch-to-controller latency and average inter-controller latency. The latency of the control plane, utilizing the actual latencies of the real network topology, is calculated for every optimum placement in the network. In the case of a failure of the single link, the actual CPL for LRCP placements is calculated and evaluated to determine how good LRCP placements are. CPL metrics are used to compare latency and reliability metrics with other models. This study provides proof that the developed methodologies for large-scale networks are highly powerful in terms of searching for all feasible controller placements while assessing the outcomes. In addition, compared to previous work including latency among controllers and reliability for an event of single-link failure.La xarxa definida per programari (SDN) separa el pla de control del pla de dades. L’enfocament SDN inicial implica un únic controlador centralitzat, que pot no escalar correctament a mesura que la xarxa creixi de mida. Els controladors distribuïts han sorgit per abordar els inconvenients d’un únic controlador centralitzat. . Un dels reptes de recerca més importants per a les arquitectures de controladors distribuïts és gestionar de manera eficaç els controladors, que inclou l’assignació de controladors suficients a les ubicacions de xarxa adequades. Per abordar aquests problemes, fem les següents contribucions. Aquesta tesi amplia el mètode de resolució del Problema de Col·locació de Control (CPP) basat en els algorismes de K-means i K-center per incloure un Problema de Col·locació de Controladors Jeràrquics (HCPP), situat a un nivell alt de Super Controller (SC), un nivell de controladors mestres (MC) i el nivell més baix de controladors de domini (DC). La mètrica d’optimització és la latència entre el controlador i els commutadors assignats a aquest. L’arquitectura i la metodologia proposades s’implementen utilitzant la topologia de NREN d’Europa occidental de l’Internet Topology Zoo. La topologia de la xarxa es divideix en clústers i es determina el nombre òptim de controladors de domini (DC) i la seva ubicació per a cada clúster. L’optimització de la ubicació de MC determina el nombre òptim de MC i la seva col·locació òptima. Com a segona contribució, es defineix una latència acumulada per resoldre el CPP, que té en compte tant la latència entre el controlador i els seus commutadors associats com la latència entre controladors. Sota la restricció de la latència, es formula un problema d’optimització segons la programació lineal de nombres enters mixts (MILP). L’objectiu de la investigació és reduir la latència acumulada alhora que es redueix el nombre de controladors de xarxa i optimitza la seva col·locació per aconseguir un equilibri òptim. El rendiment del mètode desenvolupat s’avalua en la topologia de xarxa real d’Internet2 OS3E. Per aconseguir el tercer objectiu, es va desenvolupar una mètrica que inclou la fiabilitat. També s’ha de tenir en compte la latència de comunicació entre controladors perquè un retard baix entre el commutador i el controlador no sempre implica un retard curt del controladorcontrolador per a una ubicació concreta dels controladors. Com a tercera contribució, proposem una nova mètrica per al CPP per millorar la fiabilitat dels controladors que tingui en compte tant la latència de la comunicació com la fiabilitat de la comunicació entre commutadors i controladors, així com entre controladors. La fiabilitat es té en compte quan falla un únic enllaç identificant la col·locació òptima dels controladors per aconseguir baixes latències en el trànsit del pla de control. L’objectiu d’aquest projecte és reduir la latència mitjana. Com a quarta contribució, aquest estudi avalua el model d’optimització Joint Latency and Reliability-aware Controller Placement (LRCP). Com a mètrica d’avaluació, la latència del pla de control (CPL) es defineix com la suma de la latència mitjana de commutador a controlador i la latència mitjana entre controladors. La latència del pla de control, utilitzant les latències reals de la topologia de xarxa real, es calcula per a cada col·locació òptima a la xarxa. En el cas d’una fallida en un únicenllaç, es calcula i s’avalua el CPL real de les ubicacions LRCP per determinar com de bones són les ubicacions LRCP. Les mètriques CPL s’utilitzen per comparar les mètriques de latència i fiabilitat amb altres models. Aquest estudi proporciona la prova que les metodologies desenvolupades per a xarxes a gran escala són molt potents pel que fa a la recerca de totes les ubicacions de controladors factibles mentre s’avaluen els resultats. A més, en comparació amb el treball anterior, inclou la latència entre els controladors i la fiabilitat per a un esdeveniment de fallada d’un enllaç únic.Las redes definidas por software (SDN) separan el plano de control del plano de datos. El enfoque inicial de SDN implica un único controlador centralizado, que puede no escalar adecuadamente a medida que una red crece en tamaño. Los controladores distribuidos han surgido para abordar las desventajas de un único controlador centralizado. Uno de los retos de investigación más importantes para las arquitecturas de controladores distribuidos es la gestión eficaz de los controladores, que incluye la asignación de suficientes controladores en las ubicaciones adecuadas. Para hacer frente a estos problemas, realizamos las siguientes contribuciones principales: Esta tesis amplía el método de resolución del Problema de Colocación de Controles (CPP) basado en los algoritmos K-means y K-center para incluir un Problema de Colocación de Controladores Jerárquicos (HCPP), situado en un nivel alto de Super-controladores (SC), un nivel medio de Controladores Maestros (MC), y el nivel más bajo de controladores de dominio (DC). La métrica de optimización es la latencia entre el controlador y los conmutadores asignados al mismo. . La arquitectura y la metodología propuestas se implementan utilizando la topología de las NREN de Europa Occidental del TopologyZoo. La topología completa de la red se divide en clústeres, y se determina el número óptimo de controladores de dominio (CD) y su colocación para cada clúster. La optimización de la colocación de los MC determina el número óptimo de MC y su colocación óptima. Como segunda contribución, se define una latencia acumulada para resolver el CPP, que tiene en cuenta tanto la latencia entre el controlador y sus conmutadores asociados como la latencia entre los controladores. Bajo la restricción de la latencia, se formula un problema de optimización según la programación lineal de enteros mixtos (MILP). El objetivo es reducir la latencia acumulada al tiempo que se reduce el número de controladores de la red y se optimiza su ubicación para lograr un equilibrio óptimo. El rendimiento del método desarrollado se evalúa en la topología de Internet2 OS3E. Para lograr el tercer objetivo, se desarrolló una métrica que incluye la fiabilidad. La latencia de la comunicación entre controladores también debe tenerse en cuenta, ya que un bajo retardo entre controladores y conmutadores no siempre implica un corto retardo entre controladores para una determinada ubicación de los mismos. Como tercera contribución proponemos una nueva métrica para el CPP para mejorar la fiabilidad de los controladores que tiene en cuenta tanto la latencia de la comunicación como la fiabilidad de la comunicación entre los conmutadores y los controladores, así como entre los controladores. Se tiene en cuenta la fiabilidad cuando falla un solo enlace. Este aspecto concluye con la identificación de la ubicación óptima de los controladores para lograr bajas latencias en el tráfico del plano de control. El objetivo es reducir la latencia media. Como cuarta contribución, este estudio evalúa el modelo de optimización Joint Latency and Reliability-aware Controller Placement (LRCP). Como métrica de evaluación, la latencia del plano de control (CPL) se define como la suma de la latencia media entre conmutadores y controladores y la latencia media entre controladores. La latencia del plano de control, utilizando las latencias reales de la topología de la red, se calcula para cada ubicación óptima en la red. En el caso de un fallo de un enlace, se calcula y evalúa la CPL real para las colocaciones de LRCP con el fin de determinar lo buenas que son las colocaciones de LRCP. Las métricas CPL se utilizan para comparar las métricas de latencia y fiabilidad con otros modelos. Este estudio demuestra que las metodologías desarrolladas para redes a gran escala son muy potentes en cuanto a la búsqueda de todas las ubicaciones factibles de los controladores mientras se evalúan los resultados. Además, en comparación con los trabajos anteriores, que incluyen la latencia entre controladores y la fiabilidad para un caso de fallo de un solo enlacePostprint (published version

    A Survey of Machine Learning Techniques for Video Quality Prediction from Quality of Delivery Metrics

    Get PDF
    A growing number of video streaming networks are incorporating machine learning (ML) applications. The growth of video streaming services places enormous pressure on network and video content providers who need to proactively maintain high levels of video quality. ML has been applied to predict the quality of video streams. Quality of delivery (QoD) measurements, which capture the end-to-end performances of network services, have been leveraged in video quality prediction. The drive for end-to-end encryption, for privacy and digital rights management, has brought about a lack of visibility for operators who desire insights from video quality metrics. In response, numerous solutions have been proposed to tackle the challenge of video quality prediction from QoD-derived metrics. This survey provides a review of studies that focus on ML techniques for predicting the QoD metrics in video streaming services. In the context of video quality measurements, we focus on QoD metrics, which are not tied to a particular type of video streaming service. Unlike previous reviews in the area, this contribution considers papers published between 2016 and 2021. Approaches for predicting QoD for video are grouped under the following headings: (1) video quality prediction under QoD impairments, (2) prediction of video quality from encrypted video streaming traffic, (3) predicting the video quality in HAS applications, (4) predicting the video quality in SDN applications, (5) predicting the video quality in wireless settings, and (6) predicting the video quality in WebRTC applications. Throughout the survey, some research challenges and directions in this area are discussed, including (1) machine learning over deep learning; (2) adaptive deep learning for improved video delivery; (3) computational cost and interpretability; (4) self-healing networks and failure recovery. The survey findings reveal that traditional ML algorithms are the most widely adopted models for solving video quality prediction problems. This family of algorithms has a lot of potential because they are well understood, easy to deploy, and have lower computational requirements than deep learning techniques

    PROFILING - CONCEPTS AND APPLICATIONS

    Get PDF
    Profiling is an approach to put a label or a set of labels on a subject, considering the characteristics of this subject. The New Oxford American Dictionary defines profiling as: “recording and analysis of a person’s psychological and behavioral characteristics, so as to assess or predict his/her capabilities in a certain sphere or to assist in identifying a particular subgroup of people”. This research extends this definition towards things demonstrating that many methods used for profiling of people may be applied for a different type of subjects, namely things. The goal of this research concerns proposing methods for discovery of profiles of users and things with application of Data Science methods. The profiles are utilized in vertical and 2 horizontal scenarios and concern such domains as smart grid and telecommunication (vertical scenarios), and support provided both for the needs of authorization and personalization (horizontal usage).:The thesis consists of eight chapters including an introduction and a summary. First chapter describes motivation for work that was carried out for the last 8 years together with discussion on its importance both for research and business practice. The motivation for this work is much broader and emerges also from business importance of profiling and personalization. The introduction summarizes major research directions, provides research questions, goals and supplementary objectives addressed in the thesis. Research methodology is also described, showing impact of methodological aspects on the work undertaken. Chapter 2 provides introduction to the notion of profiling. The definition of profiling is introduced. Here, also a relation of a user profile to an identity is discussed. The papers included in this chapter show not only how broadly a profile may be understood, but also how a profile may be constructed considering different data sources. Profiling methods are introduced in Chapter 3. This chapter refers to the notion of a profile developed using the BFI-44 personality test and outcomes of a survey related to color preferences of people with a specific personality. Moreover, insights into profiling of relations between people are provided, with a focus on quality of a relation emerging from contacts between two entities. Chapters from 4 to 7 present different scenarios that benefit from application of profiling methods. Chapter 4 starts with introducing the notion of a public utility company that in the thesis is discussed using examples from smart grid and telecommunication. Then, in chapter 4 follows a description of research results regarding profiling for the smart grid, focusing on a profile of a prosumer and forecasting demand and production of the electric energy in the smart grid what can be influenced e.g. by weather or profiles of appliances. Chapter 5 presents application of profiling techniques in the field of telecommunication. Besides presenting profiling methods based on telecommunication data, in particular on Call Detail Records, also scenarios and issues related to privacy and trust are addressed. Chapter 6 and Chapter 7 target at horizontal applications of profiling that may be of benefit for multiple domains. Chapter 6 concerns profiling for authentication using un-typical data sources such as Call Detail Records or data from a mobile phone describing the user behavior. Besides proposing methods, also limitations are discussed. In addition, as a side research effect a methodology for evaluation of authentication methods is proposed. Chapter 7 concerns personalization and consists of two diverse parts. Firstly, behavioral profiles to change interface and behavior of the system are proposed and applied. The performance of solutions personalizing content either locally or on the server is studied. Then, profiles of customers of shopping centers are created based on paths identified using Call Detail Records. The analysis demonstrates that the data that is collected for one purpose, may significantly influence other business scenarios. Chapter 8 summarizes the research results achieved by the author of this document. It presents contribution over state of the art as well as some insights into the future work planned

    Network-Based Management for Optimising Video Delivery

    Get PDF
    The past decade has witnessed a massive increase in Internet video traffic. The Cisco Visual Forecast index indicates that, by 2022, (79%) of the world's mobile data traffic will be video traffic. In order to increase the video streaming market revenue, service providers need to provide services to the end-users characterised by high Quality of Experience (QoE). However, delivering good-quality video services is a very challenging task due to the stringent constraints related to bandwidth and latency requirements in video streaming. Among the available streaming services, HTTP adaptive streaming (HAS) has become the de facto standard for multimedia delivery over the Internet. HAS is a pull-based approach, since the video player at the client side is responsible for adapting the requested video based on the estimated network conditions. Furthermore, HAS can traverse any firewall or proxy server that lets through standard HTTP data traffic over content delivery networks. Despite the great benefits HAS solutions bring, they also face challenges relating to quality fluctuations when they compete for a shared link. To overcome these issues, the network and video providers must exchange information and cooperate. In this context, Software Defined Networking (SDN) is an emerging technology used to deploy such architecture by providing centralised control for efficient and flexible network management. One of the first problems addressed in this thesis is that of providing QoE-level fairness for the competing HAS players and efficient resource allocation for the available network resources. This has been achieved by presenting a dynamic programming-based algorithm, based on the concept of Max-Min fairness, to provide QoE-level fairness among the competing HAS players. In order to deploy the proposed algorithm, an SDN-based architecture has been presented, named BBGDASH, that leverages the flexibility of the SDN’s management and control to deploy the proposed algorithm on the application and the network level. Another question answered by this thesis is that of how the proposed guidance approach maintains a balance between stability and scalability. To answer this question, a scalable guidance mechanism has been presented that provides guidance to the client without moving the entire control logic to an additional entity or relying purely on the client-side decision. To do so, the guidance scheme provides each client with the optimal bitrate levels to adapt the requested bitrate within the provided levels. Although the proposed BGGDASH can improve the QoE within a wired network, deploying it in a wireless network environment could result in sub-optimal decisions being made due to the high level of fluctuations in the wireless environment. In order to cope with this issue, two time series-based forecasting approaches have been presented to identify the optimal set of bitrate levels for each client based on the network conditions. Additionally, the implementation of the BBGDASH architecture has been extended by proposing an intelligent streaming architecture (named BBGDASH+). Finally, in order to evaluate the feasibility of deploying the bounding bitrate guidance with different network conditions, it has been evaluated under different network conditions to provide generic evaluations. The results show that the proposed algorithms can significantly improve the end-users QoE compared to other compared approaches

    Progressive introduction of network softwarization in operational telecom networks: advances at architectural, service and transport levels

    Get PDF
    Technological paradigms such as Software Defined Networking, Network Function Virtualization and Network Slicing are altogether offering new ways of providing services. This process is widely known as Network Softwarization, where traditional operational networks adopt capabilities and mechanisms inherit form the computing world, such as programmability, virtualization and multi-tenancy. This adoption brings a number of challenges, both from the technological and operational perspectives. On the other hand, they provide an unprecedented flexibility opening opportunities to developing new services and new ways of exploiting and consuming telecom networks. This Thesis first overviews the implications of the progressive introduction of network softwarization in operational networks for later on detail some advances at different levels, namely architectural, service and transport levels. It is done through specific exemplary use cases and evolution scenarios, with the goal of illustrating both new possibilities and existing gaps for the ongoing transition towards an advanced future mode of operation. This is performed from the perspective of a telecom operator, paying special attention on how to integrate all these paradigms into operational networks for assisting on their evolution targeting new, more sophisticated service demands.Programa de Doctorado en Ingeniería Telemática por la Universidad Carlos III de MadridPresidente: Eduardo Juan Jacob Taquet.- Secretario: Francisco Valera Pintor.- Vocal: Jorge López Vizcaín

    Exploring intelligent service migration in a highly mobile network

    Get PDF
    Mobile services allow services to be migrated or replicated closer to users as they move around. This is now regarded as a viable mechanism to provide good Quality of Service to users in highly mobile environments such as vehicular networks. The vehicular environment is rapidly becoming a significant part of the internet and this presents various challenges that must be addressed; this is due to continuous handovers as mobile devices change their point of attachment to these networks resulting in a loss of service. Therefore, this explains the need to build a framework for intelligent service migration. This thesis addresses these issues. It starts by discussing the requirements for intelligent service migration. Then it investigates a low latency Quality of Service Aware Framework as well as an experimental transport protocol that would be favoured by vehicular networks. Furthermore, two analytical models are developed using the Zero-Server Markov Chain technique which is a way of analysing scenarios when the server is not continuously available to serve. Using the Zero-Server Markov Chain, the first analytical model looks at lost service due to continuous handovers and the communication dynamics of vehicular networks, while the second model analyses how service migration affects service delivery in these networks. Formulas are developed to yield the average number of packets in the system, the response time, the probability of blocking and a new parameter called the probability of lost service. These formulas are then applied to the Middlesex VANET Testbed to look at reactive and proactive service migration. These techniques are then incorporated into a new Service Management Framework to provide sustainable Quality of Service and Quality of Experience to mobile users in vehicular networks. This thesis also shows that this new approach is better than current approaches as it addresses key issues in intelligent service migration in such environments, and hence can play a significant part in the development of Intelligent Transport Systems for Smart Cities

    Estimation of the QoE for video streaming services based on facial expressions and gaze direction

    Get PDF
    As the multimedia technologies evolve, the need to control their quality becomes even more important making the Quality of Experience (QoE) measurements a key priority. Machine Learning (ML) can support this task providing models to analyse the information extracted by the multimedia. It is possible to divide the ML models applications in the following categories: 1) QoE modelling: ML is used to define QoE models which provide an output (e.g., perceived QoE score) for any given input (e.g., QoE influence factor). 2) QoE monitoring in case of encrypted traffic: ML is used to analyze passive traffic monitored data to obtain insight into degradations perceived by end-users. 3) Big data analytics: ML is used for the extraction of meaningful and useful information from the collected data, which can further be converted to actionable knowledge and utilized in managing QoE. The QoE estimation quality task can be carried out by using two approaches: the objective approach and subjective one. As the two names highlight, they are referred to the pieces of information that the model analyses. The objective approach analyses the objective features extracted by the network connection and by the used media. As objective parameters, the state-of-the-art shows different approaches that use also the features extracted by human behaviour. The subjective approach instead, comes as a result of the rating approach, where the participants were asked to rate the perceived quality using different scales. This approach had the problem of being a time-consuming approach and for this reason not all the users agree to compile the questionnaire. Thus the direct evolution of this approach is the ML model adoption. A model can substitute the questionnaire and evaluate the QoE, depending on the data that analyses. By modelling the human response to the perceived quality on multimedia, QoE researchers found that the parameters extracted from the users could be different, like Electroencephalogram (EEG), Electrocardiogram (ECG), waves of the brain. The main problem with these techniques is the hardware. In fact, the user must wear electrodes in case of ECG and EEG, and also if the obtained results from these methods are relevant, their usage in a real context could be not feasible. For this reason, my studies have been focused on the developing of a Machine Learning framework completely unobtrusively based on the Facial reactions

    Implementing Efficient and Multi-Hop Image Acquisition In Remote Monitoring IoT systems using LoRa Technology

    Get PDF
    Remote sensing or monitoring through the deployment of wireless sensor networks (WSNs) is considered an economical and convenient manner in which to collect information without cumbersome human intervention. Unfortunately, due to challenging deployment conditions, such as large geographic area, and lack of electricity and network infrastructure, designing such wireless sensor networks for large-scale farms or forests is difficult and expensive. Many WSN-appropriate wireless technologies, such as Wi-Fi, Bluetooth, Zigbee and 6LoWPAN, have been widely adopted in remote sensing. The performance of these technologies, however, is not sufficient for use across large areas. Generally, as the geographical scope expands, more devices need to be employed to expand network coverage, so the number and cost of devices in wireless sensor networks will increase dramatically. Besides, this type of deployment usually not only has a high probability of failure and high transmission costs, but also imposes additional overhead on system management and maintenance. LoRa is an emerging physical layer standard for long range wireless communication. By utilizing chirp spread spectrum modulation, LoRa features a long communication range and broad signal coverage. At the same time, LoRa also has low power consumption. Thus, LoRa outperforms similar technologies in terms of hardware cost, power consumption and radio coverage. It is also considered to be one of the promising solutions for the future of the Internet of Things (IoT). As the research and development of LoRa are still in its early stages, it lacks sufficient support for multi-packet transport and complex deployment topologies. Therefore, LoRa is not able to further expand its network coverage and efficiently support big data transfers like other conventional technologies. Besides, due to the smaller payload and data rate in LoRa physical design, it is more challenging to implement these features in LoRa. These shortcomings limit the potential for LoRa to be used in more productive application scenarios. This thesis addresses the problem of multi-packet and multi-hop transmission using LoRa by proposing two novel protocols, namely Multi-Packet LoRa (MPLR) and Multi-Hop LoRa (MHLR). LoRa's ability to transmit large messages is first evaluated in this thesis, and then the protocols are well designed and implemented to enrich LoRa's possibilities in image transmission applications and multi-hop topologies. MPLR introduces a reliable transport mechanism for multi-packet sensory data, making its network not limited to the transmission of small sensor data only. In collaboration with a data channel reservation technique, MPLR is able to greatly mitigate data collisions caused by the increased transmission time in laboratory experiments. MHLR realizes efficient routing in LoRa multi-hop transmission by utilizing the power of machine learning. The results of both indoor and outdoor experiments show that the machine learning based routing is effective in wireless sensor networks

    Digital Towns

    Get PDF
    This open access book explores the digital transformation of small and rural towns, in particular, how to measure the evolution and development of digital towns. In addition to access to resources, competition from urban and global markets, and population trends, rural communities present lesser access and use of digital technologies and have lower digital competencies and skills than their urban counterparts. Consequently, they experience less beneficial outcomes from increased digitalisation than urban areas. This book defines what a digital town is and explores digitalisation from the perspective of the four basic economic sectors in towns - individuals and households, businesses, the public sector, and civil society - and three types of enabling infrastructure - digital connectivity, education, and governance. Particular attention is paid to how digitalisation efforts are measured by intergovernmental and international organisations for each sector and enabling infrastructure. The book concludes with a Digital Town Readiness Framework that offers local communities, policymakers, and scholars an initial set of indicators upon which to develop digital town initiatives, and measure progress. For those ready to embrace the opportunity, this book is a pathfinder on the road to a more equitable and impactful digital society and digital economy
    corecore