43 research outputs found
Recommended from our members
On thermal sensor calibration and software techniques for many-core thermal management
The high power density of a many-core processor results in increased temperature which negatively impacts system reliability and performance. Dynamic thermal management applies thermal-aware techniques at run time to avoid overheating using temperature information collected from on-chip thermal sensors. Temperature sensing and thermal control schemes are two critical technologies for successfully maintaining thermal safety. In this dissertation, on-line thermal sensor calibration schemes are developed to provide accurate temperature information.
Software-based dynamic thermal management techniques are proposed using calibrated thermal sensors. Due to process variation and silicon aging, on-chip thermal sensors require periodic calibration before use in DTM. However, the calibration cost for thermal sensors can be prohibitively high as the number of on-chip sensors increases. Linear models which are suitable for on-line calculation are employed to estimate temperatures at multiple sensor locations using performance counters. The estimated temperature and the actual sensor thermal profile show a very high similarity with correlation coefficient ~0.9 for SPLASH2 and SPEC2000 benchmarks.
A calibration approach is proposed to combine potentially inaccurate temperature values obtained from two sources: thermal sensor readings and temperature estimations. A data fusion strategy based on Bayesian inference, which combines information from these two sources, is demonstrated. The result shows the strategy can effectively recalibrate sensor readings in response to inaccuracies caused by process variation and environmental noise. The average absolute error of the corrected sensor temperature readings is
A dynamic task allocation strategy is proposed to address localized overheating in many-core systems. Our approach employs reinforcement learning, a dynamic machine learning algorithm that performs task allocation based on current temperatures and a prediction regarding which assignment will minimize the peak temperature. Our results show that the proposed technique is fast (scheduling performed in \u3c1 \u3ems) and can efficiently reduce peak temperature by up to 8 degree C in a 49-core processor (6% on average) versus a leading competing task allocation approach for a series of SPLASH-2 benchmarks. Reinforcement learning has also been applied to 3D integrated circuits to allocate tasks with thermal awareness
Improving Message Dissemination in Opportunistic Networks
Data transmission has become a need in various fields, like in social networks with the diverse interaction applications, or in the scientific and engineering areas where for example the use of sensors to capture data is growing, or in emergency situations where there is the imperative need to have a communication system to coordinate rescue operations. Wireless networks have been able to solve these issues to a great extent, but what can we do when a fixed supporting infrastructure is not available or becomes inoperative because of saturation? Opportunistic wireless networks are an alternative to consider in these situations, since their operation does not depend on the existence of a telecommunications infrastructure but they provide connectivity through the organized cooperation of users.
This research thesis focuses on these types of networks and is aimed at improving the dissemination of information in opportunistic networks analyzing the main causes that influence the performance of data transmission. Opportunistic networks do not depend on a fixed topology but depend on the number and mobility of users, the type and quantity of information generated and sent, as well as the physical characteristics of the mobile devices that users have to transmit the data. The combination of these elements impacts on the duration of the contact time between mobile users, directly affecting the information delivery probability.
This thesis starts by presenting a thorough "state of the art" study where we present the most important contributions related to this area and the solutions offered for the evaluation of the opportunistic networks, such as simulation models, routing protocols, simulation tools, among others. After offering this broad background, we evaluate the consumption of the resources of the mobile devices that affect the performance of the the applications of opportunistic networks, both from the energetic and the memory point of view.
Next, we analyze the performance of opportunistic networks considering either pedestrian and vehicular environments. The studied approaches include the use of additional fixed nodes and different data transmission technologies, to improve the duration of the contact between mobile devices.
Finally, we propose a diffusion scheme to improve the performance of data transmission based on extending the duration of the contact time and the likelihood that users will collaborate in this process. This approach is complemented by the efficient management of the resources of the mobile devices.La transmisión de datos se ha convertido en una necesidad en diversos ámbitos, como en las redes sociales con sus diversas aplicaciones, o en las áreas científicas y de ingeniería donde, por ejemplo, el uso de sensores para capturar datos está creciendo, o en situaciones de emergencia donde impera la necesidad de tener un sistema de comunicación para coordinar las operaciones de rescate. Las redes inalámbricas actuales han sido capaces de resolver estos problemas en gran medida, pero ¿qué podemos hacer cuando una infraestructura de soporte fija no está disponible o estas se vuelven inoperantes debido a la saturación de peticiones de red? Las redes inalámbricas oportunísticas son una alternativa a considerar en estas situaciones, ya que su funcionamiento no depende de la existencia de una infraestructura de telecomunicaciones sino que la conectividad es a través de la cooperación organizada de los usuarios.
Esta tesis de investigación se centra en estos tipos de redes oportunísticas y tiene como objetivo mejorar la difusión de información analizando las principales causas que influyen en el rendimiento de la transmisión de datos. Las redes oportunísticas no dependen de una topología fija, sino que dependen del número y la movilidad de los usuarios, del tipo y cantidad de información generada y enviada, así como de las características físicas de los dispositivos móviles que los usuarios tienen para transmitir los datos. La combinación de estos elementos influye en la duración del tiempo de contacto entre usuarios móviles, afectando directamente a la probabilidad de entrega de información.
Esta tesis comienza presentando un exhaustivo estudio del ``estado del arte", donde presentamos las contribuciones más importantes relacionadas con esta área y las soluciones existentes para la evaluación de las redes oportunísticas, tales como modelos de simulación, protocolos de enrutamiento, herramientas de simulación, entre otros. Tras ofrecer esta amplia compilación de investigaciones, se evalúa el consumo de recursos de los dispositivos móviles que afectan al rendimiento de las aplicaciones de redes oportunísticas, desde el punto de vista energético así como de la memoria.
A continuación, analizamos el rendimiento de las redes oportunísticas considerando tanto los entornos peatonales como vehiculares. Los enfoques estudiados incluyen el uso de nodos fijos adicionales y diferentes tecnologías de transmisión de datos, para mejorar la duración del contacto entre dispositivos móviles.
Finalmente, proponemos un esquema de difusión para mejorar el rendimiento de la transmisión de datos basado en la extensión de la duración del tiempo de contacto, y de la probabilidad de que los usuarios colaboren en este proceso. Este enfoque se complementa con la gestión eficiente de los recursos de los dispositivos móviles.La transmissió de dades s'ha convertit en una necessitat en diversos àmbits, com ara en les xarxes socials amb les diverses aplicacions d'interacció, o en les àrees científiques i d'enginyeria, en les quals, per exemple, l'ús de sensors per a capturar dades creix en l'actualitat, o en situacions d'emergència en què impera la necessitat de tenir un sistema de comunicació per a coordinar les operacions de rescat. Les xarxes sense fil han sigut capaces de resoldre aquests problemes en gran manera, però què podem fer quan una infraestructura de suport fixa no està disponible, o bé aquestes es tornen inoperants a causa de la saturació de peticions de xarxa? Les xarxes sense fil oportunistes són una alternativa que cal considerar en aquestes situacions, ja que el funcionament d'aquestes xarxes no depèn de l'existència d'una infraestructura de telecomunicacions, sinó que la connectivitat s'hi aconsegueix a través de la cooperació organitzada dels usuaris.
Aquesta tesi de recerca se centra en aquest tipus de xarxes, i té com a objectiu millorar la difusió d'informació en xarxes oportunistes tot analitzant les principals causes que influeixen en el rendiment de la transmissió de dades. Les xarxes oportunistes no depenen d'una topologia fixa, sinó del nombre i la mobilitat dels usuaris, del tipus i la quantitat d'informació generada i enviada, i de les característiques físiques dels dispositius mòbils que els usuaris tenen per a transmetre les dades. La combinació d'aquests elements influeix en la durada del temps de contacte entre usuaris mòbils, i afecta directament la probabilitat de lliurament d'informació.
Aquesta tesi comença amb un estudi exhaustiu de l'estat de la qüestió, en què presentem les contribucions més importants relacionades amb aquesta àrea i les solucions oferides per a l'avaluació de les xarxes oportunistes, com ara models de simulació, protocols d'encaminament o eines de simulació, entre d'altres. Després de mostrar aquest ampli panorama, s'avalua el consum dels recursos dels dispositius mòbils que afecten l'acompliment de les aplicacions de xarxes oportunistes, tant des del punt de vista energètic com de la memòria.
A continuació, analitzem l'acompliment de xarxes oportunistes considerant tant els entorns de vianants com els vehiculars. Els enfocaments estudiats inclouen l'ús de nodes fixos addicionals i diferents tecnologies de transmissió de dades per a millorar la durada del contacte entre dispositius mòbils.
Finalment, proposem un esquema de difusió per a millorar el rendiment de la transmissió de dades basat en l'extensió de la durada del temps de contacte, i de la probabilitat que els usuaris col·laboren en aquest procés. Aquest enfocament es complementa amb la gestió eficient dels recursos dels dispositius mòbils.Herrera Tapia, J. (2017). Improving Message Dissemination in Opportunistic Networks [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/86129TESI
POWER AND PERFORMANCE STUDIES OF THE EXPLICIT MULTI-THREADING (XMT) ARCHITECTURE
Power and thermal constraints gained critical importance in the design of microprocessors over the past decade. Chipmakers failed to keep power at bay while sustaining the performance growth of serial computers at the rate expected by consumers. As an alternative, they turned to fitting an increasing number of simpler cores on a single die. While this is a step forward for relaxing the constraints, the issue of power is far from resolved and it is joined by new challenges which we explain next.
As we move into the era of many-cores, processors consisting of 100s, even 1000s of cores, single-task parallelism is the natural path for building faster general-purpose computers. Alas, the introduction of parallelism to the mainstream general-purpose domain brings another long elusive problem to focus: ease of parallel programming. The result is the dual challenge where power efficiency and ease-of-programming are vital for the prevalence of up and coming many-core architectures.
The observations above led to the lead goal of this dissertation: a first order validation of the claim that even under power/thermal constraints, ease-of-programming and competitive performance need not be conflicting objectives for a massively-parallel general-purpose processor. As our platform, we choose the eXplicit Multi-Threading (XMT) many-core architecture for fine grained parallel programs developed at the University of Maryland. We hope that our findings will be a trailblazer for future commercial products.
XMT scales up to thousand or more lightweight cores and aims at improving single task execution time while making the task for the programmer as easy as possible. Performance advantages and ease-of-programming of XMT have been shown in a number of publications, including a study that we present in this dissertation. Feasibility of the hardware concept has been exhibited via FPGA and ASIC (per our partial involvement) prototypes.
Our contributions target the study of power and thermal envelopes of an envisioned 1024-core XMT chip (XMT1024) under programs that exist in popular parallel benchmark suites. First, we compare XMT against an area and power equivalent commercial high-end many-core GPU. We demonstrate that XMT can provide an average speedup of 8.8x in irregular parallel programs that are common and important in general purpose computing. Even under the worst-case power estimation assumptions for XMT, average speedup is only reduced by half. We further this study by experimentally evaluating the performance advantages of Dynamic Thermal Management (DTM), when applied to XMT1024. DTM techniques are frequently used in current single and multi-core processors, however until now their effects on single-tasked many-cores have not been examined in detail. It is our purpose to explore how existing techniques can be tailored for XMT to improve performance. Performance improvements up to 46% over a generic global management technique has been demonstrated. The insights we provide can guide designers of other similar many-core architectures.
A significant infrastructure contribution of this dissertation is a highly configurable cycle-accurate simulator, XMTSim. To our knowledge, XMTSim is currently the only publicly-available shared-memory many-core simulator with extensive capabilities for estimating power and temperature, as well as evaluating dynamic power and thermal management algorithms. As a major component of the XMT programming toolchain, it is not only used as the infrastructure in this work but also contributed to other publications and dissertations
High-Speed Performance, Power and Thermal Co-simulation For SoC Design
This dissertation presents a multi-faceted effort at developing standard System Design Language based tools that allow designers to the model power and thermal behavior of SoCs, including heterogeneous SoCs that include non-digital components. The research contributions made in this dissertation include:
• SystemC-based power/performance co-simulation for the Intel XScale microprocessor. We performed detailed characterization of the power dissipation patterns of a variety of system components and used these results to build detailed power models, including a highly accurate, validated instruction-level power model of the XScale processor. We also proposed a scalable, efficient and validated methodology for incorporating fast, accurate power modeling capabilities into system description languages such as SystemC. This was validated against physical measurements of hardware power dissipation.
• Modeling the behavior of non-digital SoC components within standard System Design Languages. We presented an approach for modeling the functionality, performance, power, and thermal behavior of a complex class of non-digital components — MEMS microhotplate-based gas sensors — within a SystemC design framework. The components modeled include both digital components (such as microprocessors, busses and memory) and MEMS devices comprising a gas sensor SoC. The first SystemC models of a MEMS-based SoC and the first SystemC models of MEMS thermal behavior were described. Techniques for significantly improving simulation speed were proposed, and their impact quantified.
• Vertically Integrated Execution-Driven Power, Performance and Thermal Co-Simulation For SoCs. We adapted the above techniques and used numerical methods to model the system of differential equations that governs on-chip thermal diffusion. This allows a single high-speed simulation to span performance, power and thermal modeling of a design. It also allows feedback behaviors, such as the impact of temperature on power dissipation or performance, to be modeled seamlessly. We validated the thermal equation-solving engine on test layouts against detailed low-level tools, and illustrated the power of such a strategy by demonstrating a series of studies that designers can perform using such tools. We also assessed how simulation and accuracy are impacted by spatial and temporal resolution used for thermal modeling
New approaches to data access in large-scale distributed system
Mención Internacional en el título de doctorA great number of scientific projects need supercomputing resources, such as, for example, those carried out in physics, astrophysics, chemistry, pharmacology, etc. Most of them generate, as well, a great amount of data; for example, a some minutes long experiment in a particle accelerator generates several terabytes of data.
In the last years, high-performance computing environments have evolved towards large-scale distributed systems such as Grids, Clouds, and Volunteer Computing environments. Managing a great volume of data in these environments means an added huge problem since the data have to travel from one site to another through the internet.
In this work a novel generic I/O architecture for large-scale distributed systems used for high-performance and high-throughput computing will be proposed. This solution is based on applying parallel I/O techniques to remote data access. Novel replication and data search schemes will also be proposed; schemes that, combined with the above techniques, will allow to improve the performance of those applications that execute in these environments. In addition, it will be proposed to develop simulation tools that allow to test these and other ideas without needing to use real platforms due to their technical and logistic limitations. An initial prototype of this solution has been evaluated and the results show a noteworthy improvement regarding to data access compared to existing solutions.Un gran número de proyectos científicos necesitan recursos de supercomputación como, por ejemplo, los llevados a cabo en física, astrofísica, química, farmacología, etc. Muchos de ellos generan, además, una gran cantidad de datos; por ejemplo, un experimento de unos minutos de duración en un acelerador de partículas genera varios terabytes de datos.
Los entornos de computación de altas prestaciones han evolucionado en los últimos años hacia sistemas distribuidos a gran escala tales como Grids, Clouds y entornos de computación voluntaria. En estos entornos gestionar un gran volumen de datos supone un problema añadido de importantes dimensiones ya que los datos tienen que viajar de un sitio a otro a través de internet.
En este trabajo se propondrá una nueva arquitectura de E/S genérica para sistemas distribuidos a gran escala usados para cómputo de altas prestaciones y de alta productividad. Esta solución se basa en la aplicación de técnicas de E/S paralela al acceso remoto a los datos. Así mismo, se estudiarán y propondrán nuevos esquemas de replicación y búsqueda de datos que, en combinación con las técnicas anteriores, permitan mejorar las prestaciones de aquellas aplicaciones que ejecuten en este tipo de entornos. También se propone desarrollar herramientas de simulación que permitan probar estas y otras ideas sin necesidad de recurrir a una plataforma real debido a las limitaciones técnicas y logísticas que ello supone. Se ha evaluado un prototipo inicial de esta solución y los resultados muestran una mejora significativa en el acceso a los datos sobre las soluciones existentes.Programa Oficial de Doctorado en Ciencia y Tecnología InformáticaPresidente: David Expósito Singh.- Secretario: María de los Santos Pérez Hernández.- Vocal: Juan Manuel Tirado Mart
Recommended from our members
QOE-AWARE CONTENT DISTRIBUTION SYSTEMS FOR ADAPTIVE BITRATE VIDEO STREAMING
A prodigious increase in video streaming content along with a simultaneous rise in end system capabilities has led to the proliferation of adaptive bit rate video streaming users in the Internet. Today, video streaming services range from Video-on-Demand services like traditional IP TV to more recent technologies such as immersive 3D experiences for live sports events. In order to meet the demands of these services, the multimedia and networking research community continues to strive toward efficiently delivering high quality content across the Internet while also trying to minimize content storage and delivery costs.
The introduction of flexible and adaptable technologies such as compute and storage clouds, Network Function Virtualization and Software Defined Networking continue to fuel content provider revenue. Today, content providers such as Google and Facebook build their own Software-Defined WANs to efficiently serve millions of users worldwide, while NetFlix partners with ISPs such as ATT (using OpenConnect) and cloud providers such as Amazon EC2 to serve their content and manage the delivery of several petabytes of high-quality video content for millions of subscribers at a global scale, respectively. In recent years, the unprecedented growth of video traffic in the Internet has seen several innovative systems such as Software Defined Networks and Information Centric Networks as well as inventive protocols such as QUIC, in an effort to keep up with the effects of this remarkable growth. While most existing systems continue to sub-optimally satisfy user requirements, future video streaming systems will require optimal management of storage and bandwidth resources that are several orders of magnitude larger than what is implemented today. Moreover, Quality-of-Experience metrics are becoming increasingly fine-grained in order to accurately quantify diverse content and consumer needs.
In this dissertation, we design and investigate innovative adaptive bit rate video streaming systems and analyze the implications of recent technologies on traditional streaming approaches using real-world experimentation methods. We provide useful insights for current and future content distribution network administrators to tackle Quality-of-Experience dilemmas and serve high quality video content to several users at a global scale. In order to show how Quality-of-Experience can benefit from core network architectural modifications, we design and evaluate prototypes for video streaming in Information Centric Networks and Software-Defined Networks. We also present a real-world, in-depth analysis of adaptive bitrate video streaming over protocols such as QUIC and MPQUIC to show how end-to-end protocol innovation can contribute to substantial Quality-of-Experience benefits for adaptive bit rate video streaming systems. We investigate a cross-layer approach based on QUIC and observe that application layer-based information can be successfully used to determine transport layer parameters for ABR streaming applications
Solving key design issues for massively multiplayer online games on peer-to-peer architectures
Massively Multiplayer Online Games (MMOGs) are increasing in both popularity and
scale on the Internet and are predominantly implemented by Client/Server architectures.
While such a classical approach to distributed system design offers many benefits, it suffers
from significant technical and commercial drawbacks, primarily reliability and scalability
costs. This realisation has sparked recent research interest in adapting MMOGs
to Peer-to-Peer (P2P) architectures.
This thesis identifies six key design issues to be addressed by P2P MMOGs, namely
interest management, event dissemination, task sharing, state persistency, cheating mitigation,
and incentive mechanisms. Design alternatives for each issue are systematically
compared, and their interrelationships discussed. How well representative P2P MMOG
architectures fulfil the design criteria is also evaluated. It is argued that although P2P
MMOG architectures are developing rapidly, their support for task sharing and incentive
mechanisms still need to be improved.
The design of a novel framework for P2P MMOGs, Mediator, is presented. It employs a
self-organising super-peer network over a P2P overlay infrastructure, and addresses the
six design issues in an integrated system. The Mediator framework is extensible, as it
supports flexible policy plug-ins and can accommodate the introduction of new superpeer
roles. Key components of this framework have been implemented and evaluated
with a simulated P2P MMOG.
As the Mediator framework relies on super-peers for computational and administrative
tasks, membership management is crucial, e.g. to allow the system to recover from
super-peer failures. A new technology for this, namely Membership-Aware Multicast
with Bushiness Optimisation (MAMBO), has been designed, implemented and evaluated.
It reuses the communication structure of a tree-based application-level multicast
to track group membership efficiently. Evaluation of a demonstration application shows
i
that MAMBO is able to quickly detect and handle peers joining and leaving. Compared
to a conventional supervision architecture, MAMBO is more scalable, and yet incurs
less communication overheads. Besides MMOGs, MAMBO is suitable for other P2P
applications, such as collaborative computing and multimedia streaming.
This thesis also presents the design, implementation and evaluation of a novel task
mapping infrastructure for heterogeneous P2P environments, Deadline-Driven Auctions
(DDA). DDA is primarily designed to support NPC host allocation in P2P MMOGs, and
specifically in the Mediator framework. However, it can also support the sharing of computational
and interactive tasks with various deadlines in general P2P applications. Experimental
and analytical results demonstrate that DDA efficiently allocates computing
resources for large numbers of real-time NPC tasks in a simulated P2P MMOG with approximately
1000 players. Furthermore, DDA supports gaming interactivity by keeping
the communication latency among NPC hosts and ordinary players low. It also supports
flexible matchmaking policies, and can motivate application participants to contribute
resources to the system
Computation offloading for fast and energy-efficient edge computing
In recent years, the demand for computing power has increased considerably due to the popularity of applications that involve computationally intensive tasks such as machine learning or computer vision. At the same time, users increasingly run such applications on smartphones or wearables, which have limited computational power. The research community has proposed computation offloading to meet the demand for computing power. Resource-constrained devices offload workload to remote resource providers. These providers perform the computations and return the results via the network. Computation offloading has two major benefits. First, it accelerates the execution of computationally intensive tasks and therefore reduces waiting times. Second, it decreases the energy consumption of the offloading device, which is especially attractive for devices that run on battery. After years in which cloud servers were the primary resource providers, computation offloading in edge computing systems is currently gaining popularity. Edge-based systems leverage end-user devices such as smartphones, laptops, or desktop PCs instead of cloud servers as computational resource providers. Computation offloading in such environments leads to lower latencies, better utilization of end-user devices, and lower costs in comparison to traditional cloud computing.
In this thesis, we present a computation offloading approach for fast and energy-efficient edge computing. We build upon the Tasklet system – a middleware-based computation offloading system. The Tasklet system allows devices to offload heterogeneous tasks to heterogeneous providers. We address three challenges of computation offloading in the edge. First, many applications are data-intensive, which necessitates a time-consuming transfer of input data ahead of a remote execution. To overcome this challenge, we introduce DataVinci – an approach that proactively places input data on suitable devices to accelerate task execution. DataVinci additionally offers task placement strategies that exploit data locality. Second, modern applications are often user-facing and responsive. They require sub-second execution of computationally intensive tasks to ensure proper user experience. We design the decentralized scheduling approach DecArt for such applications. Third, deciding whether a local or remote execution of an upcoming task will consume less energy is non-trivial. This decision is particularly challenging as task complexity and result data size vary across executions, even if the source code is similar. We introduce the energy-aware scheduling approach Voltaire, which uses machine learning and device-specific energy profiles for making precise offloading decisions. We integrate DataVinci, DecArt, and Voltaire into the Tasklet system and evaluate the benefits in extensive experiments
A Decentralized Session Management Framework for Heterogeneous Ad-Hoc and Fixed Networks
Wireless technologies are continuously evolving. Second generation cellular networks have gained worldwide acceptance. Wireless LANs are commonly deployed in corporations or university campuses, and their diffusion in public hotspots is growing. Third generation cellular systems are yet to affirm everywhere; still, there is an impressive amount of research ongoing for deploying beyond 3G systems. These new wireless technologies combine the characteristics of WLAN based and cellular networks to provide increased bandwidth. The common direction where all the efforts in wireless technologies are headed is towards an IP-based communication.
Telephony services have been the killer application for cellular systems; their evolution to packet-switched networks is a natural path. Effective IP telephony signaling protocols, such as the Session Initiation Protocol (SIP) and the H 323 protocol are needed to establish IP-based telephony sessions. However, IP telephony is just one service example of IP-based communication. IP-based multimedia sessions are expected to become popular and offer a wider range of communication capabilities than pure telephony.
In order to conjoin the advances of the future wireless technologies with the potential of IP-based multimedia communication, the next step would be to obtain ubiquitous communication capabilities. According to this vision, people must be able to communicate also when no support from an infrastructured network is available, needed or desired. In order to achieve ubiquitous communication, end devices must integrate all the capabilities necessary for IP-based distributed and decentralized communication. Such capabilities are currently missing. For example, it is not possible to utilize native IP telephony signaling protocols in a totally decentralized way.
This dissertation presents a solution for deploying the SIP protocol in a decentralized fashion without support of infrastructure servers. The proposed solution is mainly designed to fit the needs of decentralized mobile environments, and can be applied to small scale ad-hoc networks or also bigger networks with hundreds of nodes. A framework allowing discovery of SIP users in ad-hoc networks and the establishment of SIP sessions among them, in a fully distributed and secure way, is described and evaluated. Security support allows ad-hoc users to authenticate the sender of a message, and to verify the integrity of a received message.
The distributed session management framework has been extended in order to achieve interoperability with the Internet, and the native Internet applications. With limited extensions to the SIP protocol, we have designed and experimentally validated a SIP gateway allowing SIP signaling between ad-hoc networks with private addressing space and native SIP applications in the Internet. The design is completed by an application level relay that permits instant messaging sessions to be established in heterogeneous environments. The resulting framework constitutes a flexible and effective approach for the pervasive deployment of real time applications.The invention of the phone has radically changed the way people communicate, as it allowed persons to get in contact instantly no matter of their location. However, phone communication has been confined for decades to a fixed location, be it one's own house or a phone boot. The widespread affirmation of cellular technologies has had for fixed telephony a similar impact that the invention of the phone has had on communications years before. With mobile phones, people are enabled to talk with each other anytime and anywhere.
Internet has also revolutionized the way people communicate. E-mails have soon become one of the Internet killer applications. Later on, instant messaging, popularly known as chatting, has gained huge consensus among net surfers. Only recently, the use of the Internet for voice communication is becoming mainstream, and the so called Voice over IP (VoIP) applications (Skype is probably the most famous for the masses) are becoming common use.
Despite its popularity, Internet still suffers from the inherent limitations that affected early telephony: it is fixed. The usage of Internet on the move still does not constitute the easiest and most satisfactory user experience, due to capabilities and limitations of the access technology, terminals, services and applications.
Efforts for mobilizing the Internet are ongoing both in the industrial and in the academic worlds, but several bricks are needed to build the wall of mobile Internet. This dissertation provides one of these bricks, describing a solution that allows the deployment of multimedia applications (chat, VoIP, gaming) in mobile environments. In other words, this dissertation gives solutions for facilitating ubiquitous Internet-based communication, anytime and anywhere.
The vision that we want to become true is that Internet must become mobile in the same way as fixed telephony has become mobile thanks to the cellular technology. More than this, we do not want that users are limited by the presence of an infrastructure to communicate with each other. In order to achieve this, we present solutions to deploy Internet-based services and applications in environments where no support from servers is available. In other words, we enable direct device-to-device, user-to-user Internet communication.
Our contribution is mainly focused on the steps needed to establish the communication, the so called session establishment or signaling phase. We have validated our signaling framework by building a chat application that utilizes its features and works in server-less environments.
The custom server-less solution does not prohibit to connect at the same time with the Internet, so that one can engage in a chess game using direct communication with a person in the proximity while having a chat in progress with a friend using standard Internet services. The challenge that we had to face is that Internet services and applications are usually built implying support from a centralized server. In order to deploy direct user-to-user Internet services, while maintaining interoperability with mainstream services, we had to enhance native Internet services to work without infrastructure support, without sacrificing interoperability with standard Internet applications.
To conclude, we have placed our brick on the still yet to be completed wall of mobile Internet. Our hope is that one day, thanks also to this brick, everybody will be able to enjoy Internet-based applications as easily as now it is possible to use mobile telephony services
High-level services for networks-on-chip
Future technology trends envision that next-generation Multiprocessors Systems-on- Chip (MPSoCs) will be composed of a combination of a large number of processing and storage elements interconnected by complex communication architectures. Communication and interconnection between these basic blocks play a role of crucial importance when the number of these elements increases. Enabling reliable communication channels between cores becomes therefore a challenge for system designers. Networks-on-Chip (NoCs) appeared as a strategy for connecting and managing the communication between several design elements and IP blocks, as required in complex Systems-on-Chip (SoCs). The topic can be considered as a multidisciplinary synthesis of multiprocessing, parallel computing, networking, and on- chip communication domains. Networks-on-Chip, in addition to standard communication services, can be employed for providing support for the implementation of system-level services. This dissertation will demonstrate how high-level services can be added to an MPSoC platform by embedding appropriate hardware/software support in the network interfaces (NIs) of the NoC. In this dissertation, the implementation of innovative modules acting in parallel with protocol translation and data transmission in NIs is proposed and evaluated. The modules can support the execution of the high-level services in the NoC at a relatively low cost in terms of area and energy consumption. Three types of services will be addressed and discussed: security, monitoring, and fault tolerance. With respect to the security aspect, this dissertation will discuss the implementation of an innovative data protection mechanism for detecting and preventing illegal accesses to protected memory blocks and/or memory mapped peripherals. The second aspect will be addressed by proposing the implementation of a monitoring system based on programmable multipurpose monitoring probes aimed at detecting NoC internal events and run-time characteristics. As last topic, new architectural solutions for the design of fault tolerant network interfaces will be presented and discussed