478 research outputs found

    Advancement in infotainment system in automotive sector with vehicular cloud network and current state of art

    Get PDF
    The automotive industry has been incorporating various technological advancement on top-end versions of the vehicle order to improvise the degree of comfortability as well as enhancing the safer driving system. Infotainment system is one such pivotal system which not only makes the vehicle smart but also offers abundance of information as well as entertainment to the driver and passenger. The capability to offer extensive relay of service through infotainment system is highly dependent on vehicular adhoc network as well as back end support of cloud environment. However, it is know that such legacy system of vehicular adhoc network is also characterized by various problems associated with channel capacity, latency, heterogeneous network processing, and many more. Therefore, this paper offers a comprehensive insight to the research work being carried out towards leveraging the infotainment system in order to obtain the true picture of strength, limitation, and open end problems associated with infotainment system

    Smart handoff technique for internet of vehicles communication using dynamic edge-backup node

    Get PDF
    © 2020 The Authors. Published by MDPI. This is an open access article available under a Creative Commons licence. The published version can be accessed at the following link on the publisher’s website: https://doi.org/10.3390/electronics9030524A vehicular adhoc network (VANET) recently emerged in the the Internet of Vehicles (IoV); it involves the computational processing of moving vehicles. Nowadays, IoV has turned into an interesting field of research as vehicles can be equipped with processors, sensors, and communication devices. IoV gives rise to handoff, which involves changing the connection points during the online communication session. This presents a major challenge for which many standardized solutions are recommended. Although there are various proposed techniques and methods to support seamless handover procedure in IoV, there are still some open research issues, such as unavoidable packet loss rate and latency. On the other hand, the emerged concept of edge mobile computing has gained crucial attention by researchers that could help in reducing computational complexities and decreasing communication delay. Hence, this paper specifically studies the handoff challenges in cluster based handoff using new concept of dynamic edge-backup node. The outcomes are evaluated and contrasted with the network mobility method, our proposed technique, and other cluster-based technologies. The results show that coherence in communication during the handoff method can be upgraded, enhanced, and improved utilizing the proposed technique.Published onlin

    Packet analysis for network forensics: A comprehensive survey

    Get PDF
    Packet analysis is a primary traceback technique in network forensics, which, providing that the packet details captured are sufficiently detailed, can play back even the entire network traffic for a particular point in time. This can be used to find traces of nefarious online behavior, data breaches, unauthorized website access, malware infection, and intrusion attempts, and to reconstruct image files, documents, email attachments, etc. sent over the network. This paper is a comprehensive survey of the utilization of packet analysis, including deep packet inspection, in network forensics, and provides a review of AI-powered packet analysis methods with advanced network traffic classification and pattern identification capabilities. Considering that not all network information can be used in court, the types of digital evidence that might be admissible are detailed. The properties of both hardware appliances and packet analyzer software are reviewed from the perspective of their potential use in network forensics

    On the Exploration of FPGAs and High-Level Synthesis Capabilities on Multi-Gigabit-per-Second Networks

    Full text link
    Tesis doctoral inĂ©dita leĂ­da en la Universidad AutĂłnoma de Madrid, Escuela PolitĂ©cnica Superior, Departamento de Tecnología Electrónica y de las Comunicaciones. Fecha de lectura: 24-01-2020Traffic on computer networks has faced an exponential grown in recent years. Both links and communication equipment had to adapt in order to provide a minimum quality of service required for current needs. However, in recent years, a few factors have prevented commercial off-the-shelf hardware from being able to keep pace with this growth rate, consequently, some software tools are struggling to fulfill their tasks, especially at speeds higher than 10 Gbit/s. For this reason, Field Programmable Gate Arrays (FPGAs) have arisen as an alternative to address the most demanding tasks without the need to design an application specific integrated circuit, this is in part to their flexibility and programmability in the field. Needless to say, developing for FPGAs is well-known to be complex. Therefore, in this thesis we tackle the use of FPGAs and High-Level Synthesis (HLS) languages in the context of computer networks. We focus on the use of FPGA both in computer network monitoring application and reliable data transmission at very high-speed. On the other hand, we intend to shed light on the use of high level synthesis languages and boost FPGA applicability in the context of computer networks so as to reduce development time and design complexity. In the first part of the thesis, devoted to computer network monitoring. We take advantage of the FPGA determinism in order to implement active monitoring probes, which consist on sending a train of packets which is later used to obtain network parameters. In this case, the determinism is key to reduce the uncertainty of the measurements. The results of our experiments show that the FPGA implementations are much more accurate and more precise than the software counterpart. At the same time, the FPGA implementation is scalable in terms of network speed — 1, 10 and 100 Gbit/s. In the context of passive monitoring, we leverage the FPGA architecture to implement algorithms able to thin cyphered traffic as well as removing duplicate packets. These two algorithms straightforward in principle, but very useful to help traditional network analysis tools to cope with their task at higher network speeds. On one hand, processing cyphered traffic bring little benefits, on the other hand, processing duplicate traffic impacts negatively in the performance of the software tools. In the second part of the thesis, devoted to the TCP/IP stack. We explore the current limitations of reliable data transmission using standard software at very high-speed. Nowadays, the network is becoming an important bottleneck to fulfill current needs, in particular in data centers. What is more, in recent years the deployment of 100 Gbit/s network links has started. Consequently, there has been an increase scrutiny of how networking functionality is deployed, furthermore, a wide range of approaches are currently being explored to increase the efficiency of networks and tailor its functionality to the actual needs of the application at hand. FPGAs arise as the perfect alternative to deal with this problem. For this reason, in this thesis we develop Limago an FPGA-based open-source implementation of a TCP/IP stack operating at 100 Gbit/s for Xilinx’s FPGAs. Limago not only provides an unprecedented throughput, but also, provides a tiny latency when compared to the software implementations, at least fifteen times. Limago is a key contribution in some of the hottest topic at the moment, for instance, network-attached FPGA and in-network data processing

    Coordinated Computation of Multi-layer Paths via Inter-layer PCE Communication: Standards, Interoperability and Deployment

    Get PDF
    The Path Computation Element (PCE) is positioned nowadays as one of the solutions that almost every carrier will eventually deploy. The PCE architecture as well as a number of components, including the PCE Communication Protocol (PCEP), have been standardized by the IETF. However, a number of challenges remain to be solved on its way from standards to deployment. In particular, the existing proposals for multilayer path computation within the PCE framework need to be further developed and tested, before considering their possible integration into operational networks. This is especially true for the interoperability of the various PCE implementations and the extensions such as the Virtual Network Topology Manager (VNTM) which cannot be taken for granted. This paper presents a functional implementation of coordinated computation of multilayer paths supported through inter-layer PCE communication, where one PCE is developed by industry and the other as an open-source effort. To this end, we consider an IP/MPLS network deployed over a Wavelength Switched Optical Network (WSON), each of which deploying its own PCE, in an attempt to create an inter-operable multilayer solution. We discuss the key challenges that the research community will face in this area, which in turn will drive a considerable part of the upcoming efforts in terms of standardizationPostprint (published version

    Quality-driven resource utilization methods for video streaming in wireless communication networks

    Get PDF
    This research is focused on the optimisation of resource utilisation in wireless mobile networks with the consideration of the users’ experienced quality of video streaming services. The study specifically considers the new generation of mobile communication networks, i.e. 4G-LTE, as the main research context. The background study provides an overview of the main properties of the relevant technologies investigated. These include video streaming protocols and networks, video service quality assessment methods, the infrastructure and related functionalities of LTE, and resource allocation algorithms in mobile communication systems. A mathematical model based on an objective and no-reference quality assessment metric for video streaming, namely Pause Intensity, is developed in this work for the evaluation of the continuity of streaming services. The analytical model is verified by extensive simulation and subjective testing on the joint impairment effects of the pause duration and pause frequency. Various types of the video contents and different levels of the impairments have been used in the process of validation tests. It has been shown that Pause Intensity is closely correlated with the subjective quality measurement in terms of the Mean Opinion Score and this correlation property is content independent. Based on the Pause Intensity metric, an optimised resource allocation approach is proposed for the given user requirements, communication system specifications and network performances. This approach concerns both system efficiency and fairness when establishing appropriate resource allocation algorithms, together with the consideration of the correlation between the required and allocated data rates per user. Pause Intensity plays a key role here, representing the required level of Quality of Experience (QoE) to ensure the best balance between system efficiency and fairness. The 3GPP Long Term Evolution (LTE) system is used as the main application environment where the proposed research framework is examined and the results are compared with existing scheduling methods on the achievable fairness, efficiency and correlation. Adaptive video streaming technologies are also investigated and combined with our initiatives on determining the distribution of QoE performance across the network. The resulting scheduling process is controlled through the prioritization of users by considering their perceived quality for the services received. Meanwhile, a trade-off between fairness and efficiency is maintained through an online adjustment of the scheduler’s parameters. Furthermore, Pause Intensity is applied to act as a regulator to realise the rate adaptation function during the end user’s playback of the adaptive streaming service. The adaptive rates under various channel conditions and the shape of the QoE distribution amongst the users for different scheduling policies have been demonstrated in the context of LTE. Finally, the work for interworking between mobile communication system at the macro-cell level and the different deployments of WiFi technologies throughout the macro-cell is presented. A QoEdriven approach is proposed to analyse the offloading mechanism of the user’s data (e.g. video traffic) while the new rate distribution algorithm reshapes the network capacity across the macrocell. The scheduling policy derived is used to regulate the performance of the resource allocation across the fair-efficient spectrum. The associated offloading mechanism can properly control the number of the users within the coverages of the macro-cell base station and each of the WiFi access points involved. The performance of the non-seamless and user-controlled mobile traffic offloading (through the mobile WiFi devices) has been evaluated and compared with that of the standard operator-controlled WiFi hotspots

    Design Space Exploration and Resource Management of Multi/Many-Core Systems

    Get PDF
    The increasing demand of processing a higher number of applications and related data on computing platforms has resulted in reliance on multi-/many-core chips as they facilitate parallel processing. However, there is a desire for these platforms to be energy-efficient and reliable, and they need to perform secure computations for the interest of the whole community. This book provides perspectives on the aforementioned aspects from leading researchers in terms of state-of-the-art contributions and upcoming trends
    • 

    corecore