194 research outputs found

    Aerial base station placement in temporary-event scenarios

    Get PDF
    Die Anforderungen an den Netzdatenverkehr sind in den letzten Jahren dramatisch gestiegen, was ein großes Interesse an der Entwicklung neuartiger Lösungen zur Erhöhung der Netzkapazität in Mobilfunknetzen erzeugt hat. Besonderes Augenmerk wurde auf das Problem der Kapazitätsverbesserung bei temporären Veranstaltungen gelegt, bei denen das Umfeld im Wesentlichen dynamisch ist. Um der Dynamik der sich verändernden Umgebung gerecht zu werden und die Bodeninfrastruktur durch zusätzliche Kapazität zu unterstützen, wurde der Einsatz von Luftbasisstationen vorgeschlagen. Die Luftbasisstationen können in der Nähe des Nutzers platziert werden und aufgrund der im Vergleich zur Bodeninfrastruktur höheren Lage die Vorteile der Sichtlinienkommunikation nutzen. Dies reduziert den Pfadverlust und ermöglicht eine höhere Kanalkapazität. Das Optimierungsproblem der Maximierung der Netzkapazität durch die richtige Platzierung von Luftbasisstationen bildet einen Schwerpunkt der Arbeit. Es ist notwendig, das Optimierungsproblem rechtzeitig zu lösen, um auf Veränderungen in der dynamischen Funkumgebung zu reagieren. Die optimale Platzierung von Luftbasisstationen stellt jedoch ein NP-schweres Problem dar, wodurch die Lösung nicht trivial ist. Daher besteht ein Bedarf an schnellen und skalierbaren Optimierungsalgorithmen. Als Erstes wird ein neuartiger Hybrid-Algorithmus (Projected Clustering) vorgeschlagen, der mehrere Lösungen auf der Grundlage der schnellen entfernungsbasierten Kapazitätsapproximierung berechnet und sie auf dem genauen SINR-basierten Kapazitätsmodell bewertet. Dabei werden suboptimale Lösungen vermieden. Als Zweites wird ein neuartiges verteiltes, selbstorganisiertes Framework (AIDA) vorgeschlagen, welches nur lokales Wissen verwendet, den Netzwerkmehraufwand verringert und die Anforderungen an die Kommunikation zwischen Luftbasisstationen lockert. Bei der Formulierung des Platzierungsproblems konnte festgestellt werden, dass Unsicherheiten in Bezug auf die Modellierung der Luft-Bodensignalausbreitung bestehen. Da dieser Aspekt im Rahmen der Analyse eine wichtige Rolle spielt, erfolgte eine Validierung moderner Luft-Bodensignalausbreitungsmodelle, indem reale Messungen gesammelt und das genaueste Modell für die Simulationen ausgewählt wurden.As the traffic demands have grown dramatically in recent years, so has the interest in developing novel solutions that increase the network capacity in cellular networks. The problem of capacity improvement is even more complex when applied to a dynamic environment during a disaster or temporary event. The use of aerial base stations has received much attention in the last ten years as the solution to cope with the dynamics of the changing environment and to supplement the ground infrastructure with extra capacity. Due to higher elevations and possibility to place aerial base stations in close proximity to the user, path loss is significantly smaller in comparison to the ground infrastructure, which in turn enables high data capacity. We are studying the optimization problem of maximizing network capacity by proper placement of aerial base stations. To handle the changes in the dynamic radio environment, it is necessary to promptly solve the optimization problem. However, we show that the optimal placement of aerial base stations is the NP-hard problem and its solution is non-trivial, and thus, there is a need for fast and scalable optimization algorithms. This dissertation investigates how to solve the placement problem efficiently and to support the dynamics of temporary events. First, we propose a novel hybrid algorithm (Projected Clustering), which calculates multiple solutions based on the fast distance-based capacity approximation and evaluates them on the accurate SINR-based capacity model, avoiding sub-optimal solutions. Second, we propose a novel distributed, self-organized framework (AIDA), which conducts a decision-making process using only local knowledge, decreasing the network overhead and relaxing the requirements for communication between aerial base stations. During the formulation of the placement problem, we found that there is still considerable uncertainty with regard to air-to-ground propagation modeling. Since this aspect plays an important role in our analysis, we validated state-of-the-art air-to-ground propagation models by collecting real measurements and chose the most accurate model for the simulations

    A survey on cost-effective context-aware distribution of social data streams over energy-efficient data centres

    Get PDF
    Social media have emerged in the last decade as a viable and ubiquitous means of communication. The ease of user content generation within these platforms, e.g. check-in information, multimedia data, etc., along with the proliferation of Global Positioning System (GPS)-enabled, always-connected capture devices lead to data streams of unprecedented amount and a radical change in information sharing. Social data streams raise a variety of practical challenges, including derivation of real-time meaningful insights from effectively gathered social information, as well as a paradigm shift for content distribution with the leverage of contextual data associated with user preferences, geographical characteristics and devices in general. In this article we present a comprehensive survey that outlines the state-of-the-art situation and organizes challenges concerning social media streams and the infrastructure of the data centres supporting the efficient access to data streams in terms of content distribution, data diffusion, data replication, energy efficiency and network infrastructure. We systematize the existing literature and proceed to identify and analyse the main research points and industrial efforts in the area as far as modelling, simulation and performance evaluation are concerned

    S-RLNC based MAC Optimization for Multimedia Data Transmission over LTE/LTE-A Network

    Get PDF
    The high pace emergence in communication systems and associated demands has triggered academia-industries to achieve more efficient solution for Quality of Service (QoS) delivery for which recently introduced Long Term Evolution (LTE) or LTE-Advanced has been found as a promising solution. However, enabling QoS and Quality of Experience (QoE) delivery for multimedia data over LTE has always been a challenging task. QoS demands require reliable data transmission with minimum signalling overheads, computational complexity, minimum latency etc, for which classical Hybrid Automatic Repeat Request (HREQ) based LTE-MAC is not sufficient. To alleviate these issues, in this paper a novel and robust Multiple Generation Mixing (MGM) assisted Systematic Random Linear Network Coding (S-RLNC) model is developed to be used at the top of LTE MAC protocol stack for multimedia data transmission over LTE/LTE-A system. Our proposed model incorporated interleaving and coding approach along with MGM to ensure secure, resource efficient and reliable multiple data delivery over LTE systems. The simulation results reveal that our proposed S-RLNC-MGM based MAC can ensure QoS/QoE delivery over LTE systems for multimedia data communication

    Issues and Challenges Facing Low Latency in Tactile Internet

    Get PDF
    Tactile Internet is considered as the next step towards a revolutionary impact on the society, this is due to the introduction of different types of applications mainly the haptic ones that require strict Quality of Service guarantee especially in terms of latency. This would be a major challenge towards the design of new communication technologies and protocols in order to provide ultra-low latency. This article discusses the diverse technologies, communication protocols, and the necessary infrastructure to provide low latency based principally on the fifth generation (5G) of mobile network that is considered as the key enablers of the Tactile Internet. Furthermore, current research direction along with future challenges and open issues are discussed extensively

    Towards Efficient and Scalable Data-Intensive Content Delivery: State-of-the-Art, Issues and Challenges

    Get PDF
    This chapter presents the authors’ work for the Case Study entitled “Delivering Social Media with Scalability” within the framework of High-Performance Modelling and Simulation for Big Data Applications (cHiPSet) COST Action 1406. We identify some core research areas and give an outline of the publications we came up within the framework of the aforementioned action. The ease of user content generation within social media platforms, e.g. check-in information, multimedia data, etc., along with the proliferation of Global Positioning System (GPS)-enabled, always-connected capture devices lead to data streams of unprecedented amount and a radical change in information sharing. Social data streams raise a variety of practical challenges: derivation of real-time meaningful insights from effectively gathered social information, a paradigm shift for content distribution with the leverage of contextual data associated with user preferences, geographical characteristics and devices in general, etc. In this article we present the methodology we followed, the results of our work and the outline of a comprehensive survey, that depicts the state-of-the-art situation and organizes challenges concerning social media streams and the infrastructure of the data centers supporting the efficient access to data streams in terms of content distribution, data diffusion, data replication, energy efficiency and network infrastructure. The challenges of enabling better provisioning of social media data have been identified and they were based on the context of users accessing these resources. The existing literature has been systematized and the main research points and industrial efforts in the area were identified and analyzed. In our works, in the framework of the Action, we came up with potential solutions addressing the problems of the area and described how these fit in the general ecosystem

    A Comprehensive Survey of the Tactile Internet: State of the art and Research Directions

    Get PDF
    The Internet has made several giant leaps over the years, from a fixed to a mobile Internet, then to the Internet of Things, and now to a Tactile Internet. The Tactile Internet goes far beyond data, audio and video delivery over fixed and mobile networks, and even beyond allowing communication and collaboration among things. It is expected to enable haptic communication and allow skill set delivery over networks. Some examples of potential applications are tele-surgery, vehicle fleets, augmented reality and industrial process automation. Several papers already cover many of the Tactile Internet-related concepts and technologies, such as haptic codecs, applications, and supporting technologies. However, none of them offers a comprehensive survey of the Tactile Internet, including its architectures and algorithms. Furthermore, none of them provides a systematic and critical review of the existing solutions. To address these lacunae, we provide a comprehensive survey of the architectures and algorithms proposed to date for the Tactile Internet. In addition, we critically review them using a well-defined set of requirements and discuss some of the lessons learned as well as the most promising research directions

    4G/5G cellular networks metrology and management

    Get PDF
    La prolifération d'applications et de services sophistiqués s'accompagne de diverses exigences de performances, ainsi que d'une croissance exponentielle du trafic pour le lien montant (uplink) et descendant (downlink). Les réseaux cellulaires tels que 4G et 5G évoluent pour prendre en charge cette quantité diversifiée et énorme de données. Le travail de cette thèse vise le renforcement de techniques avancées de gestion et supervision des réseaux cellulaires prenant l'explosion du trafic et sa diversité comme deux des principaux défis dans ces réseaux. La première contribution aborde l'intégration de l'intelligence dans les réseaux cellulaires via l'estimation du débit instantané sur le lien montant pour de petites granularités temporelles. Un banc d'essai 4G temps réel est déployé dans ce but de fournir un benchmark exhaustif des métriques de l'eNB. Des estimations précises sont ainsi obtenues. La deuxième contribution renforce le découpage 5G en temps réel au niveau des ressources radio dans un système multicellulaire. Pour cela, deux modèles d'optimisation ont été proposés. Du fait de leurs temps d'exécution trop long, des heuristiques ont été développées et évaluées en comparaisons des modèles optimaux. Les résultats sont prometteurs, les deux heuristiques renforçant fortement le découpage du RAN en temps réel.The proliferation of sophisticated applications and services comes with diverse performance requirements as well as an exponential traffic growth for both upload and download. The cellular networks such as 4G and 5G are advocated to support this diverse and huge amount of data. This thesis work targets the enforcement of advanced cellular network supervision and management techniques taking the traffic explosion and diversity as two main challenges in these networks. The first contribution tackles the intelligence integration in cellular networks through the estimation of users uplink instantaneous throughput at small time granularities. A real time 4G testbed is deployed for such aim with an exhaustive metrics benchmark. Accurate estimations are achieved.The second contribution enforces the real time 5G slicing from radio resources perspective in a multi-cell system. For that, two exact optimization models are proposed. Due to their high convergence time, heuristics are developed and evaluated with the optimal models. Results are promising, as two heuristics are highly enforcing the real time RAN slicing

    High-Performance Modelling and Simulation for Big Data Applications

    Get PDF
    This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications

    Survey: Benefits of integrating both wireless sensors networks and cloud computing infrastructure

    Get PDF
    Cloud computing has the capabilities of powerful processing and scalable storage with the ability of offline and online data analysis and mining of the collected sensed data from body areas networks. Cloud computing can be considered as the main enabler for modern manufacturing industries. Cloud computing can efficiently serve key areas of manufacturing by aspects of the pay-as-you-go business model, scaling up and down production according to certain demands, more customized solutions, and flexible deployments. In cloud manufacturing, the distributed sensors and resources can be managed in centralized architecture that allows cloud users to request more specific product design, testing at all the stages of the product. This study covers the main points of Integrating Both Wireless Sensors Networks and Cloud Computing Infrastructure and gives a view of the various advantage and disadvantages of methods in integration

    High-Performance Modelling and Simulation for Big Data Applications

    Get PDF
    This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications
    corecore