13 research outputs found

    Evaluation of 3GPP Technology Candidate Towards Fourth Generation Mobile

    Full text link
    [ES] LTE-Advanced es una de las tecnologías candidatas para convertirse en la próxima generación de comunicaciones móviles (4G). Es responsabilidad de la Unión Internacional de las Telecomunicaciones (UIT) evaluar esta tecnología a través de los Grupos de Evaluación Externos (GEE), entre los cuales se encuentra el consorcio WINNER+ (Wireless World Initiative New Radio +). El Grupo de Comunicaciones Móviles (GCM) del Instituto de Telecomunicaciones y Aplicaciones Multimedia, como socio de WINNER+, está analizando diferentes técnicas para optimizar la red de acceso radio LTEAdvanced. Esta tesina de máster se enmarca dentro de este trabajo, y especialmente, en la comparación de los turbo-códigos (TC) y Low Density Partity Check (LDPC) para anchos de banda de hasta 100 MHz. Los resultados obtenidos muestran que tanto los TC como los LDPC son buenos codificadores para esos tamaños de bloque. Los códigos LDPC representan una mejora de 0.5 dB como máximo respecto a los TC. Además, se ha realizado un estudio de prestaciones de la capa física de LTE en el enlace ascendente y descendente, junto con una propuesta de calibración de este tipo de simulaciones de enlace.[EN] LTE-Advanced is one promising candidate technology to become part of the next generation mobile (4G). It is up to the International Telecommunication Union (ITU) standardization body to assess this technology through the External Evaluation Groups (EEG), being one of them the WINNER+ project (Wireless World Initiative New Radio +). The Mobile Communications Group (MCG) of the Institute of Telecommunications and Multimedia Applications, as a partner of WINNER+, is currently analyzing and proposing different techniques with the aim of optimizing the LTE-Advanced radio access network. This Master Thesis is part of this activity and, especially, on the comparison of Turbo (TC) and Low Density Parity Check (LDPC) codes for bandwidths up to 100 MHz. Results prove that both TC and LDPC codes are good encoders for those block sizes. The LDPC codes only entail a maximum 0.5 dB improvement as compared with TC. In addition to this assessment, a performance study of LTE downlink/uplink (DL/ UL) physical layer together with a calibration proposal for link level simulations has been carried out.Cabrejas Peñuelas, J. (2009). Evaluation of 3GPP Technology Candidate Towards Fourth Generation Mobile. http://hdl.handle.net/10251/27347.Archivo delegad

    A comparative investigation on the application and performance of Femtocell against Wi-Fi networks in an indoor environment

    Get PDF
    Due to the strenuous demands on the available spectrum and bandwidth, alongside the ever increasing rate at which data traffic is growing and the poor quality of experience QoE) faced with indoor communications, in order for cellular networks to remain dominant in areas pertaining to voice and data services, cellular service providers have to reform their marketing and service delivery strategies together with their overall network rchitecture. To accomplish this leap forward in performance, cellular service operators need to employ a network topology, which makes use of a mix of macrocells and small cells, effectively evolving the network, bringing it closer to the end-­‐user. This investigation explores the use of small cell technology, specifically Femtocell technology in comparison to the already employed Wi-­‐Fi technology as a viable solution to poor indoor communications.The performance evolution is done by comparing key areas in the every day use of Internet communications. These include HTTP testing, RTP testing and VoIP testing. Results are explained and the modes of operation of both technologies are compared

    MIMO Techniques in UTRA Long Term Evolution

    Get PDF

    The strategies associated with the migration of networks to 4G

    Get PDF
    The networks need to provide higher speeds than those offered today. For it, considering that in the spectrum radio technologies is the scarcest resource in the development of these technologies and the new developments is essential to maximize the performance of bits per hertz transmitted. Long Term Evolution optimize spectral efficiency modulations with new air interface, and more advanced algorithms radius. These capabilities is the fact that LTE is an IPbased technology that enables end-to-end offer high transmission rates per user and very low latency, ie delay in the response times of the network around only 10 milliseconds, so you can offer any realtime application. LTE is the latest standard in mobile network technology and 3GPP ensure competitiveness in the future, may be considered a technology bridge between 3G networks - current 3.5G and future 4G networks, which are expected to reach speeds of up to 1G . LTE operators provide a simplified architecture but both robust, supporting services on IP technology. The objectives to be achieved through its implementation are ambitious, first users have a wide range of added services like capabilities that currently enjoys with residential broadband access at competitive prices, while the operator will have a network fully IP-based environment, reducing the complexity and cost of the same, which will give operators the opportunity to migrate to LTE directly. A major advantage of LTE is its ability to fuse with existing networks, ensuring interconnection with the same, increasing his current coverage and allowing a data connection established by a user in the environment continue when fade the coverage LTE. Moreover, the operator has the advantage of deploying network gradually, starting initially at areas of high demand for broadband services and expand progressively in line with this. RESUMEN. Las redes necesitan proporcionar velocidades mayores a las ofertadas a día de hoy. Para ello, teniendo en cuenta que en tecnologías radio el espectro es el recurso más escaso, en la evolución de estas tecnologías y en los nuevos desarrollos es esencial maximizar el rendimiento de bits por hercio transmitido. Long Term Evolution optimiza la eficiencia espectral con nuevas modulaciones en la interfaz aire, así como los algoritmos radio más avanzado. A estas capacidades se suma el hecho de que LTE es una tecnología basada en IP de extremo a extremo que permite ofrecer altas velocidades de transmisión por usuario y latencias muy bajas, es decir, retardos en los tiempos de respuesta de la red en torno a sólo 10 milisegundos, por lo que permite ofrecer cualquier tipo de aplicación en tiempo real. LTE es el último estándar en tecnología de redes móviles y asegurará la competitividad de 3GPP en el futuro, pudiendo ser considerada una tecnología puente entre las redes 3G – 3.5G actuales y las futuras redes 4G, de las que se esperan alcanzar velocidades de hasta 1G. LTE proporcionará a las operadoras una arquitectura simplificada pero robusta a la vez, soportando servicios sobre tecnología IP. Los objetivos que se persiguen con su implantación son ambiciosos, por una parte los usuarios dispondrá de una amplia oferta de servicios añadidos con capacidades similares a las que disfruta actualmente con accesos a banda ancha residencial y a precios competitivos, mientras que el operador dispondrá de una red basada en entorno totalmente IP, reduciendo la complejidad y el costo de la misma, lo que dará a las operadoras la oportunidad de migrar a LTE directamente. Una gran ventaja de LTE es su capacidad para fusionarse con las redes existentes, asegurando la interconexión con las mismas, aumentando su actual cobertura y permitiendo que una conexión de datos establecida por un usuario en el entorno LTE continúe cuando la cobertura LTE se desvanezca. Por otra parte el operador tiene la ventaja de desplegar la red LTE de forma gradual, comenzando inicialmente por las áreas de gran demanda de servicios de banda ancha y ampliarla progresivamente en función de ésta

    Spatial modulation: theory to practice

    Get PDF
    Spatial modulation (SM) is a transmission technique proposed for multiple–input multiple– output (MIMO) systems, where only one transmit antenna is active at a time, offering an increase in the spectral efficiency equal to the base–two logarithm of the number of transmit antennas. The activation of only one antenna at each time instance enhances the average bit error ratio (ABER) as inter–channel interference (ICI) is avoided, and reduces hardware complexity, algorithmic complexity and power consumption. Thus, SM is an ideal candidate for large scale MIMO (tens and hundreds of antennas). The analytical ABER performance of SM is studied and different frameworks are proposed in other works. However, these frameworks have various limitations. Therefore, a closed–form analytical bound for the ABER performance of SM over correlated and uncorrelated, Rayleigh, Rician and Nakagami–m channels is proposed in this work. Furthermore, in spite of the low–complexity implementation of SM, there is still potential for further reductions, by limiting the number of possible combinations by exploiting the sphere decoder (SD) principle. However, existing SD algorithms do not consider the basic and fundamental principle of SM, that at any given time, only one antenna is active. Therefore, two modified SD algorithms tailored to SM are proposed. It is shown that the proposed sphere decoder algorithms offer an optimal performance, with a significant reduction of the computational complexity. Finally, the logarithmic increase in spectral efficiency offered by SM and the requirement that the number of antennas must be a power of two would require a large number of antennas. To overcome this limitation, two new MIMO modulation systems generalised spatial modulation (GNSM) and variable generalised spatial modulation (VGSM) are proposed, where the same symbol is transmitted simultaneously from more than one transmit antenna at a time. Transmitting the same data symbol from more than one antenna reduces the number of transmit antennas needed and retains the key advantages of SM. In initial development simple channel models can be used, however, as the system develops it should be tested on more realistic channels, which include the interactions between the environment and antennas. Therefore, a full analysis of the ABER performance of SM over urban channel measurements is carried out. The results using the urban measured channels confirm the theoretical work done in the field of SM. Finally, for the first time, the performance of SM is tested in a practical testbed, whereby the SM principle is validated

    Performance analysis of 4G wireless networks using system level simulator

    Get PDF
    Doutoramento em Engenharia ElectrotécnicaIn the last decade, mobile wireless communications have witnessed an explosive growth in the user’s penetration rate and their widespread deployment around the globe. In particular, a research topic of particular relevance in telecommunications nowadays is related to the design and implementation of mobile communication systems of 4th generation (4G). 4G networks will be characterized by the support of multiple radio access technologies in a core network fully compliant with the Internet Protocol (all IP paradigms). Such networks will sustain the stringent quality of service (QoS) requirements and the expected high data rates from the type of multimedia applications (i.e. YouTube and Skype) to be available in the near future. Therefore, 4G wireless communications system will be of paramount importance on the development of the information society in the near future. As 4G wireless services will continue to increase, this will put more and more pressure on the spectrum availability. There is a worldwide recognition that methods of spectrum managements have reached their limit and are no longer optimal, therefore new paradigms must be sought. Studies show that most of the assigned spectrum is under-utilized, thus the problem in most cases is inefficient spectrum management rather spectrum shortage. There are currently trends towards a more liberalized approach of spectrum management, which are tightly linked to what is commonly termed as Cognitive Radio (CR). Furthermore, conventional deployment of 4G wireless systems (one BS in cell and mobile deploy around it) are known to have problems in providing fairness (users closer to the BS are more benefited relatively to the cell edge users) and in covering some zones affected by shadowing, therefore the use of relays has been proposed as a solution. To evaluate and analyse the performances of 4G wireless systems software tools are normally used. Software tools have become more and more mature in recent years and their need to provide a high level evaluation of proposed algorithms and protocols is now more important. The system level simulation (SLS) tools provide a fundamental and flexible way to test all the envisioned algorithms and protocols under realistic conditions, without the need to deal with the problems of live networks or reduced scope prototypes. Furthermore, the tools allow network designers a rapid collection of a wide range of performance metrics that are useful for the analysis and optimization of different algorithms. This dissertation proposes the design and implementation of conventional system level simulator (SLS), which afterwards enhances for the 4G wireless technologies namely cognitive Radios (IEEE802.22) and Relays (IEEE802.16j). SLS is then used for the analysis of proposed algorithms and protocols.FC

    Distributed optimisation techniques for wireless networks

    Get PDF
    Alongside the ever increasing traffic demand, the fifth generation (5G) cellular network architecture is being proposed to provide better quality of service, increased data rate, decreased latency, and increased capacity. Without any doubt, the 5G cellular network will comprise of ultra-dense networks and multiple input multiple output technologies. This will make the current centralised solutions impractical due to increased complexity. Moreover, the amount of coordination information that needs to be transported over the backhaul links will be increased. Distributed or decentralised solutions are promising to provide better alternatives. This thesis proposes new distributed algorithms for wireless networks which aim to reduce the amount of system overheads in the backhaul links and the system complexity. The analysis of conflicts amongst transmitters, and resource allocation are conducted via the use of game theory, convex optimisation, and auction theory. Firstly, game-theoretic model is used to analyse a mixed quality of service (QoS) strategic non-cooperative game (SNG), for a two-user multiple-input single-output (MISO) interference channel. The players are considered to have different objectives. Following this, the mixed QoS SNG is extended to a multicell multiuser network in terms of signal-to-interference-and-noise ratio (SINR) requirement. In the multicell multiuser setting, each transmitter is assumed to be serving real time users (RTUs) and non-real time users (NRTUs), simultaneously. A novel mixed QoS SNG algorithm is proposed, with its operating point identified as the Nash equilibrium-mixed QoS (NE-mixed QoS). Nash, Kalai-Smorodinsky, and Egalitarian bargain solutions are then proposed to improve the performance of the NE-mixed QoS. The performance of the bargain solutions are observed to be comparable to the centralised solutions. Secondly, user offloading and user association problems are addressed for small cells using auction theory. The main base station wishes to offload some of its users to privately owned small cell access points. A novel bid-wait-auction (BWA) algorithm, which allows single-item bidding at each auction round, is designed to decompose the combinatorial mathematical nature of the problem. An analysis on the existence and uniqueness of the dominant strategy equilibrium is conducted. The BWA is then used to form the forward BWA (FBWA) and the backward BWA (BBWA). It is observed that the BBWA allows more users to be admitted as compared to the FBWA. Finally, simultaneous multiple-round ascending auction (SMRA), altered SMRA (ASMRA), sequential combinatorial auction with item bidding (SCAIB), and repetitive combinatorial auction with item bidding (RCAIB) algorithms are proposed to perform user offloading and user association for small cells. These algorithms are able to allow bundle bidding. It is then proven that, truthful bidding is individually rational and leads to Walrasian equilibrium. The performance of the proposed auction based algorithms is evaluated. It is observed that the proposed algorithms match the performance of the centralised solutions when the guest users have low target rates. The SCAIB algorithm is shown to be the most preferred as it provides high admission rate and competitive revenue to the bidders

    On detection of OFDM signals for cognitive radio applications

    Get PDF
    As the requirement for wireless telecommunications services continues to grow, it has become increasingly important to ensure that the Radio Frequency (RF) spectrum is managed efficiently. As a result of the current spectrum allocation policy, it has been found that portions of RF spectrum belonging to licensed users are often severely underutilised, at particular times and geographical locations. Awareness of this problem has led to the development of Dynamic Spectrum Access (DSA) and Cognitive Radio (CR) as possible solutions. In one variation of the shared-use model for DSA, it is proposed that the inefficient use of licensed spectrum could be overcome by enabling unlicensed users to opportunistically access the spectrum when the licensed user is not transmitting. In order for an unlicensed device to make decisions, it must be aware of its own RF environment and, therefore, it has been proposed that DSA could been abled using CR. One approach that has be identified to allow the CR to gain information about its operating environment is spectrum sensing. An interesting solution that has been identified for spectrum sensing is cyclostationary detection. This property refers to the inherent periodic nature of the second order statistics of many communications signals. One of the most common modulation formats in use today is Orthogonal Frequency Division Multiplexing (OFDM), which exhibits cyclostationarity due to the addition of a Cyclic Prefix (CP). This thesis examines several statistical tests for cyclostationarity in OFDM signals that may be used for spectrum sensing in DSA and CR. In particular, focus is placed on statistical tests that rely on estimation of the Cyclic Autocorrelation Function (CAF). Based on splitting the CAF into two complex component functions, several new statistical tests are introduced and are shown to lead to an improvement in detection performance when compared to the existing algorithms. The performance of each new algorithm is assessed in Additive White Gaussian Noise (AWGN), impulsive noise and when subjected to impairments such as multipath fading and Carrier Frequency Offset (CFO). Finally, each algorithm is targeted for Field Programmable Gate Array (FPGA) implementation using a Xilinx 7 series device. In order to keep resource costs to a minimum, it is suggested that the new algorithms are implemented on the FPGA using hardware sharing, and a simple mathematical re-arrangement of certain tests statistics is proposed to circumvent a costly division operation.As the requirement for wireless telecommunications services continues to grow, it has become increasingly important to ensure that the Radio Frequency (RF) spectrum is managed efficiently. As a result of the current spectrum allocation policy, it has been found that portions of RF spectrum belonging to licensed users are often severely underutilised, at particular times and geographical locations. Awareness of this problem has led to the development of Dynamic Spectrum Access (DSA) and Cognitive Radio (CR) as possible solutions. In one variation of the shared-use model for DSA, it is proposed that the inefficient use of licensed spectrum could be overcome by enabling unlicensed users to opportunistically access the spectrum when the licensed user is not transmitting. In order for an unlicensed device to make decisions, it must be aware of its own RF environment and, therefore, it has been proposed that DSA could been abled using CR. One approach that has be identified to allow the CR to gain information about its operating environment is spectrum sensing. An interesting solution that has been identified for spectrum sensing is cyclostationary detection. This property refers to the inherent periodic nature of the second order statistics of many communications signals. One of the most common modulation formats in use today is Orthogonal Frequency Division Multiplexing (OFDM), which exhibits cyclostationarity due to the addition of a Cyclic Prefix (CP). This thesis examines several statistical tests for cyclostationarity in OFDM signals that may be used for spectrum sensing in DSA and CR. In particular, focus is placed on statistical tests that rely on estimation of the Cyclic Autocorrelation Function (CAF). Based on splitting the CAF into two complex component functions, several new statistical tests are introduced and are shown to lead to an improvement in detection performance when compared to the existing algorithms. The performance of each new algorithm is assessed in Additive White Gaussian Noise (AWGN), impulsive noise and when subjected to impairments such as multipath fading and Carrier Frequency Offset (CFO). Finally, each algorithm is targeted for Field Programmable Gate Array (FPGA) implementation using a Xilinx 7 series device. In order to keep resource costs to a minimum, it is suggested that the new algorithms are implemented on the FPGA using hardware sharing, and a simple mathematical re-arrangement of certain tests statistics is proposed to circumvent a costly division operation
    corecore