601 research outputs found
Application of advanced technology to space automation
Automated operations in space provide the key to optimized mission design and data acquisition at minimum cost for the future. The results of this study strongly accentuate this statement and should provide further incentive for immediate development of specific automtion technology as defined herein. Essential automation technology requirements were identified for future programs. The study was undertaken to address the future role of automation in the space program, the potential benefits to be derived, and the technology efforts that should be directed toward obtaining these benefits
Proceedings of the Second International Mobile Satellite Conference (IMSC 1990)
Presented here are the proceedings of the Second International Mobile Satellite Conference (IMSC), held June 17-20, 1990 in Ottawa, Canada. Topics covered include future mobile satellite communications concepts, aeronautical applications, modulation and coding, propagation and experimental systems, mobile terminal equipment, network architecture and control, regulatory and policy considerations, vehicle antennas, and speech compression
High-Capacity Communications from Martian Distances
High capacity communications from Martian distances, required for the envisioned human exploration and desirable for data-intensive science missions, is challenging. NASA s Deep Space Network currently requires large antennas to close RF telemetry links operating at kilobit-per-second data rates. To accommodate higher rate communications, NASA is considering means to achieve greater effective aperture at its ground stations. This report, focusing on the return link from Mars to Earth, demonstrates that without excessive research and development expenditure, operational Mars-to-Earth RF communications systems can achieve data rates up to 1 Gbps by 2020 using technology that today is at technology readiness level (TRL) 4-5. Advanced technology to achieve the needed increase in spacecraft power and transmit aperture is feasible at an only moderate increase in spacecraft mass and technology risk. In addition, both power-efficient, near-capacity coding and modulation and greater aperture from the DSN array will be required. In accord with these results and conclusions, investment in the following technologies is recommended:(1) lightweight (1 kg/sq m density) spacecraft antenna systems; (2) a Ka-band receive ground array consisting of relatively small (10-15 m) antennas; (3) coding and modulation technology that reduces spacecraft power by at least 3 dB; and (4) efficient generation of kilowatt-level spacecraft RF power
TDRSS telecommunications study. Phase 1: Final report
A parametric analysis of the telecommunications support capability of the Tracking and Data Relay Satellite System (TDRSS) was performed. Emphasis was placed on maximizing support capability provided to the user while minimizing impact on the user spacecraft. This study evaluates the present TDRSS configuration as presented in the TDRSS Definition Phase Study Report, December 1973 to determine potential changes for improving the overall performance. In addition, it provides specifications of the user transponder equipment to be used in the TDRSS
REAL-TIME ADAPTIVE PULSE COMPRESSION ON RECONFIGURABLE, SYSTEM-ON-CHIP (SOC) PLATFORMS
New radar applications need to perform complex algorithms and process a large quantity of data to generate useful information for the users. This situation has motivated the search for better processing solutions that include low-power high-performance processors, efficient algorithms, and high-speed interfaces. In this work, hardware implementation of adaptive pulse compression algorithms for real-time transceiver optimization is presented, and is based on a System-on-Chip architecture for reconfigurable hardware devices. This study also evaluates the performance of dedicated coprocessors as hardware accelerator units to speed up and improve the computation of computing-intensive tasks such matrix multiplication and matrix inversion, which are essential units to solve the covariance matrix. The tradeoffs between latency and hardware utilization are also presented. Moreover, the system architecture takes advantage of the embedded processor, which is interconnected with the logic resources through high-performance buses, to perform floating-point operations, control the processing blocks, and communicate with an external PC through a customized software interface. The overall system functionality is demonstrated and tested for real-time operations using a Ku-band testbed together with a low-cost channel emulator for different types of waveforms
Mitigation of nonlinear receiver effects in modern radar: advanced signal processing techniques
This thesis presents a study into nonlinearities in the radar receiver and investigates
advanced digital signal processing (DSP) techniques capable of mitigating the resultant
deleterious effects. The need for these mitigation techniques has become more prevalent
as the use of commercial radar sensors has increased rapidly over the last decade. While
advancements in low-cost radio frequency (RF) technologies have made mass-produced
radar systems more feasible, they also pose a significant risk to the functionality of the
sensor. One of the major compromises when employing low-cost commercial off-theshelf
(COTS) components in the radar receiver is system linearity. This linearity trade-off
leaves the radar susceptible to interfering signals as the RF receiver can now be driven
into the weakly nonlinear regime. Radars are not designed to operate in the nonlinear
regime as distortion is observed in the radar output if they do. If radars are to maintain
operational performance in an RF environment that is becoming increasingly crowded,
novel techniques that allow the sensor to operate in the nonlinear regime must be developed.
Advanced DSP techniques offer a low-cost low-impact solution to the nonlinear
receiver problem in modern radar. While there is very little work published on this topic
in the radar literature, inspiration can be taken from the related field of communications
where techniques have been successfully employed.
It is clear from the communications literature that for any mitigation algorithm to be
successful, the mechanisms driving the nonlinear distortion in the receiver must be understood
in great detail. Therefore, a behavioural modelling technique capable of capturing
both the nonlinear amplitude and phase effects in the radar receiver is presented before
any mitigation techniques are studied. Two distinct groups of mitigation algorithms
are then developed specifically for radar systems with their performance tested in the
medium pulse repetition frequency (MPRF) mode of operation. The first of these is the
look-up table (LUT) approach which has the benefit of being mode independent and computationally
inexpensive to implement. The limitations of this communications-based
technique are discussed with particular emphasis placed on its performance against receiver
nonlinearities that exhibit complex nonlinear memory effects. The second group
of mitigation algorithms to be developed is the forward modelling technique. While this
novel technique is both mode dependent and computationally intensive to implement,
it has a unique formalisation that allows it to be extended to include nonlinear memory
effects in a well-defined manner. The performance of this forward modelling technique
is analysed and discussed in detail.
It was shown in this study that nonlinearities generated in the radar receiver can be
successfully mitigated using advanced DSP techniques. For this to be the case however,
the behaviour of the RF receiver must be characterised to a high degree of accuracy both
in the linear and weakly nonlinear regimes. In the case where nonlinear memory effects
are significant in the radar receiver, it was shown that memoryless mitigation techniques
can become decorrelated drastically reducing their effectiveness. Importantly however, it
was demonstrated that the LUT and forward modelling techniques can both be extended
to compensate for complex nonlinear memory effects generated in the RF receiver. It was
also found that the forward modelling technique dealt with the nonlinear memory effects
in a far more robust manner than the LUT approach leading to a superior mitigation
performance in the memory rich case
Recommended from our members
Array Architectures and Physical Layer Design for Millimeter-Wave Communications Beyond 5G
Ever increasing demands in mobile data rates have resulted in exploration of millimeter-wave (mmW) frequencies for the next generation (5G) wireless networks. Communications at mmW frequencies is presented with two keys challenges. Firstly, high propagation loss requires base stations (BSs) and user equipment (UEs) to use a large number of antennas and narrow beams to close the link with sufficient received signal power. Consequently, communications using narrow beams create a new challenge in channel estimation and link establishment based on fine angular probing. Current mmW system use analog phased arrays that can probe only one angle at the time which results in high latency during link establishment and channel tracking. It is desirable to design low latency beam training by exploring both physical layer designs and array architectures that could replace current 5G approaches and pave the way to the communications for frequency bands in higher mmW band and sub-THz region where larger antenna arrays and communications bandwidth can be exploited. To this end, we propose a novel signal processing techniques exploiting unique properties of mmW channel, and show both theoretically, in simulation and experiments its advantages over conventional approaches. Secondly, we explore different array architecture design and analyze their trade-offs between spectral efficiency and power consumption and area. For comprehensive comparison, we have developed a methodology for optimal design of system parameters for different array architecture candidates based on the spectral efficiency target, and use these parameters to estimate the array area and power consumption based on the circuits reported in the literature. We show that the hybrid analog and digital architectures have severe scalability concerns in radio frequency signal distribution with increased array size and spatial multiplexing levels, while the fully-digital array architectures have the best performance and power/area trade-offs.The developed approaches are based on a cross-disciplinary research that combines innovation in model based signal processing, machine learning, and radio hardware. This work is the first to apply compressive sensing (CS), a signal processing tool that exploits sparsity of mmW channel model, to accelerate beam training of mmW cellular system. The algorithm is designed to address practical issues including the requirement of cell discovery and synchronization that involves estimation of angular channel together with carrier frequency offset and timing offsets. We have analyzed the algorithm performance in the 5G compliant simulation and showed that an order of magnitude saving is achieved in initial access latency for the desired channel estimation accuracy. Moreover, we are the first to develop and implement a neural network assisted compressive beam alignment to deal with hardware impairments in mmW radios. We have used 60GHz mmW testbed to perform experiments and show that neural networks approach enhances alignment rate compared to CS. To further accelerate beam training, we proposed a novel frequency selective probing beams using the true-time-delay (TTD) analog array architecture. Our approach utilizes different subcarriers to scan different directions, and achieves a single-shot beam alignment, the fastest approach reported to date. Our comprehensive analysis of different array architectures and exploration of emerging architectures enabled us to develop an order of magnitude faster and energy efficient approaches for initial access and channel estimation in mmW systems
Proceedings of the Mobile Satellite Conference
A satellite-based mobile communications system provides voice and data communications to mobile users over a vast geographic area. The technical and service characteristics of mobile satellite systems (MSSs) are presented and form an in-depth view of the current MSS status at the system and subsystem levels. Major emphasis is placed on developments, current and future, in the following critical MSS technology areas: vehicle antennas, networking, modulation and coding, speech compression, channel characterization, space segment technology and MSS experiments. Also, the mobile satellite communications needs of government agencies are addressed, as is the MSS potential to fulfill them
Spaceborne Imaging Radar Symposium
An overview of the present state of the art in the different scientific and technological fields related to spaceborne imaging radars was presented. The data acquired with the SEASAT SAR (1978) and Shuttle Imaging Radar, SIR-A (1981) clearly demonstrated the important emphasis in the 80's is going to be on in-depth research investigations conducted with the more flexible and sophisticated SIR series instruments and on long term monitoring of geophysical phenomena conducted from free-flying platforms such as ERS-1 and RADARSAT
- …