2,126 research outputs found

    Wideband data-independent beamforming for subarrays

    Get PDF
    The desire to operate large antenna arrays for e.g. RADAR applications over a wider frequency range is currently limited by the hardware, which due to weight, cost and size only permits complex multipliers behind each element. In contrast, wideband processing would have to rely on tap delay lines enabling digital filters for every element.As an intermediate step, in this thesis we consider a design where elements are grouped into subarrays, within which elements are still individually controlled by narrowband complex weights, but where each subarray output is given a tap delay line or finite impulse response digital filter for further wideband processing. Firstly, this thesis explores how a tap delay line attached to every subarray can be designed as a delay-and-sum beamformer. This filter is set to realised a fractional delay design based on a windowed sinc function. At the element level, we show that designing a narrowband beam w.r.t. a centre frequency of wideband operation is suboptimal,and suggest an optimisation technique that can yield sufficiently accurate gain over a frequency band of interest for an arbitrary look direction, which however comes at the cost of reduced aperture efficiency, as well as significantly increased sidelobes. We also suggest an adaptive method to enhance the frequency characteristic of a partial wideband array design, by utilising subarrays pointing in different directions in different frequency bands - resolved by means of a filter bank - to adaptively suppress undesired components in the beam patterns of the subarrays. Finally, the thesis proposes a novel array design approach obtained by rotational tiling of subarrays such that the overall array aperture is densely constructed from the same geometric subarray by rotation and translation only. Since the grating lobes of differently oriented subarrays do not necessarily align, an effective grating lobe attenuation w.r.t. the main beam is achieved. Based on a review of findings from geometry,a number of designs are highlight and transformed into numerical examples, and the theoretically expected grating lobe suppression is compared to uniformly spaced arrays.Supported by a number of models and simulations, the thesis thus suggests various numerical and hardware design techniques, mainly the addition of tap-delay-line per subarray and some added processing overhead, that can help to construct a large partial wideband array close in wideband performance to currently existing hardware.The desire to operate large antenna arrays for e.g. RADAR applications over a wider frequency range is currently limited by the hardware, which due to weight, cost and size only permits complex multipliers behind each element. In contrast, wideband processing would have to rely on tap delay lines enabling digital filters for every element.As an intermediate step, in this thesis we consider a design where elements are grouped into subarrays, within which elements are still individually controlled by narrowband complex weights, but where each subarray output is given a tap delay line or finite impulse response digital filter for further wideband processing. Firstly, this thesis explores how a tap delay line attached to every subarray can be designed as a delay-and-sum beamformer. This filter is set to realised a fractional delay design based on a windowed sinc function. At the element level, we show that designing a narrowband beam w.r.t. a centre frequency of wideband operation is suboptimal,and suggest an optimisation technique that can yield sufficiently accurate gain over a frequency band of interest for an arbitrary look direction, which however comes at the cost of reduced aperture efficiency, as well as significantly increased sidelobes. We also suggest an adaptive method to enhance the frequency characteristic of a partial wideband array design, by utilising subarrays pointing in different directions in different frequency bands - resolved by means of a filter bank - to adaptively suppress undesired components in the beam patterns of the subarrays. Finally, the thesis proposes a novel array design approach obtained by rotational tiling of subarrays such that the overall array aperture is densely constructed from the same geometric subarray by rotation and translation only. Since the grating lobes of differently oriented subarrays do not necessarily align, an effective grating lobe attenuation w.r.t. the main beam is achieved. Based on a review of findings from geometry,a number of designs are highlight and transformed into numerical examples, and the theoretically expected grating lobe suppression is compared to uniformly spaced arrays.Supported by a number of models and simulations, the thesis thus suggests various numerical and hardware design techniques, mainly the addition of tap-delay-line per subarray and some added processing overhead, that can help to construct a large partial wideband array close in wideband performance to currently existing hardware

    Optimal Space-Time-Frequency Design of Microphone Networks

    Get PDF
    Consider a sensing system using a large number of N microphones placed in multiple dimensions to monitor a acoustic field. Using all the microphones at once is impractical because of the amount data generated. Instead, we choose a subset of D microphones to be active. Specifically, we wish to find the D set of microphones that minimizes the largest interference gain at multiple frequencies while monitoring a target of interest. A direct, combinatorial approach - testing all N choose D subsets of microphones - is impractical because of problem size. Instead, we use a convex optimization technique that induces sparsity through a l1-penalty to determine which subset of microphones to use. Our work investigates not only the optimal placement (space) of microphones but also how to process the output of each microphone (time/frequency). We explore this problem for both single and multi-frequency sources, optimizing both microphone weights and positions simultaneously. In addition, we explore this problem for random sources where the output of each of the N microphones is processed by an individual multirate filterbank. The N processed filterbank outputs are then combined to form one final signal. In this case, we fix all the analysis filters and optimize over all the synthesis filters. We show how to convert the continuous frequency problem to a discrete frequency approximation that is computationally tractable. In this random source/multirate filterbank case, we once again optimize over space-time-frequency simultaneously

    compressive synthetic aperture sonar imaging with distributed optimization

    Get PDF
    Synthetic aperture sonar (SAS) provides high-resolution acoustic imaging by processing coherently the backscattered acoustic signal recorded over consecutive pings. Traditionally, object detection and classification tasks rely on high-resolution seafloor mapping achieved with widebeam, broadband SAS systems. However, aspect- or frequency-specific information is crucial for improving the performance of automatic target recognition algorithms. For example, low frequencies can be partly transmitted through objects or penetrate the seafloor providing information about internal structure and buried objects, while multiple views provide information about the object's shape and dimensions. Sub-band and limited-view processing, though, degrades the SAS resolution. In this paper, SAS imaging is formulated as an l1-norm regularized least-squares optimization problem which improves the resolution by promoting a parsimonious representation of the data. The optimization problem is solved in a distributed and computationally efficient way with an algorithm based on the alternating direction method of multipliers. The resulting SAS image is the consensus outcome of collaborative filtering of the data from each ping. The potential of the proposed method for high-resolution, narrowband, and limited-aspect SAS imaging is demonstrated with simulated and experimental data.Synthetic aperture sonar (SAS) provides high-resolution acoustic imaging by processing coherently the backscattered acoustic signal recorded over consecutive pings. Traditionally, object detection and classification tasks rely on high-resolution seafloor mapping achieved with widebeam, broadband SAS systems. However, aspect- or frequency-specific information is crucial for improving the performance of automatic target recognition algorithms. For example, low frequencies can be partly transmitted through objects or penetrate the seafloor providing information about internal structure and buried objects, while multiple views provide information about the object's shape and dimensions. Sub-band and limited-view processing, though, degrades the SAS resolution. In this paper, SAS imaging is formulated as an l1-norm regularized least-squares optimization problem which improves the resolution by promoting a parsimonious representation of the data. The optimization problem is solved in a distributed and computati..

    Antenna selection in massive mimo based on matching pursuit

    Get PDF
    As wireless services proliferate, the demand for available spectrum also grows. As a result, the spectral efficiency is still an issue addressed by many researchers looking for solutions to provide quality of service to a growing number of users. massive MIMO is an attractive technology for the next wireless systems since it can alleviate the expected spectral shortage. This work proposes two antenna selection strategies to be applied in the downlink of a massive MIMO system, aiming at reducing the transmission power. The proposed algorithms can also be employed to select a subset of active sensors in centralized sensor networks. The proposed strategy to select the antennas is inspired by the matching pursuit technique. The presented results show that an efficient selection can be obtained with reduced computational complexity.Com a proliferação de serviços wireless, a demanda por espectro disponível também cresce. Logo, a eficiência espectral é um assunto de grande interesse na comunidade científica, que procura por meios para fornecer qualidade de serviço ao crescente número de usuários. massive MIMO é uma técnica repleta de atrativos a ser empregada na futura geração wireless, já que aproveita o espectro existente eficientemente. Este trabalho propõe duas estratégias de seleção de antenas para serem empregadas no downlink de um sistema massive MIMO, visando a redução da potência de transmissão. Os algoritmos propostos podem também ser usados para selecionar um subconjunto de sensores ativos em uma rede centralizada de sensores. A estratégia proposta para seleção de antenas é inspirada na técnica matching pursuit. Os resultados apresentados indicam que uma seleção eficiente pode ser obtida com baixa complexidade computacional

    Quantifying Potential Energy Efficiency Gain in Green Cellular Wireless Networks

    Full text link
    Conventional cellular wireless networks were designed with the purpose of providing high throughput for the user and high capacity for the service provider, without any provisions of energy efficiency. As a result, these networks have an enormous Carbon footprint. In this paper, we describe the sources of the inefficiencies in such networks. First we present results of the studies on how much Carbon footprint such networks generate. We also discuss how much more mobile traffic is expected to increase so that this Carbon footprint will even increase tremendously more. We then discuss specific sources of inefficiency and potential sources of improvement at the physical layer as well as at higher layers of the communication protocol hierarchy. In particular, considering that most of the energy inefficiency in cellular wireless networks is at the base stations, we discuss multi-tier networks and point to the potential of exploiting mobility patterns in order to use base station energy judiciously. We then investigate potential methods to reduce this inefficiency and quantify their individual contributions. By a consideration of the combination of all potential gains, we conclude that an improvement in energy consumption in cellular wireless networks by two orders of magnitude, or even more, is possible.Comment: arXiv admin note: text overlap with arXiv:1210.843

    Advanced algorithms for audio and image processing

    Get PDF
    The objective of the thesis is the development of a set of innovative algorithms around the topic of beamforming in the field of acoustic imaging, audio and image processing, aimed at significantly improving the performance of devices that exploit these computational approaches. Therefore the context is the improvement of devices (ultrasound machines and video/audio devices) already on the market or the development of new ones which, through the proposed studies, can be introduced on new the markets with the launch of innovative high-tech start-ups. This is the motivation and the leitmotiv behind the doctoral work carried out. In fact, in the first part of the work an innovative image reconstruction algorithm in the field of ultrasound biomedical imaging is presented, which is connected to the development of such equipment that exploits the computing opportunities currently offered nowadays at low cost by GPUs (Moore\u2019s law). The proposed target is to obtain a new pipeline of the reconstruction of the image abandoning the architecture of such hardware based In the first part of the thesis I faced the topic of the reconstruction of ultrasound images for applications hypothesized on a software based device through image reconstruction algorithms processed in the frequency domain. An innovative beamforming algorithm based on seismic migration is presented, in which a transformation of the RF data is carried out and the reconstruction algorithm can evaluate a masking of the k-space of the data, speeding up the reconstruction process and reducing the computational burden. The analysis and development of the algorithms responsible for carrying out the thesis has been approached from a feasibility point in an off-line context and on the Matlab platform, processing both synthetic simulated generated data and real RF data: the subsequent development of these algorithms within of the future ultrasound biomedical equipment will exploit an high-performance computing framework capable of processing customized kernel pipelines (henceforth called \u2019filters\u2019) on CPU/GPU. The type of filters implemented involved the topic of Plane Wave Imaging (PWI), an alternative method of acquiring the ultrasound image compared to the state of the art of the traditional standard B-mode which currently exploit sequential sequence of insonification of the sample under examination through focused beams transmitted by the probe channels. The PWI mode is interesting and opens up new scenarios compared to the usual signal acquisition and processing techniques, with the aim of making signal processing in general and image reconstruction in particular faster and more flexible, and increasing importantly the frame rate opens up and improves clinical applications. The innovative idea is to introduce in an offline seismic reconstruction algorithm for ultrasound imaging a further filter, named masking matrix. The masking matrices can be computed offline knowing the system parameters, since they do not depend from acquired data. Moreover, they can be pre-multiplied to propagation matrices, without affecting the overall computational load. Subsequently in the thesis, the topic of beamforming in audio processing on super-direct linear arrays of microphones is addressed. The aim is to make an in depth analysis of two main families of data-independent approaches and algorithms present in the literature by comparing their performances and the trade-off between directivity and frequency invariance, which is not yet known at to the state-of-the-art. The goal is to validate the best algorithm that allows, from the perspective of an implementation, to experimentally verify performance, correlating it with the characteristics and error statistics. Frequency-invariant beam patterns are often required by systems using an array of sensors to process broadband signals. In some experimental conditions, the array spatial aperture is shorter than the involved wavelengths. In these conditions, superdirective beamforming is essential for an efficient system. I present a comparison between two methods that deal with a data-independent beamformer based on a filter-and-sum structure. Both methods (the first one numerical, the second one analytic) formulate a mathematical convex minimization problem, in which the variables to be optimized are the filters coefficients or frequency responses. In the described simulations, I have chosen a geometry and a set-up of parameters that allows us to make a fair comparison between the performances of the two different design methods analyzed. In particular, I addressed a small linear array for audio capture with different purposes (hearing aids, audio surveillance system, video-conference system, multimedia device, etc.). The research activity carried out has been used for the launch of a high-tech device through an innovative start-up in the field of glasses/audio devices (https://acoesis.com/en/). It has been proven that the proposed algorithm gives the possibility of obtaining higher performances than the state of the art of similar algorithms, additionally providing the possibility of connecting directivity or better generalized directivity to the statistics of phase errors and gain of sensors, extremely important in superdirective arrays in the case of real and industrial implementation. Therefore, the method proposed by the comparison is innovative because it quantitatively links the physical construction characteristics of the array to measurable and experimentally verifiable quantities, making the real implementation process controllable. The third topic faced is the reconstruction of the Room Impluse Response (RIR) using audio processing blind methods. Given an unknown audio source, the estimation of time differences-of-arrivals (TDOAs) can be efficiently and robustly solved using blind channel identification and exploiting the cross-correlation identity (CCI). Prior blind works have improved the estimate of TDOAs by means of different algorithmic solutions and optimization strategies, while always sticking to the case N = 2 microphones. But what if we can obtain a direct improvement in performance by just increasing N? In the fourth Chapter I tried to investigate this direction, showing that, despite the arguable simplicity, this is capable of (sharply) improving upon state-of-the-art blind channel identification methods based on CCI, without modifying the computational pipeline. Inspired by our results, we seek to warm up the community and the practitioners by paving the way (with two concrete, yet preliminary, examples) towards joint approaches in which advances in the optimization are combined with an increased number of microphones, in order to achieve further improvements. Sound source localisation applications can be tackled by inferring the time-difference-of-arrivals (TDOAs) between a sound-emitting source and a set of microphones. Among the referred applications, one can surely list room-aware sound reproduction, room geometry\u2019s estimation, speech enhancement. Despite a broad spectrum of prior works estimate TDOAs from a known audio source, even when the signal emitted from the acoustic source is unknown, TDOAs can be inferred by comparing the signals received at two (or more) spatially separated microphones, using the notion of cross-corrlation identity (CCI). This is the key theoretical tool, not only, to make the ordering of microphones irrelevant during the acquisition stage, but also to solve the problem as blind channel identification, robustly and reliably inferring TDOAs from an unknown audio source. However, when dealing with natural environments, such \u201cmutual agreement\u201d between microphones can be tampered by a variety of audio ambiguities such as ambient noise. Furthermore, each observed signal may contain multiple distorted or delayed replicas of the emitting source due to reflections or generic boundary effects related to the (closed) environment. Thus, robustly estimating TDOAs is surely a challenging problem and CCI-based approaches cast it as single-input/multi-output blind channel identification. Such methods promote robustness in the estimate from the methodological standpoint: using either energy-based regularization, sparsity or positivity constraints, while also pre-conditioning the solution space. Last but not least, the Acoustic Imaging is an imaging modality that exploits the propagation of acoustic waves in a medium to recover the spatial distribution and intensity of sound sources in a given region. Well known and widespread acoustic imaging applications are, for example, sonar and ultrasound. There are active and passive imaging devices: in the context of this thesis I consider a passive imaging system called Dual Cam that does not emit any sound but acquires it from the environment. In an acoustic image each pixel corresponds to the sound intensity of the source, the whose position is described by a particular pair of angles and, in the case in which the beamformer can, as in our case, work in near-field, from a distance on which the system is focused. In the last part of this work I propose the use of a new modality characterized by a richer information content, namely acoustic images, for the sake of audio-visual scene understanding. Each pixel in such images is characterized by a spectral signature, associated to a specific direction in space and obtained by processing the audio signals coming from an array of microphones. By coupling such array with a video camera, we obtain spatio-temporal alignment of acoustic images and video frames. This constitutes a powerful source of self-supervision, which can be exploited in the learning pipeline we are proposing, without resorting to expensive data annotations. However, since 2D planar arrays are cumbersome and not as widespread as ordinary microphones, we propose that the richer information content of acoustic images can be distilled, through a self-supervised learning scheme, into more powerful audio and visual feature representations. The learnt feature representations can then be employed for downstream tasks such as classification and cross-modal retrieval, without the need of a microphone array. To prove that, we introduce a novel multimodal dataset consisting in RGB videos, raw audio signals and acoustic images, aligned in space and synchronized in time. Experimental results demonstrate the validity of our hypothesis and the effectiveness of the proposed pipeline, also when tested for tasks and datasets different from those used for training. Chapter 6 closes the thesis, presenting a development activity of a new Dual Cam POC to build-up from it a spin-off, assuming to apply for an innovation project for hi-tech start- ups (such as a SME instrument H2020) for a 50Keuro grant, following the idea of the technology transfer. A deep analysis of the reference market, technologies and commercial competitors, business model and the FTO of intellectual property is then conducted. Finally, following the latest technological trends (https://www.flir.eu/products/si124/) a new version of the device (planar audio array) with reduced dimensions and improved technical characteristics is simulated, simpler and easier to use than the current one, opening up new interesting possibilities of development not only technical and scientific but also in terms of business fallout
    corecore