255,089 research outputs found

    Channel coding for progressive images in a 2-D time-frequency OFDM block with channel estimation errors.

    Get PDF
    Coding and diversity are very effective techniques for improving transmission reliability in a mobile wireless environment. The use of diversity is particularly important for multimedia communications over fading channels. In this work, we study the transmission of progressive image bitstreams using channel coding in a 2-D time-frequency resource block in an OFDM network, employing time and frequency diversities simultaneously. In particular, in the frequency domain, based on the order of diversity and the correlation of individual subcarriers, we construct symmetric n -channel FEC-based multiple descriptions using channel erasure codes combined with embedded image coding. In the time domain, a concatenation of RCPC codes and CRC codes is employed to protect individual descriptions. We consider the physical channel conditions arising from various coherence bandwidths and coherence times, leading to a range of orders of diversities available in the time and frequency domains. We investigate the effects of different error patterns on the delivered image quality due to various fade rates. We also study the tradeoffs and compare the relative effectiveness associated with the use of erasure codes in the frequency domain and convolutional codes in the time domain under different physical environments. Both the effects of intercarrier interference and channel estimation errors are included in our study. Specifically, the effects of channel estimation errors, frequency selectivity and the rate of the channel variations are taken into consideration for the construction of the 2-D time-frequency block. We provide results showing the gain that the proposed model achieves compared to a system without temporal coding. In one example, for a system experiencing flat fading, low Doppler, and imperfect CSI, we find that the increase in PSNR compared to a system without time diversity is as much as 9.4 dB

    On the robustness of model-based algorithms for photoacoustic tomography: Comparison between time and frequency domains

    Get PDF
    For photoacoustic image reconstruction, certain parameters such as sensor positions and speed of sound have a major impact on the reconstruction process and must be carefully determined before data acquisition. Uncertainties in these parameters can lead to errors produced by a modeling mismatch, hindering the reconstruction process and severely affecting the resulting image quality. Therefore, in this work, we study how modeling errors arising from uncertainty in sensor locations affect the images obtained by matrix model-based reconstruction algorithms based on time domain and frequency domain models of the photoacoustic problem. The effects on the reconstruction performance with respect to the uncertainty in the knowledge of the sensors location are compared and analyzed both in a qualitative and quantitative fashion for both time and frequency models. Ultimately, our study shows that the frequency domain approach is more sensitive to this kind of modeling errors. These conclusions are supported by numerical experiments and a theoretical sensitivity analysis of the mathematical operator for the direct problem.Fil: Hirsch, Lucas. Universidad de Buenos Aires. Facultad de Ingeniería. Departamento de Física; ArgentinaFil: González, Martín Germán. Consejo Nacional de Investigaciones Científicas y Técnicas; Argentina. Universidad de Buenos Aires. Facultad de Ingeniería. Departamento de Física; ArgentinaFil: Rey Vega, Leonardo Javier. Consejo Nacional de Investigaciones Científicas y Técnicas; Argentina. Universidad de Buenos Aires. Facultad de Ingeniería. Departamento de Física; Argentin

    Holistic Dynamic Frequency Transformer for Image Fusion and Exposure Correction

    Full text link
    The correction of exposure-related issues is a pivotal component in enhancing the quality of images, offering substantial implications for various computer vision tasks. Historically, most methodologies have predominantly utilized spatial domain recovery, offering limited consideration to the potentialities of the frequency domain. Additionally, there has been a lack of a unified perspective towards low-light enhancement, exposure correction, and multi-exposure fusion, complicating and impeding the optimization of image processing. In response to these challenges, this paper proposes a novel methodology that leverages the frequency domain to improve and unify the handling of exposure correction tasks. Our method introduces Holistic Frequency Attention and Dynamic Frequency Feed-Forward Network, which replace conventional correlation computation in the spatial-domain. They form a foundational building block that facilitates a U-shaped Holistic Dynamic Frequency Transformer as a filter to extract global information and dynamically select important frequency bands for image restoration. Complementing this, we employ a Laplacian pyramid to decompose images into distinct frequency bands, followed by multiple restorers, each tuned to recover specific frequency-band information. The pyramid fusion allows a more detailed and nuanced image restoration process. Ultimately, our structure unifies the three tasks of low-light enhancement, exposure correction, and multi-exposure fusion, enabling comprehensive treatment of all classical exposure errors. Benchmarking on mainstream datasets for these tasks, our proposed method achieves state-of-the-art results, paving the way for more sophisticated and unified solutions in exposure correction

    Concept for collision avoidance in machine tools based on geometric simulation and sensor data

    Get PDF
    Collisions are a major cause of unplanned downtime in small series manufacturing with machine tools. Existing solutions based on geometric simulation do not cover collisions due to setup errors. Therefore a concept is developed to enable a sensor-based matching of the setup with the simulation, thus detecting discrepancies. Image processing in the spatial and frequency domain is used to compensate for harsh conditions in the machine, including swarf, fluids and suboptimal illumination

    Joint Multi-Pitch Detection Using Harmonic Envelope Estimation for Polyphonic Music Transcription

    Get PDF
    In this paper, a method for automatic transcription of music signals based on joint multiple-F0 estimation is proposed. As a time-frequency representation, the constant-Q resonator time-frequency image is employed, while a novel noise suppression technique based on pink noise assumption is applied in a preprocessing step. In the multiple-F0 estimation stage, the optimal tuning and inharmonicity parameters are computed and a salience function is proposed in order to select pitch candidates. For each pitch candidate combination, an overlapping partial treatment procedure is used, which is based on a novel spectral envelope estimation procedure for the log-frequency domain, in order to compute the harmonic envelope of candidate pitches. In order to select the optimal pitch combination for each time frame, a score function is proposed which combines spectral and temporal characteristics of the candidate pitches and also aims to suppress harmonic errors. For postprocessing, hidden Markov models (HMMs) and conditional random fields (CRFs) trained on MIDI data are employed, in order to boost transcription accuracy. The system was trained on isolated piano sounds from the MAPS database and was tested on classic and jazz recordings from the RWC database, as well as on recordings from a Disklavier piano. A comparison with several state-of-the-art systems is provided using a variety of error metrics, where encouraging results are indicated

    Wavefront Control and Image Restoration with Less Computing

    Get PDF
    PseudoDiversity is a method of recovering the wavefront in a sparse- or segmented- aperture optical system typified by an interferometer or a telescope equipped with an adaptive primary mirror consisting of controllably slightly moveable segments. (PseudoDiversity should not be confused with a radio-antenna-arraying method called pseudodiversity.) As in the cases of other wavefront- recovery methods, the streams of wavefront data generated by means of PseudoDiversity are used as feedback signals for controlling electromechanical actuators of the various segments so as to correct wavefront errors and thereby, for example, obtain a clearer, steadier image of a distant object in the presence of atmospheric turbulence. There are numerous potential applications in astronomy, remote sensing from aircraft and spacecraft, targeting missiles, sighting military targets, and medical imaging (including microscopy) through such intervening media as cells or water. In comparison with prior wavefront-recovery methods used in adaptive optics, PseudoDiversity involves considerably simpler equipment and procedures and less computation. For PseudoDiversity, there is no need to install separate metrological equipment or to use any optomechanical components beyond those that are already parts of the optical system to which the method is applied. In Pseudo- Diversity, the actuators of a subset of the segments or subapertures are driven to make the segments dither in the piston, tilt, and tip degrees of freedom. Each aperture is dithered at a unique frequency at an amplitude of a half wavelength of light. During the dithering, images on the focal plane are detected and digitized at a rate of at least four samples per dither period. In the processing of the image samples, the use of different dither frequencies makes it possible to determine the separate effects of the various dithered segments or apertures. The digitized image-detector outputs are processed in the spatial-frequency (Fourier-transform) domain to obtain measures of the piston, tip, and tilt errors over each segment or subaperture. Once these measures are known, they are fed back to the actuators to correct the errors. In addition, measures of errors that remain after correction by use of the actuators are further utilized in an algorithm in which the image is phase-corrected in the spatial-frequency domain and then transformed back to the spatial domain at each time step and summed with the images from all previous time steps to obtain a final image having a greater signal-to-noise ratio (and, hence, a visual quality) higher than would otherwise be attainable

    Zero-shot Medical Image Translation via Frequency-Guided Diffusion Models

    Full text link
    Recently, the diffusion model has emerged as a superior generative model that can produce high quality and realistic images. However, for medical image translation, the existing diffusion models are deficient in accurately retaining structural information since the structure details of source domain images are lost during the forward diffusion process and cannot be fully recovered through learned reverse diffusion, while the integrity of anatomical structures is extremely important in medical images. For instance, errors in image translation may distort, shift, or even remove structures and tumors, leading to incorrect diagnosis and inadequate treatments. Training and conditioning diffusion models using paired source and target images with matching anatomy can help. However, such paired data are very difficult and costly to obtain, and may also reduce the robustness of the developed model to out-of-distribution testing data. We propose a frequency-guided diffusion model (FGDM) that employs frequency-domain filters to guide the diffusion model for structure-preserving image translation. Based on its design, FGDM allows zero-shot learning, as it can be trained solely on the data from the target domain, and used directly for source-to-target domain translation without any exposure to the source-domain data during training. We evaluated it on three cone-beam CT (CBCT)-to-CT translation tasks for different anatomical sites, and a cross-institutional MR imaging translation task. FGDM outperformed the state-of-the-art methods (GAN-based, VAE-based, and diffusion-based) in metrics of Frechet Inception Distance (FID), Peak Signal-to-Noise Ratio (PSNR), and Structural Similarity Index Measure (SSIM), showing its significant advantages in zero-shot medical image translation

    Fast Stochastic Wiener Filter for Super-Resolution Image Restoration with Information Theoretic Visual Quality Assessment

    Get PDF
    Super-resolution (SR) refers to reconstructing a single high resolution (HR) image from a set of subsampled, blurred and noisy low resolution (LR) images. The reconstructed image suffers from degradations such as blur, aliasing, photo-detector noise and registration and fusion error. Wiener filter can be used to remove artifacts and enhance the visual quality of the reconstructed images. In this paper, we introduce a new fast stochastic Wiener filter for SR reconstruction and restoration that can be implemented efficiently in the frequency domain. Our derivation depends on the continuous-discrete-continuous (CDC) model that represents most of the degradations encountered during the image-gathering and image-display processes. We incorporate a new parameter that accounts for LR images registration and fusion errors. Also, we speeded up the performance of the filter by constraining it to work on small patches of the images. Beside this, we introduce two figures of merits: information rate and maximum realizable fidelity, which can be used to assess the visual quality of the resultant images. Simulations and experimental results demonstrate that the derived Wiener filter that can be implemented efficiently in the frequency domain can reduce aliasing, blurring, and noise and result in a sharper reconstructed image. Also, Quantitative assessment using the proposed figures coincides with the visual qualitative assessment. Finally, we evaluate our filter against other SR techniques and its results were very competitive

    Homography-Based Correction of Positional Errors in MRT Survey

    Full text link
    The Mauritius Radio Telescope (MRT) images show systematics in the positional errors of sources when compared to source positions in the Molonglo Reference Catalogue (MRC). We have applied two-dimensional homography to correct positional errors in the image domain and avoid re-processing the visibility data. Positions of bright (above 15-σ\sigma) sources, common to MRT and MRC catalogues, are used to set up an over-determined system to solve for the 2-D homography matrix. After correction, the errors are found to be within 10% of the beamwidth for these bright sources and the systematics are eliminated from the images.Comment: 4 pages, 4 figures, The Low-Frequency Radio Universe, Proceedings of a conference held at NCRA-TIFR, Pune, 8-12 December 2008, ASP Conference Series, Vol. 407, 2009, Eds: D.J. Saikia, D.A. Green, Y. Gupta and T. Ventur
    corecore