1,380 research outputs found

    Requirements for tracking radar for falling spheres

    Get PDF
    Error analysis on radar tracking of falling sphere

    Spatial patterns in timing of the diurnal temperature cycle

    Get PDF
    This paper investigates the structural difference in timing of the diurnal temperature cycle (DTC) over land resulting from choice of measuring device or model framework. It is shown that the timing can be reliably estimated from temporally sparse observations acquired from a constellation of low Earth-orbiting satellites given record lengths of at least three months. Based on a year of data, the spatial patterns of mean DTC timing are compared between temperature estimates from microwave Ka-band, geostationary thermal infrared (TIR), and numerical weather prediction model output from the Global Modeling and Assimilation Office (GMAO). It is found that the spatial patterns can be explained by vegetation effects, sensing depth differences and more speculatively the orientation of orographic relief features. In absolute terms, the GMAO model puts the peak of the DTC on average at 12:50 local solar time, 23 min before TIR with a peak temperature at 13:13 (both averaged over Africa and Europe). Since TIR is the shallowest observation of the land surface, this small difference represents a structural error that possibly affects the model's ability to assimilate observations that are closely tied to the DTC. The equivalent average timing for Ka-band is 13:44, which is influenced by the effect of increased sensing depth in desert areas. For non-desert areas, the Ka-band observations lag the TIR observations by only 15 min, which is in agreement with their respective theoretical sensing depth. The results of this comparison provide insights into the structural differences between temperature measurements and models, and can be used as a first step to account for these differences in a coherent way

    Att-TasNet: attending to encodings in time-domain audio speech separation of noisy, reverberant speech mixtures

    Get PDF
    Separation of speech mixtures in noisy and reverberant environments remains a challenging task for state-of-the-art speech separation systems. Time-domain audio speech separation networks (TasNets) are among the most commonly used network architectures for this task. TasNet models have demonstrated strong performance on typical speech separation baselines where speech is not contaminated with noise. When additive or convolutive noise is present, performance of speech separation degrades significantly. TasNets are typically constructed of an encoder network, a mask estimation network and a decoder network. The design of these networks puts the majority of the onus for enhancing the signal on the mask estimation network when used without any pre-processing of the input data or post processing of the separation network output data. Use of multihead attention (MHA) is proposed in this work as an additional layer in the encoder and decoder to help the separation network attend to encoded features that are relevant to the target speakers and conversely suppress noisy disturbances in the encoded features. As shown in this work, incorporating MHA mechanisms into the encoder network in particular leads to a consistent performance improvement across numerous quality and intelligibility metrics on a variety of acoustic conditions using the WHAMR corpus, a data-set of noisy reverberant speech mixtures. The use of MHA is also investigated in the decoder network where it is demonstrated that smaller performance improvements are consistently gained within specific model configurations. The best performing MHA models yield a mean 0.6 dB scale invariant signal-to-distortion (SISDR) improvement on noisy reverberant mixtures over a baseline 1D convolution encoder. A mean 1 dB SISDR improvement is observed on clean speech mixtures

    The University of Sheffield CHiME-7 UDASE challenge speech enhancement system

    Get PDF
    The CHiME-7 unsupervised domain adaptation speech enhancement (UDASE) challenge targets domain adaptation to unlabelled speech data. This paper describes the University of Sheffield team’s system submitted to the challenge. A generative adversarial network (GAN) methodology based on a conformer-based metric GAN (CMGAN) is employed as opposed to the unsupervised RemixIT strategy used in the CHiME-7 baseline system. The discriminator of the GAN is trained to predict the output score of a Deep Noise Suppression Mean Opinion Score (DNSMOS) metric. Additional data augmentation strategies are employed which provide the discriminator with historical training data outputs as well as more diverse training examples from an additional pseudo-generator. The proposed approach, denoted as CMGAN+/+, achieves significant improvement in DNSMOS evaluation metrics with the best proposed system achieving 3.51 OVR-MOS, a 24% improvement over the baseline

    Perceive and predict: self-supervised speech representation based loss functions for speech enhancement

    Get PDF
    Recent work in the domain of speech enhancement has explored the use of self-supervised speech representations to aid in the training of neural speech enhancement models. However, much of this work focuses on using the deepest or final outputs of self supervised speech representation models, rather than the earlier feature encodings. The use of self supervised representations in such a way is often not fully motivated. In this work it is shown that the distance between the feature encodings of clean and noisy speech correlate strongly with psychoacoustically motivated measures of speech quality and intelligibility, as well as with human Mean Opinion Score (MOS) ratings. Experiments using this distance as a loss function are performed and improved performance over the use of STFT spectrogram distance based loss as well as other common loss functions from speech enhancement literature is demonstrated using objective measures such as perceptual evaluation of speech quality (PESQ) and short-time objective intelligibility (STOI)

    The Ursinus Weekly, December 6, 1954

    Get PDF
    West Chester choir sings at vespers • Does Ursinus have adequate parking facilities? • Candlelight Communion to be held Dec. 9 in Bomberger • Improper procedure key in MSGA trial • Rosicrucians add members at after-dinner dessert • Dr. Dugger speaks, shows slides to pre-medical group • Christmas to visit UC in many forms • Beta Sig presents Bill Haley on Jan. 7 • Charles Hudnut wins award in national poetry contest • Seniors hold prom; Elect lord and lady • 300 attend seventeenth annual Messiah performance • Chesterfield holds contest Home for the holidays • Dr. Oliver Gogarty discusses poets • Band practicing for May Day; Marches at basketball games • Editorials • Test of time • Matmen boast 7 lettermen; Two MAC champions return • Gridmen elect Neborak MVP, 1955 captain • Heller, Bauser All Philadelphia 3rd hockey team • Susquehanna, Nat. Aggies bow; Juniata mars record by 78-56 • Westerhoff\u27s proposed revisions passed by MSGAhttps://digitalcommons.ursinus.edu/weekly/1462/thumbnail.jp

    The USFD Spoken Language Translation System for IWSLT 2014

    Get PDF
    The University of Sheffield (USFD) participated in the International Workshop for Spoken Language Translation (IWSLT) in 2014. In this paper, we will introduce the USFD SLT system for IWSLT. Automatic speech recognition (ASR) is achieved by two multi-pass deep neural network systems with adaptation and rescoring techniques. Machine translation (MT) is achieved by a phrase-based system. The USFD primary system incorporates state-of-the-art ASR and MT techniques and gives a BLEU score of 23.45 and 14.75 on the English-to-French and English-to-German speech-to-text translation task with the IWSLT 2014 data. The USFD contrastive systems explore the integration of ASR and MT by using a quality estimation system to rescore the ASR outputs, optimising towards better translation. This gives a further 0.54 and 0.26 BLEU improvement respectively on the IWSLT 2012 and 2014 evaluation data
    corecore