2,837 research outputs found

    Underwater Localization in a Confined Space Using Acoustic Positioning and Machine Learning

    Get PDF
    Localization is a critical step in any navigation system. Through localization, the vehicle can estimate its position in the surrounding environment and plan how to reach its goal without any collision. This thesis focuses on underwater source localization, using sound signals for position estimation. We propose a novel underwater localization method based on machine learning techniques in which source position is directly estimated from collected acoustic data. The position of the sound source is estimated by training Random Forest (RF), Support Vector Machine (SVM), Feedforward Neural Network (FNN), and Convolutional Neural Network (CNN). To train these data-driven methods, data are collected inside a confined test tank with dimensions of 6m x 4.5m x 1.7m. The transmission unit, which includes Xilinx LX45 FPGA and transducer, generates acoustic signal. The receiver unit collects and prepares propagated sound signals and transmit them to a computer. It consists of 4 hydrophones, Red Pitay analog front-end board, and NI 9234 data acquisition board. We used MATLAB 2018 to extract pitch, Mel-Frequency Cepstrum Coefficients (MFCC), and spectrogram from the sound signals. These features are used by MATLAB Toolboxes to train RF, SVM, FNN, and CNN. Experimental results show that CNN archives 4% of Mean Absolute Percentage Error (MAPE) in the test tank. The finding of this research can pave the way for Autonomous Underwater Vehicle (AUV) and Remotely Operated Vehicle (ROV) navigation in underwater open spaces

    INSPIRE Newsletter Fall 2022

    Get PDF
    https://scholarsmine.mst.edu/inspire-newsletters/1011/thumbnail.jp

    Confusion modelling for lip-reading

    Get PDF
    Lip-reading is mostly used as a means of communication by people with hearing di�fficulties. Recent work has explored the automation of this process, with the aim of building a speech recognition system entirely driven by lip movements. However, this work has so far produced poor results because of factors such as high variability of speaker features, diffi�culties in mapping from visual features to speech sounds, and high co-articulation of visual features. The motivation for the work in this thesis is inspired by previous work in dysarthric speech recognition [Morales, 2009]. Dysathric speakers have poor control over their articulators, often leading to a reduced phonemic repertoire. The premise of this thesis is that recognition of the visual speech signal is a similar problem to recog- nition of dysarthric speech, in that some information about the speech signal has been lost in both cases, and this brings about a systematic pattern of errors in the decoded output. This work attempts to exploit the systematic nature of these errors by modelling them in the framework of a weighted finite-state transducer cascade. Results indicate that the technique can achieve slightly lower error rates than the conventional approach. In addition, it explores some interesting more general questions for automated lip-reading

    GiD 2008. 4th Conference on advances and applications of GiD

    Get PDF
    The extended use of simulation programs has leaned on the advances in user-friendly interfaces and in the capability to generate meshes for any generic complex geometry. More than ten years of development have made Gid grow to become one of the more popular pre ans postprocessing systems at international level. The constant dialogue between the GiD development team and the users has guided the development of giD to cover the pre-post needs of many disciplines in science and engineering. Following gthis philosophy, the biannual GiD Conference has become an important forum for discussion and interchange of experiences among the GiD community. This monograph includes the contributions of the participants to the fourth edition of the GiD Conference held in the island of Ibiza from 8-9 May 2008

    Computer-aided image quality assessment in automated 3D breast ultrasound images

    Get PDF
    Automated 3D breast ultrasound (ABUS) is a valuable, non-ionising adjunct to X-ray mammography for breast cancer screening and diagnosis for women with dense breasts. High image quality is an important prerequisite for diagnosis and has to be guaranteed at the time of acquisition. The high throughput of images in a screening scenario demands for automated solutions. In this work, an automated image quality assessment system rating ABUS scans at the time of acquisition was designed and implemented. Quality assessment of present diagnostic ultrasound images has rarely been performed demanding thorough analysis of potential image quality aspects in ABUS. Therefore, a reader study was initiated, making two clinicians rate the quality of clinical ABUS images. The frequency of specific quality aspects was evaluated revealing that incorrect positioning and insufficiently applied contact fluid caused the most relevant image quality issues. The relative position of the nipple in the image, the acoustic shadow caused by the nipple as well as the shape of the breast contour reflect patient positioning and ultrasound transducer handling. Morphological and histogram-based features utilized for machine learning to reproduce the manual classification as provided by the clinicians. At 97 % specificity, the automatic classification achieved sensitivities of 59 %, 45 %, and 46 % for the three aforementioned aspects, respectively. The nipple is an important landmark in breast imaging, which is generally---but not always correctly---pinpointed by the technicians. An existing nipple detection algorithm was extended by probabilistic atlases and exploited for automatic detection of incorrectly annotated nipple marks. The nipple detection rate was increased from 82 % to 85 % and the classification achieved 90 % sensitivity at 89 % specificity. A lack of contact fluid between transducer and skin can induce reverberation patterns and acoustic shadows, which can possibly obscure lesions. Parameter maps were computed in order to localize these artefact regions and yielded a detection rate of 83 % at 2.6 false positives per image. Parts of the presented work were integrated to clinical workflow making up a novel image quality assessment system that supported technicians in their daily routine by detecting images of insufficient quality and indicating potential improvements for a repeated scan while the patient was still in the examination room. First evaluations showed that the proposed method sensitises technicians for the radiologists' demands on diagnostically valuable images

    Deciphering Speech: a Zero-Resource Approach to Cross-Lingual Transfer in ASR

    Get PDF
    We present a method for cross-lingual training an ASR system using absolutely no transcribed training data from the target language, and with no phonetic knowledge of the language in question. Our approach uses a novel application of a decipherment algorithm, which operates given only unpaired speech and text data from the target language. We apply this decipherment to phone sequences generated by a universal phone recogniser trained on out-of-language speech corpora, which we follow with flat-start semi-supervised training to obtain an acoustic model for the new language. To the best of our knowledge, this is the first practical approach to zero-resource cross-lingual ASR which does not rely on any hand-crafted phonetic information. We carry out experiments on read speech from the GlobalPhone corpus, and show that it is possible to learn a decipherment model on just 20 minutes of data from the target language. When used to generate pseudo-labels for semi-supervised training, we obtain WERs that range from 32.5% to just 1.9% absolute worse than the equivalent fully supervised models trained on the same data.Comment: Submitted to Interspeech 202

    Deciphering Speech: a Zero-Resource Approach to Cross-Lingual Transfer in ASR

    Get PDF

    Advancing Climate Change Research and Hydrocarbon Leak Detection : by Combining Dissolved Carbon Dioxide and Methane Measurements with ADCP Data

    Get PDF
    With the emergence of largescale, comprehensive environmental monitoring projects, there is an increased need to combine state-of-the art technologies to address complicated problems such as ocean acidifi cation and hydrocarbon leak detection
    • …
    corecore