3,042 research outputs found

    Unsupervised extraction of recurring words from infant-directed speech

    Get PDF
    To date, most computational models of infant word segmentation have worked from phonemic or phonetic input, or have used toy datasets. In this paper, we present an algorithm for word extraction that works directly from naturalistic acoustic input: infant-directed speech from the CHILDES corpus. The algorithm identifies recurring acoustic patterns that are candidates for identification as words or phrases, and then clusters together the most similar patterns. The recurring patterns are found in a single pass through the corpus using an incremental method, where only a small number of utterances are considered at once. Despite this limitation, we show that the algorithm is able to extract a number of recurring words, including some that infants learn earliest, such as Mommy and the child’s name. We also introduce a novel information-theoretic evaluation measure

    Fast and easy blind deblurring using an inverse filter and PROBE

    Full text link
    PROBE (Progressive Removal of Blur Residual) is a recursive framework for blind deblurring. Using the elementary modified inverse filter at its core, PROBE's experimental performance meets or exceeds the state of the art, both visually and quantitatively. Remarkably, PROBE lends itself to analysis that reveals its convergence properties. PROBE is motivated by recent ideas on progressive blind deblurring, but breaks away from previous research by its simplicity, speed, performance and potential for analysis. PROBE is neither a functional minimization approach, nor an open-loop sequential method (blur kernel estimation followed by non-blind deblurring). PROBE is a feedback scheme, deriving its unique strength from the closed-loop architecture rather than from the accuracy of its algorithmic components

    FAST: A multi-processed environment for visualization of computational fluid dynamics

    Get PDF
    Three-dimensional, unsteady, multi-zoned fluid dynamics simulations over full scale aircraft are typical of the problems being investigated at NASA Ames' Numerical Aerodynamic Simulation (NAS) facility on CRAY2 and CRAY-YMP supercomputers. With multiple processor workstations available in the 10-30 Mflop range, we feel that these new developments in scientific computing warrant a new approach to the design and implementation of analysis tools. These larger, more complex problems create a need for new visualization techniques not possible with the existing software or systems available as of this writing. The visualization techniques will change as the supercomputing environment, and hence the scientific methods employed, evolves even further. The Flow Analysis Software Toolkit (FAST), an implementation of a software system for fluid mechanics analysis, is discussed

    Scientific Visualization Using the Flow Analysis Software Toolkit (FAST)

    Get PDF
    Over the past few years the Flow Analysis Software Toolkit (FAST) has matured into a useful tool for visualizing and analyzing scientific data on high-performance graphics workstations. Originally designed for visualizing the results of fluid dynamics research, FAST has demonstrated its flexibility by being used in several other areas of scientific research. These research areas include earth and space sciences, acid rain and ozone modelling, and automotive design, just to name a few. This paper describes the current status of FAST, including the basic concepts, architecture, existing functionality and features, and some of the known applications for which FAST is being used. A few of the applications, by both NASA and non-NASA agencies, are outlined in more detail. Described in the Outlines are the goals of each visualization project, the techniques or 'tricks' used lo produce the desired results, and custom modifications to FAST, if any, done to further enhance the analysis. Some of the future directions for FAST are also described

    Learning object categories from Google's image search

    Get PDF
    Current approaches to object category recognition require datasets of training images to be manually prepared, with varying degrees of supervision. We present an approach that can learn an object category from just its name, by utilizing the raw output of image search engines available on the Internet. We develop a new model, TSI-pLSA, which extends pLSA (as applied to visual words) to include spatial information in a translation and scale invariant manner. Our approach can handle the high intra-class variability and large proportion of unrelated images returned by search engines. We evaluate the models on standard test sets, showing performance competitive with existing methods trained on hand prepared datasets

    FAST: A multi-processed environment for visualization of computational fluid

    Get PDF
    Three dimensional, unsteady, multizoned fluid dynamics simulations over full scale aircraft is typical of problems being computed at NASA-Ames on CRAY2 and CRAY-YMP supercomputers. With multiple processor workstations available in the 10 to 30 Mflop range, it is felt that these new developments in scientific computing warrant a new approach to the design and implementation of analysis tools. These large, more complex problems create a need for new visualization techniques not possible with the existing software or systems available as of this time. These visualization techniques will change as the supercomputing environment, and hence the scientific methods used, evolve ever further. Visualization of computational aerodynamics require flexible, extensible, and adaptable software tools for performing analysis tasks. FAST (Flow Analysis Software Toolkit), an implementation of a software system for fluid mechanics analysis that is based on this approach is discussed

    Motion Deblurring in the Wild

    Full text link
    The task of image deblurring is a very ill-posed problem as both the image and the blur are unknown. Moreover, when pictures are taken in the wild, this task becomes even more challenging due to the blur varying spatially and the occlusions between the object. Due to the complexity of the general image model we propose a novel convolutional network architecture which directly generates the sharp image.This network is built in three stages, and exploits the benefits of pyramid schemes often used in blind deconvolution. One of the main difficulties in training such a network is to design a suitable dataset. While useful data can be obtained by synthetically blurring a collection of images, more realistic data must be collected in the wild. To obtain such data we use a high frame rate video camera and keep one frame as the sharp image and frame average as the corresponding blurred image. We show that this realistic dataset is key in achieving state-of-the-art performance and dealing with occlusions

    Artificial Intelligence for Detecting Preterm Uterine Activity in Gynacology and Obstertric Care

    Get PDF
    Preterm birth brings considerable emotional and economic costs to families and society. However, despite extensive research into understanding the risk factors, the prediction of patient mechanisms and improvements to obstetrical practice, the UK National Health Service still annually spends more than £2.95 billion on this issue. Diagnosis of labour in normal pregnancies is important for minimizing unnecessary hospitalisations, interventions and expenses. Moreover, accurate identification of spontaneous preterm labour would also allow clinicians to start necessary treatments early in women with true labour and avert unnecessary treatment and hospitalisation for women who are simply having preterm contractions, but who are not in true labour. In this research, the Electrohysterography signals have been used to detect preterm births, because Electrohysterography signals provide a strong basis for objective prediction and diagnosis of preterm birth. This has been achieved using an open dataset, which contains 262 records for women who delivered at term and 38 who delivered prematurely. Three different machine learning algorithm were used to identify these records. The results illustrate that the Random Forest performed the best of sensitivity 97%, specificity of 85%, Area under the Receiver Operator curve (AUROC) of 94% and mean square error rate of 14%
    corecore