59,038 research outputs found

    Machine learning paradigms for modeling spatial and temporal information in multimedia data mining

    Get PDF
    Multimedia data mining and knowledge discovery is a fast emerging interdisciplinary applied research area. There is tremendous potential for effective use of multimedia data mining (MDM) through intelligent analysis. Diverse application areas are increasingly relying on multimedia under-standing systems. Advances in multimedia understanding are related directly to advances in signal processing, computer vision, machine learning, pattern recognition, multimedia databases, and smart sensors. The main mission of this special issue is to identify state-of-the-art machine learning paradigms that are particularly powerful and effective for modeling and combining temporal and spatial media cues such as audio, visual, and face information and for accomplishing tasks of multimedia data mining and knowledge discovery. These models should be able to bridge the gap between low-level audiovisual features which require signal processing and high-level semantics. A number of papers have been submitted to the special issue in the areas of imaging, artificial intelligence; and pattern recognition and five contributions have been selected covering state-of-the-art algorithms and advanced related topics. The first contribution by D. Xiang et al. “Evaluation of data quality and drought monitoring capability of FY-3A MERSI data” describes some basic parameters and major technical indicators of the FY-3A, and evaluates data quality and drought monitoring capability of the Medium-Resolution Imager (MERSI) onboard the FY-3A. The second contribution by A. Belatreche et al. “Computing with biologically inspired neural oscillators: application to color image segmentation” investigates the computing capabilities and potential applications of neural oscillators, a biologically inspired neural model, to gray scale and color image segmentation, an important task in image understanding and object recognition. The major contribution of this paper is the ability to use neural oscillators as a learning scheme for solving real world engineering problems. The third paper by A. Dargazany et al. entitled “Multibandwidth Kernel-based object tracking” explores new methods for object tracking using the mean shift (MS). A bandwidth-handling MS technique is deployed in which the tracker reach the global mode of the density function not requiring a specific staring point. It has been proven via experiments that the Gradual Multibandwidth Mean Shift tracking algorithm can converge faster than the conventional kernel-based object tracking (known as the mean shift). The fourth contribution by S. Alzu’bi et al. entitled “3D medical volume segmentation using hybrid multi-resolution statistical approaches” studies new 3D volume segmentation using multiresolution statistical approaches based on discrete wavelet transform and hidden Markov models. This system commonly reduced the percentage error achieved using the traditional 2D segmentation techniques by several percent. Furthermore, a contribution by G. Cabanes et al. entitled “Unsupervised topographic learning for spatiotemporal data mining” proposes a new unsupervised algorithm, suitable for the analysis of noisy spatiotemporal Radio Frequency Identification (RFID) data. The new unsupervised algorithm depicted in this article is an efficient data mining tool for behavioral studies based on RFID technology. It has the ability to discover and compare stable patterns in a RFID signal, and is appropriate for continuous learning. Finally, we would like to thank all those who helped to make this special issue possible, especially the authors and the reviewers of the articles. Our thanks go to the Hindawi staff and personnel, the journal Manager in bringing about the issue and giving us the opportunity to edit this special issue

    SciTech News Volume 70, No. 4 (2016)

    Get PDF
    Columns and Reports From the Editor 3 Division News Science-Technology Division 4 SLA Annual Meeting 2016 Report (S. Kirk Cabeen Travel Stipend Award recipient) 6 Reflections on SLA Annual Meeting (Diane K. Foster International Student Travel Award recipient) 8 SLA Annual Meeting Report (Bonnie Hilditch International Librarian Award recipient)10 Chemistry Division 12 Engineering Division 15 Reflections from the 2016 SLA Conference (SPIE Digital Library Student Travel Stipend recipient)15 Fundamentals of Knowledge Management and Knowledge Services (IEEE Continuing Education Stipend recipient) 17 Makerspaces in Libraries: The Big Table, the Art Studio or Something Else? (by Jeremy Cusker) 19 Aerospace Section of the Engineering Division 21 Reviews Sci-Tech Book News Reviews 22 Advertisements IEEE 17 WeBuyBooks.net 2

    Structural Change Can Be Detected in Advanced-Glaucoma Eyes.

    Get PDF
    PurposeTo compare spectral-domain optical coherence tomography (SD-OCT) standard structural measures and a new three-dimensional (3D) volume optic nerve head (ONH) change detection method for detecting change over time in severely advanced-glaucoma (open-angle glaucoma [OAG]) patients.MethodsThirty-five eyes of 35 patients with very advanced glaucoma (defined as a visual field mean deviation < -21 dB) and 46 eyes of 30 healthy subjects to estimate aging changes were included. Circumpapillary retinal fiber layer thickness (cpRNFL), minimum rim width (MRW), and macular retinal ganglion cell-inner plexiform layer (GCIPL) thicknesses were measured using the San Diego Automated Layer Segmentation Algorithm (SALSA). Progression was defined as structural loss faster than 95th percentile of healthy eyes. Three-dimensional volume ONH change was estimated using the Bayesian-kernel detection scheme (BKDS), which does not require extensive retinal layer segmentation.ResultsThe number of progressing glaucoma eyes identified was highest for 3D volume BKDS (13, 37%), followed by GCPIL (11, 31%), cpRNFL (4, 11%), and MRW (2, 6%). In advanced-OAG eyes, only the mean rate of GCIPL change reached statistical significance, -0.18 ÎŒm/y (P = 0.02); the mean rates of cpRNFL and MRW change were not statistically different from zero. In healthy eyes, the mean rates of cpRNFL, MRW, and GCIPL change were significantly different from zero. (all P < 0.001).ConclusionsGanglion cell-inner plexiform layer and 3D volume BKDS show promise for identifying change in severely advanced glaucoma. These results suggest that structural change can be detected in very advanced disease. Longer follow-up is needed to determine whether changes identified are false positives or true progression

    Introduction for the Special Issue on Beyond the Hypes of Geospatial Big Data: Theories, Methods, Analytics, and Applications

    Get PDF
    We live in the era of ‘Big Data’. In particular, Geospatial data, whether captured through remote sensors (e.g., satellite imagery) or generated from large-scale simulations (e.g., climate change models) have always been significantly large in size. Over the last decade however, advances in instrumentation and computation has seen the volume, variety, velocity, and veracity of this data increase exponentially. Of the 2.5 quintillion (1018) bytes of data that are generated on a daily basis across the globe, a large portion (arguably as much as 80%) is found to be geo-referenced. Therefore, this special issue is dedicated to the innovative theories, methods, analytics, and applications of geospatial big data

    Feedforward data-aided phase noise estimation from a DCT basis expansion

    Get PDF
    This contribution deals with phase noise estimation from pilot symbols. The phase noise process is approximated by an expansion of discrete cosine transform (DCT) basis functions containing only a few terms. We propose a feedforward algorithm that estimates the DCT coefficients without requiring detailed knowledge about the phase noise statistics. We demonstrate that the resulting (linearized) mean-square phase estimation error consists of two contributions: a contribution from the additive noise, that equals the Cramer-Rao lower bound, and a noise independent contribution, that results front the phase noise modeling error. We investigate the effect of the symbol sequence length, the pilot symbol positions, the number of pilot symbols, and the number of estimated DCT coefficients it the estimation accuracy and on the corresponding bit error rate (PER). We propose a pilot symbol configuration allowing to estimate any number of DCT coefficients not exceeding the number of pilot Symbols, providing a considerable Performance improvement as compared to other pilot symbol configurations. For large block sizes, the DCT-based estimation algorithm substantially outperforms algorithms that estimate only the time-average or the linear trend of the carrier phase. Copyright (C) 2009 J. Bhatti and M. Moeneclaey

    Towards Automatic Speech Identification from Vocal Tract Shape Dynamics in Real-time MRI

    Full text link
    Vocal tract configurations play a vital role in generating distinguishable speech sounds, by modulating the airflow and creating different resonant cavities in speech production. They contain abundant information that can be utilized to better understand the underlying speech production mechanism. As a step towards automatic mapping of vocal tract shape geometry to acoustics, this paper employs effective video action recognition techniques, like Long-term Recurrent Convolutional Networks (LRCN) models, to identify different vowel-consonant-vowel (VCV) sequences from dynamic shaping of the vocal tract. Such a model typically combines a CNN based deep hierarchical visual feature extractor with Recurrent Networks, that ideally makes the network spatio-temporally deep enough to learn the sequential dynamics of a short video clip for video classification tasks. We use a database consisting of 2D real-time MRI of vocal tract shaping during VCV utterances by 17 speakers. The comparative performances of this class of algorithms under various parameter settings and for various classification tasks are discussed. Interestingly, the results show a marked difference in the model performance in the context of speech classification with respect to generic sequence or video classification tasks.Comment: To appear in the INTERSPEECH 2018 Proceeding
    • 

    corecore