812,149 research outputs found
Satellite image analysis using neural networks
The tremendous backlog of unanalyzed satellite data necessitates the development of improved methods for data cataloging and analysis. Ford Aerospace has developed an image analysis system, SIANN (Satellite Image Analysis using Neural Networks) that integrates the technologies necessary to satisfy NASA's science data analysis requirements for the next generation of satellites. SIANN will enable scientists to train a neural network to recognize image data containing scenes of interest and then rapidly search data archives for all such images. The approach combines conventional image processing technology with recent advances in neural networks to provide improved classification capabilities. SIANN allows users to proceed through a four step process of image classification: filtering and enhancement, creation of neural network training data via application of feature extraction algorithms, configuring and training a neural network model, and classification of images by application of the trained neural network. A prototype experimentation testbed was completed and applied to climatological data
Efficient Classification for Metric Data
Recent advances in large-margin classification of data residing in general
metric spaces (rather than Hilbert spaces) enable classification under various
natural metrics, such as string edit and earthmover distance. A general
framework developed for this purpose by von Luxburg and Bousquet [JMLR, 2004]
left open the questions of computational efficiency and of providing direct
bounds on generalization error.
We design a new algorithm for classification in general metric spaces, whose
runtime and accuracy depend on the doubling dimension of the data points, and
can thus achieve superior classification performance in many common scenarios.
The algorithmic core of our approach is an approximate (rather than exact)
solution to the classical problems of Lipschitz extension and of Nearest
Neighbor Search. The algorithm's generalization performance is guaranteed via
the fat-shattering dimension of Lipschitz classifiers, and we present
experimental evidence of its superiority to some common kernel methods. As a
by-product, we offer a new perspective on the nearest neighbor classifier,
which yields significantly sharper risk asymptotics than the classic analysis
of Cover and Hart [IEEE Trans. Info. Theory, 1967].Comment: This is the full version of an extended abstract that appeared in
Proceedings of the 23rd COLT, 201
Data Fine-tuning
In real-world applications, commercial off-the-shelf systems are utilized for
performing automated facial analysis including face recognition, emotion
recognition, and attribute prediction. However, a majority of these commercial
systems act as black boxes due to the inaccessibility of the model parameters
which makes it challenging to fine-tune the models for specific applications.
Stimulated by the advances in adversarial perturbations, this research proposes
the concept of Data Fine-tuning to improve the classification accuracy of a
given model without changing the parameters of the model. This is accomplished
by modeling it as data (image) perturbation problem. A small amount of "noise"
is added to the input with the objective of minimizing the classification loss
without affecting the (visual) appearance. Experiments performed on three
publicly available datasets LFW, CelebA, and MUCT, demonstrate the
effectiveness of the proposed concept.Comment: Accepted in AAAI 201
High-Dimensional Data Reduction, Image Inpainting and their Astronomical Applications
Technological advances are revolutionizing multispectral astrophysics as well as the detection and study of transient sources. This new era of multitemporal and multispectral data sets demands new ways of data representation, processing and management thus making data dimension reduction instrumental in efficient data organization, retrieval, analysis and information visualization. Other astrophysical applications of data dimension reduction which require new paradigms of data analysis include knowledge discovery, cluster analysis, feature extraction and object classification, de-correlating data elements, discovering meaningful patterns and finding essential representation of correlated variables that form a manifold (e.g. the manifold of galaxies), tagging astronomical images, multiscale analysis synchronized across all available wavelengths, denoising, etc. The second part of this paper is dedicated to a new, active area of image processing: image inpainting that consists of automated methods for filling in missing or damaged regions in images. Inpainting has multiple astronomical applications including restoring images corrupted by instrument artifacts, removing undesirable objects like bright stars and their halos, sky estimating, and pre-processing for the Fourier or wavelet transforms. Applications of high-dimensional data reduction and mitigation of instrument artifacts are demonstrated on images taken by the Spitzer Space Telescope
Unsupervised feature-learning for galaxy SEDs with denoising autoencoders
With the increasing number of deep multi-wavelength galaxy surveys, the
spectral energy distribution (SED) of galaxies has become an invaluable tool
for studying the formation of their structures and their evolution. In this
context, standard analysis relies on simple spectro-photometric selection
criteria based on a few SED colors. If this fully supervised classification
already yielded clear achievements, it is not optimal to extract relevant
information from the data. In this article, we propose to employ very recent
advances in machine learning, and more precisely in feature learning, to derive
a data-driven diagram. We show that the proposed approach based on denoising
autoencoders recovers the bi-modality in the galaxy population in an
unsupervised manner, without using any prior knowledge on galaxy SED
classification. This technique has been compared to principal component
analysis (PCA) and to standard color/color representations. In addition,
preliminary results illustrate that this enables the capturing of extra
physically meaningful information, such as redshift dependence, galaxy mass
evolution and variation over the specific star formation rate. PCA also results
in an unsupervised representation with physical properties, such as mass and
sSFR, although this representation separates out. less other characteristics
(bimodality, redshift evolution) than denoising autoencoders.Comment: 11 pages and 15 figures. To be published in A&
GliomaPredict: A Clinically Useful Tool for Assigning Glioma Patients to Specific Molecular Subtypes
Background: Advances in generating genome-wide gene expression data have accelerated the development of molecular-based tumor classification systems. Tools that allow the translation of such molecular classification schemas from research into clinical applications are still missing in the emerging era of personalized medicine.
Results: We developed GliomaPredict as a computational tool that allows the fast and reliable classification of glioma patients into one of six previously published stratified subtypes based on sets of extensively validated classifiers derived from hundreds of glioma transcriptomic profiles. Our tool utilizes a principle component analysis (PCA)-based approach to generate a visual representation of the analyses, quantifies the confidence of the underlying subtype assessment and presents results as a printable PDF file. GliomaPredict tool is implemented as a plugin application for the widely-used GenePattern framework.
Conclusions: GliomaPredict provides a user-friendly, clinically applicable novel platform for instantly assigning gene expression-based subtype in patients with gliomas thereby aiding in clinical trial design and therapeutic decisionmaking. Implemented as a user-friendly diagnostic tool, we expect that in time GliomaPredict, and tools like it, will become routinely used in translational/clinical research and in the clinical care of patients with gliomas
Overview: Computer vision and machine learning for microstructural characterization and analysis
The characterization and analysis of microstructure is the foundation of
microstructural science, connecting the materials structure to its composition,
process history, and properties. Microstructural quantification traditionally
involves a human deciding a priori what to measure and then devising a
purpose-built method for doing so. However, recent advances in data science,
including computer vision (CV) and machine learning (ML) offer new approaches
to extracting information from microstructural images. This overview surveys CV
approaches to numerically encode the visual information contained in a
microstructural image, which then provides input to supervised or unsupervised
ML algorithms that find associations and trends in the high-dimensional image
representation. CV/ML systems for microstructural characterization and analysis
span the taxonomy of image analysis tasks, including image classification,
semantic segmentation, object detection, and instance segmentation. These tools
enable new approaches to microstructural analysis, including the development of
new, rich visual metrics and the discovery of
processing-microstructure-property relationships.Comment: submitted to Materials and Metallurgical Transactions
- …