15,731 research outputs found
Titan Science with the James Webb Space Telescope (JWST)
The James Webb Space Telescope (JWST), scheduled for launch in 2018, is the
successor to the Hubble Space Telescope (HST) but with a significantly larger
aperture (6.5 m) and advanced instrumentation focusing on infrared science
(0.6-28.0 m ). In this paper we examine the potential for scientific
investigation of Titan using JWST, primarily with three of the four
instruments: NIRSpec, NIRCam and MIRI, noting that science with NIRISS will be
complementary. Five core scientific themes are identified: (i) surface (ii)
tropospheric clouds (iii) tropospheric gases (iv) stratospheric composition and
(v) stratospheric hazes. We discuss each theme in depth, including the
scientific purpose, capabilities and limitations of the instrument suite, and
suggested observing schemes. We pay particular attention to saturation, which
is a problem for all three instruments, but may be alleviated for NIRCam
through use of selecting small sub-arrays of the detectors - sufficient to
encompass Titan, but with significantly faster read-out times. We find that
JWST has very significant potential for advancing Titan science, with a
spectral resolution exceeding the Cassini instrument suite at near-infrared
wavelengths, and a spatial resolution exceeding HST at the same wavelengths. In
particular, JWST will be valuable for time-domain monitoring of Titan, given a
five to ten year expected lifetime for the observatory, for example monitoring
the seasonal appearance of clouds. JWST observations in the post-Cassini period
will complement those of other large facilities such as HST, ALMA, SOFIA and
next-generation ground-based telescopes (TMT, GMT, EELT).Comment: 50 pages, including 22 figures and 2 table
Visually grounded learning of keyword prediction from untranscribed speech
During language acquisition, infants have the benefit of visual cues to
ground spoken language. Robots similarly have access to audio and visual
sensors. Recent work has shown that images and spoken captions can be mapped
into a meaningful common space, allowing images to be retrieved using speech
and vice versa. In this setting of images paired with untranscribed spoken
captions, we consider whether computer vision systems can be used to obtain
textual labels for the speech. Concretely, we use an image-to-words multi-label
visual classifier to tag images with soft textual labels, and then train a
neural network to map from the speech to these soft targets. We show that the
resulting speech system is able to predict which words occur in an
utterance---acting as a spoken bag-of-words classifier---without seeing any
parallel speech and text. We find that the model often confuses semantically
related words, e.g. "man" and "person", making it even more effective as a
semantic keyword spotter.Comment: 5 pages, 3 figures, 5 tables; small updates, added link to code;
accepted to Interspeech 201
Object Discovery From a Single Unlabeled Image by Mining Frequent Itemset With Multi-scale Features
TThe goal of our work is to discover dominant objects in a very general
setting where only a single unlabeled image is given. This is far more
challenge than typical co-localization or weakly-supervised localization tasks.
To tackle this problem, we propose a simple but effective pattern mining-based
method, called Object Location Mining (OLM), which exploits the advantages of
data mining and feature representation of pre-trained convolutional neural
networks (CNNs). Specifically, we first convert the feature maps from a
pre-trained CNN model into a set of transactions, and then discovers frequent
patterns from transaction database through pattern mining techniques. We
observe that those discovered patterns, i.e., co-occurrence highlighted
regions, typically hold appearance and spatial consistency. Motivated by this
observation, we can easily discover and localize possible objects by merging
relevant meaningful patterns. Extensive experiments on a variety of benchmarks
demonstrate that OLM achieves competitive localization performance compared
with the state-of-the-art methods. We also evaluate our approach compared with
unsupervised saliency detection methods and achieves competitive results on
seven benchmark datasets. Moreover, we conduct experiments on fine-grained
classification to show that our proposed method can locate the entire object
and parts accurately, which can benefit to improving the classification results
significantly
Deep filter banks for texture recognition, description, and segmentation
Visual textures have played a key role in image understanding because they
convey important semantics of images, and because texture representations that
pool local image descriptors in an orderless manner have had a tremendous
impact in diverse applications. In this paper we make several contributions to
texture understanding. First, instead of focusing on texture instance and
material category recognition, we propose a human-interpretable vocabulary of
texture attributes to describe common texture patterns, complemented by a new
describable texture dataset for benchmarking. Second, we look at the problem of
recognizing materials and texture attributes in realistic imaging conditions,
including when textures appear in clutter, developing corresponding benchmarks
on top of the recently proposed OpenSurfaces dataset. Third, we revisit classic
texture representations, including bag-of-visual-words and the Fisher vectors,
in the context of deep learning and show that these have excellent efficiency
and generalization properties if the convolutional layers of a deep model are
used as filter banks. We obtain in this manner state-of-the-art performance in
numerous datasets well beyond textures, an efficient method to apply deep
features to image regions, as well as benefit in transferring features from one
domain to another.Comment: 29 pages; 13 figures; 8 table
Deep Learning Approaches in Pavement Distress Identification: A Review
This paper presents a comprehensive review of recent advancements in image
processing and deep learning techniques for pavement distress detection and
classification, a critical aspect in modern pavement management systems. The
conventional manual inspection process conducted by human experts is gradually
being superseded by automated solutions, leveraging machine learning and deep
learning algorithms to enhance efficiency and accuracy. The ability of these
algorithms to discern patterns and make predictions based on extensive datasets
has revolutionized the domain of pavement distress identification. The paper
investigates the integration of unmanned aerial vehicles (UAVs) for data
collection, offering unique advantages such as aerial perspectives and
efficient coverage of large areas. By capturing high-resolution images, UAVs
provide valuable data that can be processed using deep learning algorithms to
detect and classify various pavement distresses effectively. While the primary
focus is on 2D image processing, the paper also acknowledges the challenges
associated with 3D images, such as sensor limitations and computational
requirements. Understanding these challenges is crucial for further
advancements in the field. The findings of this review significantly contribute
to the evolution of pavement distress detection, fostering the development of
efficient pavement management systems. As automated approaches continue to
mature, the implementation of deep learning techniques holds great promise in
ensuring safer and more durable road infrastructure for the benefit of society
- …