699 research outputs found
A Discriminative Representation of Convolutional Features for Indoor Scene Recognition
Indoor scene recognition is a multi-faceted and challenging problem due to
the diverse intra-class variations and the confusing inter-class similarities.
This paper presents a novel approach which exploits rich mid-level
convolutional features to categorize indoor scenes. Traditionally used
convolutional features preserve the global spatial structure, which is a
desirable property for general object recognition. However, we argue that this
structuredness is not much helpful when we have large variations in scene
layouts, e.g., in indoor scenes. We propose to transform the structured
convolutional activations to another highly discriminative feature space. The
representation in the transformed space not only incorporates the
discriminative aspects of the target dataset, but it also encodes the features
in terms of the general object categories that are present in indoor scenes. To
this end, we introduce a new large-scale dataset of 1300 object categories
which are commonly present in indoor scenes. Our proposed approach achieves a
significant performance boost over previous state of the art approaches on five
major scene classification datasets
Fuzzy Simulated Evolution Algorithm for Topology Design on Campus Networks
The topology design of campus networks is a hard contrained combinatorial optimization problem. It consists of deciding the number, type, and location of the active network elements (nodes) and links. This choice is dictated by physical and technological constraints and must optimize several objectives. Example of objectives are monetary cost, network delay, and hop count between communicating pairs. Furthermore, due to the nondeterministic nature of network traffic and other design parameters, the objectives criteria are imprecise. Fuzzy logic provides a suitable mathematical framework in such a situation. In this paper, we present an approach based on Simulated Evolution algorithm for the design of campus network topology. The two main phases of the algorithm, namely evaluation and allocation, have been fuzzified. To diversify the search, we have also incorporated Tabu Search-based characteristics in the allocation phase of the SE algorithm. This approach is then compared with Simulated Anealing algorithm, which is another well-known heuristic. Results show that on all test cases, Simulated Evolution algorithm more intelligent search of the solutions subspace and was able to find better solutions than Simulated Anealing
Adversarial Training of Variational Auto-encoders for High Fidelity Image Generation
Variational auto-encoders (VAEs) provide an attractive solution to image
generation problem. However, they tend to produce blurred and over-smoothed
images due to their dependence on pixel-wise reconstruction loss. This paper
introduces a new approach to alleviate this problem in the VAE based generative
models. Our model simultaneously learns to match the data, reconstruction loss
and the latent distributions of real and fake images to improve the quality of
generated samples. To compute the loss distributions, we introduce an
auto-encoder based discriminator model which allows an adversarial learning
procedure. The discriminator in our model also provides perceptual guidance to
the VAE by matching the learned similarity metric of the real and fake samples
in the latent space. To stabilize the overall training process, our model uses
an error feedback approach to maintain the equilibrium between competing
networks in the model. Our experiments show that the generated samples from our
proposed model exhibit a diverse set of attributes and facial expressions and
scale up to high-resolution images very well
From Known to the Unknown: Transferring Knowledge to Answer Questions about Novel Visual and Semantic Concepts
© 2020 Elsevier B.V. Current Visual Question Answering (VQA) systems can answer intelligent questions about ‘known’ visual content. However, their performance drops significantly when questions about visually and linguistically ‘unknown’ concepts are presented during inference (‘Open-world’ scenario). A practical VQA system should be able to deal with novel concepts in real world settings. To address this problem, we propose an exemplar-based approach that transfers learning (i.e., knowledge) from previously ‘known’ concepts to answer questions about the ‘unknown’. We learn a highly discriminative joint embedding (JE) space, where visual and semantic features are fused to give a unified representation. Once novel concepts are presented to the model, it looks for the closest match from an exemplar set in the JE space. This auxiliary information is used alongside the given Image-Question pair to refine visual attention in a hierarchical fashion. Our novel attention model is based on a dual-attention mechanism that combines the complementary effect of spatial and channel attention. Since handling the high dimensional exemplars on large datasets can be a significant challenge, we introduce an efficient matching scheme that uses a compact feature description for search and retrieval. To evaluate our model, we propose a new dataset for VQA, separating unknown visual and semantic concepts from the training set. Our approach shows significant improvements over state-of-the-art VQA models on the proposed Open-World VQA dataset and other standard VQA datasets
- …