15,914 research outputs found
EnTri: Ensemble Learning with Tri-level Representations for Explainable Scene Recognition
Scene recognition based on deep-learning has made significant progress, but
there are still limitations in its performance due to challenges posed by
inter-class similarities and intra-class dissimilarities. Furthermore, prior
research has primarily focused on improving classification accuracy, yet it has
given less attention to achieving interpretable, precise scene classification.
Therefore, we are motivated to propose EnTri, an ensemble scene recognition
framework that employs ensemble learning using a hierarchy of visual features.
EnTri represents features at three distinct levels of detail: pixel-level,
semantic segmentation-level, and object class and frequency level. By
incorporating distinct feature encoding schemes of differing complexity and
leveraging ensemble strategies, our approach aims to improve classification
accuracy while enhancing transparency and interpretability via visual and
textual explanations. To achieve interpretability, we devised an extension
algorithm that generates both visual and textual explanations highlighting
various properties of a given scene that contribute to the final prediction of
its category. This includes information about objects, statistics, spatial
layout, and textural details. Through experiments on benchmark scene
classification datasets, EnTri has demonstrated superiority in terms of
recognition accuracy, achieving competitive performance compared to
state-of-the-art approaches, with an accuracy of 87.69%, 75.56%, and 99.17% on
the MIT67, SUN397, and UIUC8 datasets, respectively.Comment: Submitted to Pattern Recognition journa
Recommended from our members
Abductive reasoning in neural-symbolic learning systems
Abduction is or subsumes a process of inference. It entertains possible hypotheses and it chooses hypotheses for further scrutiny. There is a large literature on various aspects of non-symbolic, subconscious abduction. There is also a very active research community working on the symbolic (logical) characterisation of abduction, which typically treats it as a form of hypothetico-deductive reasoning. In this paper we start to bridge the gap between the symbolic and sub-symbolic approaches to abduction. We are interested in benefiting from developments made by each community. In particular, we are interested in the ability of non-symbolic systems (neural networks) to learn from experience using efficient algorithms and to perform massively parallel computations of alternative abductive explanations. At the same time, we would like to benefit from the rigour and semantic clarity of symbolic logic. We present two approaches to dealing with abduction in neural networks. One of them uses Connectionist Modal Logic and a translation of Horn clauses into modal clauses to come up with a neural network ensemble that computes abductive explanations in a top-down fashion. The other combines neural-symbolic systems and abductive logic programming and proposes a neural architecture which performs a more systematic, bottom-up computation of alternative abductive explanations. Both approaches employ standard neural network architectures which are already known to be highly effective in practical learning applications. Differently from previous work in the area, our aim is to promote the integration of reasoning and learning in a way that the neural network provides the machinery for cognitive computation, inductive learning and hypothetical reasoning, while logic provides the rigour and explanation capability to the systems, facilitating the interaction with the outside world. Although it is left as future work to determine whether the structure of one of the proposed approaches is more amenable to learning than the other, we hope to have contributed to the development of the area by approaching it from the perspective of symbolic and sub-symbolic integration
Improving Retrieval-Based Question Answering with Deep Inference Models
Question answering is one of the most important and difficult applications at
the border of information retrieval and natural language processing, especially
when we talk about complex science questions which require some form of
inference to determine the correct answer. In this paper, we present a two-step
method that combines information retrieval techniques optimized for question
answering with deep learning models for natural language inference in order to
tackle the multi-choice question answering in the science domain. For each
question-answer pair, we use standard retrieval-based models to find relevant
candidate contexts and decompose the main problem into two different
sub-problems. First, assign correctness scores for each candidate answer based
on the context using retrieval models from Lucene. Second, we use deep learning
architectures to compute if a candidate answer can be inferred from some
well-chosen context consisting of sentences retrieved from the knowledge base.
In the end, all these solvers are combined using a simple neural network to
predict the correct answer. This proposed two-step model outperforms the best
retrieval-based solver by over 3% in absolute accuracy.Comment: 8 pages, 2 figures, 8 tables, accepted at IJCNN 201
- …