109 research outputs found
Binary Classifier Inspired by Quantum Theory
Machine Learning (ML) helps us to recognize patterns from raw data. ML is
used in numerous domains i.e. biomedical, agricultural, food technology, etc.
Despite recent technological advancements, there is still room for substantial
improvement in prediction. Current ML models are based on classical theories of
probability and statistics, which can now be replaced by Quantum Theory (QT)
with the aim of improving the effectiveness of ML. In this paper, we propose
the Binary Classifier Inspired by Quantum Theory (BCIQT) model, which
outperforms the state of the art classification in terms of recall for every
category.Comment: AAAI 201
Internet of Things is a revolutionary approach for future technology enhancement: a review
Abstract Internet of Things (IoT) is a new paradigm that has changed the traditional way of living into a high tech life style. Smart city, smart homes, pollution control, energy saving, smart transportation, smart industries are such transformations due to IoT. A lot of crucial research studies and investigations have been done in order to enhance the technology through IoT. However, there are still a lot of challenges and issues that need to be addressed to achieve the full potential of IoT. These challenges and issues must be considered from various aspects of IoT such as applications, challenges, enabling technologies, social and environmental impacts etc. The main goal of this review article is to provide a detailed discussion from both technological and social perspective. The article discusses different challenges and key issues of IoT, architecture and important application domains. Also, the article bring into light the existing literature and illustrated their contribution in different aspects of IoT. Moreover, the importance of big data and its analysis with respect to IoT has been discussed. This article would help the readers and researcher to understand the IoT and its applicability to the real world
VISU at WASSA 2023 Shared Task: Detecting Emotions in Reaction to News Stories Leveraging BERT and Stacked Embeddings
Our system, VISU, participated in the WASSA 2023 Shared Task (3) of Emotion
Classification from essays written in reaction to news articles. Emotion
detection from complex dialogues is challenging and often requires
context/domain understanding. Therefore in this research, we have focused on
developing deep learning (DL) models using the combination of word embedding
representations with tailored prepossessing strategies to capture the nuances
of emotions expressed. Our experiments used static and contextual embeddings
(individual and stacked) with Bidirectional Long short-term memory (BiLSTM) and
Transformer based models. We occupied rank tenth in the emotion detection task
by scoring a Macro F1-Score of 0.2717, validating the efficacy of our
implemented approaches for small and imbalanced datasets with mixed categories
of target emotions
Towards Subject Agnostic Affective Emotion Recognition
This paper focuses on affective emotion recognition, aiming to perform in the
subject-agnostic paradigm based on EEG signals. However, EEG signals manifest
subject instability in subject-agnostic affective Brain-computer interfaces
(aBCIs), which led to the problem of distributional shift. Furthermore, this
problem is alleviated by approaches such as domain generalisation and domain
adaptation. Typically, methods based on domain adaptation confer comparatively
better results than the domain generalisation methods but demand more
computational resources given new subjects. We propose a novel framework,
meta-learning based augmented domain adaptation for subject-agnostic aBCIs. Our
domain adaptation approach is augmented through meta-learning, which consists
of a recurrent neural network, a classifier, and a distributional shift
controller based on a sum-decomposable function. Also, we present that a neural
network explicating a sum-decomposable function can effectively estimate the
divergence between varied domains. The network setting for augmented domain
adaptation follows meta-learning and adversarial learning, where the controller
promptly adapts to new domains employing the target data via a few
self-adaptation steps in the test phase. Our proposed approach is shown to be
effective in experiments on a public aBICs dataset and achieves similar
performance to state-of-the-art domain adaptation methods while avoiding the
use of additional computational resources.Comment: To Appear in MUWS workshop at the 32nd ACM International Conference
on Information and Knowledge Management (CIKM) 202
Towards a Quantum-Inspired Binary Classifier
Machine Learning classification models learn the relation between input as features and output as a class in order to predict the class for the new given input. Several research works have demonstrated the effectiveness of machine learning algorithms but the state-of-the-art algorithms are based on the classical theories of probability and logic. Quantum Mechanics (QM) has already shown its effectiveness in many fields and researchers have proposed several interesting results which cannot be obtained through classical theory. In recent years, researchers have been trying to investigate whether the QM can help to improve the classical machine learning algorithms. It is believed that the theory of QM may also inspire an effective algorithm if it is implemented properly. From this inspiration, we propose the quantum-inspired binary classifier, which is based on quantum detection theory. We used text corpora and image corpora to explore the effect of our proposed model. Our proposed model outperforms the state-of-the-art models in terms of precision, recall, and F-measure for several topics (categories) in the 20 newsgroup text corpora. Our proposed model outperformed all the baselines in terms of recall when the MNIST handwritten image dataset was used; F-measure is also higher for most of the categories and precision is also higher for some categories. Our proposed model suggests that binary classification effectiveness can be achieved by using quantum detection theory. In particular, we found that our Quantum-Inspired Binary Classifier can increase the precision, recall, and F-measure of classification where the state-of-the-art methods cannot
DILF: Differentiable Rendering-Based Multi-View Image-Language Fusion for Zero-Shot 3D Shape Understanding
Zero-shot 3D shape understanding aims to recognize âunseenâ 3D categories that are not present in training data. Recently, Contrastive LanguageâImage Pre-training (CLIP) has shown promising open-world performance in zero-shot 3D shape understanding tasks by information fusion among language and 3D modality. It first renders 3D objects into multiple 2D image views and then learns to understand the semantic relationships between the textual descriptions and images, enabling the model to generalize to new and unseen categories. However, existing studies in zero-shot 3D shape understanding rely on predefined rendering parameters, resulting in repetitive, redundant, and low-quality views. This limitation hinders the modelâs ability to fully comprehend 3D shapes and adversely impacts the textâimage fusion in a shared latent space. To this end, we propose a novel approach called Differentiable rendering-based multi-view ImageâLanguage Fusion (DILF) for zero-shot 3D shape understanding. Specifically, DILF leverages large-scale language models (LLMs) to generate textual prompts enriched with 3D semantics and designs a differentiable renderer with learnable rendering parameters to produce representative multi-view images. These rendering parameters can be iteratively updated using a textâimage fusion loss, which aids in parametersâ regression, allowing the model to determine the optimal viewpoint positions for each 3D object. Then a group-view mechanism is introduced to model interdependencies across views, enabling efficient information fusion to achieve a more comprehensive 3D shape understanding. Experimental results can demonstrate that DILF outperforms state-of-the-art methods for zero-shot 3D classification while maintaining competitive performance for standard 3D classification. The code is available at https://github.com/yuzaiyang123/DILP
Pixel and Feature Transfer Fusion for Unsupervised Cross-Dataset Person Re-Identification
Publisher Copyright: IEEERecently, unsupervised cross-dataset person reidentification (Re-ID) has attracted more and more attention, which aims to transfer knowledge of a labeled source domain to an unlabeled target domain. There are two common frameworks: one is pixel-alignment of transferring low-level knowledge, and the other is feature-alignment of transferring high-level knowledge. In this article, we propose a novel recurrent autoencoder (RAE) framework to unify these two kinds of methods and inherit their merits. Specifically, the proposed RAE includes three modules, i.e., a feature-transfer (FT) module, a pixel-transfer (PT) module, and a fusion module. The FT module utilizes an encoder to map source and target images to a shared feature space. In the space, not only features are identity-discriminative but also the gap between source and target features is reduced. The PT module takes a decoder to reconstruct original images with its features. Here, we hope that the images reconstructed from target features are in the sourcestyle. Thus, the low-level knowledge can be propagated to the target domain. After transferring both high- and low-level knowledge with the two proposed modules above, we design another bilinear pooling layer to fuse both kinds of knowledge. Extensive experiments on Market-1501, DukeMTMC-ReID, and MSMT17 datasets show that our method significantly outperforms either pixel-alignment or feature-alignment Re-ID methods and achieves new state-of-the-art results.Peer reviewe
Intelligent System for Depression Scale Estimation with Facial Expressions and Case Study in Industrial Intelligence
As a mental disorder, depression has affected people's lives, works, and so on. Researchers have proposed various industrial intelligent systems in the pattern recognition field for audiovisual depression detection. This paper presents an endâtoâend trainable intelligent system to generate highâlevel representations over the entire video clip. Specifically, a threeâdimensional (3D) convolutional neural network equipped with a module spatiotemporal feature aggregation module (STFAM) is trained from scratch on audio/visual emotion challenge (AVEC)2013 and AVEC2014 data, which can model the discriminative patterns closely related to depression. In the STFAM, channel and spatial attention mechanism and an aggregation method, namely 3D DEPâNetVLAD, are integrated to learn the compact characteristic based on the feature maps. Extensive experiments on the two databases (i.e., AVEC2013 and AVEC2014) are illustrated that the proposed intelligent system can efficiently model the underlying depression patterns and obtain better performances over the most videoâbased depression recognition approaches. Case studies are presented to describes the applicability of the proposed intelligent system for industrial intelligence.Peer reviewe
- âŠ