6,478 research outputs found
THE PROCESS AUGMENTABILITY CANVAS - HOW TO FIND THE SWEET SPOT FOR AUGMENTED REALITY
The adoption of augmented reality (AR) has been one of the defining technological trends of the past decade. While AR has experienced significant growth in consumer electronics, its potential for professional use still needs to be explored. Despite the growing interest in AR, determining its feasibility and potential to satisfy business needs remains challenging. To address this gap, we used a mixedmethod research approach to create a guiding framework called the process augmentability canvas. Drawing on a comprehensive case study of a major European maintenance, repair, and overhaul service provider, as well as state-of-the literature, we present a canvas that allows scholars and practitioners to evaluate AR’s applicability for a given process thoroughly. By providing a structured approach to analyzing AR solutions, the process augmentability canvas contributes to a better understanding of how AR can be used efficiently in organizational settings
Virtual environment trajectory analysis:a basis for navigational assistance and scene adaptivity
This paper describes the analysis and clustering of motion trajectories obtained while users navigate within a virtual environment (VE). It presents a neural network simulation that produces a set of five clusters which help to differentiate users on the basis of efficient and inefficient navigational strategies. The accuracy of classification carried out with a self-organising map algorithm was tested and improved to in excess of 85% by using learning vector quantisation. This paper considers how such user classifications could be utilised in the delivery of intelligent navigational support and the dynamic reconfiguration of scenes within such VEs. We explore how such intelligent assistance and system adaptivity could be delivered within a Multi-Agent Systems (MAS) context
Semantic-aware Node Synthesis for Imbalanced Heterogeneous Information Networks
Heterogeneous graph neural networks (HGNNs) have exhibited exceptional
efficacy in modeling the complex heterogeneity in heterogeneous information
networks (HINs). The critical advantage of HGNNs is their ability to handle
diverse node and edge types in HINs by extracting and utilizing the abundant
semantic information for effective representation learning. However, as a
widespread phenomenon in many real-world scenarios, the class-imbalance
distribution in HINs creates a performance bottleneck for existing HGNNs. Apart
from the quantity imbalance of nodes, another more crucial and distinctive
challenge in HINs is semantic imbalance. Minority classes in HINs often lack
diverse and sufficient neighbor nodes, resulting in biased and incomplete
semantic information. This semantic imbalance further compounds the difficulty
of accurately classifying minority nodes, leading to the performance
degradation of HGNNs. To tackle the imbalance of minority classes and
supplement their inadequate semantics, we present the first method for the
semantic imbalance problem in imbalanced HINs named Semantic-aware Node
Synthesis (SNS). By assessing the influence on minority classes, SNS adaptively
selects the heterogeneous neighbor nodes and augments the network with
synthetic nodes while preserving the minority semantics. In addition, we
introduce two regularization approaches for HGNNs that constrain the
representation of synthetic nodes from both semantic and class perspectives to
effectively suppress the potential noises from synthetic nodes, facilitating
more expressive embeddings for classification. The comprehensive experimental
study demonstrates that SNS consistently outperforms existing methods by a
large margin in different benchmark datasets
Comparison and contrast in perceptual categorization
People categorized pairs of perceptual stimuli that varied in both category membership and pairwise similarity. Experiments 1 and 2 showed categorization of 1 color of a pair to be reliably contrasted from that of the other. This similarity-based contrast effect occurred only when the context stimulus was relevant for the categorization of the target (Experiment 3). The effect was not simply owing to perceptual color contrast (Experiment 4), and it extended to pictures from common semantic categories (Experiment 5). Results were consistent with a sign-and-magnitude version of N. Stewart and G. D. A. Brown's (2005) similarity-dissimilarity generalized context model, in which categorization is affected by both similarity to and difference from target categories. The data are also modeled with criterion setting theory (M. Treisman & T. C. Williams, 1984), in which the decision criterion is systematically shifted toward the mean of the current stimuli
Semantic-aware Video Representation for Few-shot Action Recognition
Recent work on action recognition leverages 3D features and textual
information to achieve state-of-the-art performance. However, most of the
current few-shot action recognition methods still rely on 2D frame-level
representations, often require additional components to model temporal
relations, and employ complex distance functions to achieve accurate alignment
of these representations. In addition, existing methods struggle to effectively
integrate textual semantics, some resorting to concatenation or addition of
textual and visual features, and some using text merely as an additional
supervision without truly achieving feature fusion and information transfer
from different modalities. In this work, we propose a simple yet effective
Semantic-Aware Few-Shot Action Recognition (SAFSAR) model to address these
issues. We show that directly leveraging a 3D feature extractor combined with
an effective feature-fusion scheme, and a simple cosine similarity for
classification can yield better performance without the need of extra
components for temporal modeling or complex distance functions. We introduce an
innovative scheme to encode the textual semantics into the video representation
which adaptively fuses features from text and video, and encourages the visual
encoder to extract more semantically consistent features. In this scheme,
SAFSAR achieves alignment and fusion in a compact way. Experiments on five
challenging few-shot action recognition benchmarks under various settings
demonstrate that the proposed SAFSAR model significantly improves the
state-of-the-art performance.Comment: WACV202
Tangible user interfaces : past, present and future directions
In the last two decades, Tangible User Interfaces (TUIs) have emerged as a new interface type that interlinks the digital and physical worlds. Drawing upon users' knowledge and skills of interaction with the real non-digital world, TUIs show a potential to enhance the way in which people interact with and leverage digital information. However, TUI research is still in its infancy and extensive research is required in or- der to fully understand the implications of tangible user interfaces, to develop technologies that further bridge the digital and the physical, and to guide TUI design with empirical knowledge. This paper examines the existing body of work on Tangible User In- terfaces. We start by sketching the history of tangible user interfaces, examining the intellectual origins of this field. We then present TUIs in a broader context, survey application domains, and review frame- works and taxonomies. We also discuss conceptual foundations of TUIs including perspectives from cognitive sciences, phycology, and philoso- phy. Methods and technologies for designing, building, and evaluating TUIs are also addressed. Finally, we discuss the strengths and limita- tions of TUIs and chart directions for future research
XAI & I: Self-explanatory AI facilitating mutual understanding between AI and human experts
Traditionally, explainable artificial intelligence seeks to provide explanation and interpretability of high-performing black-box models such as deep neural networks. Interpretation of such models remains difficult, because of their high complexity. An alternative method is to instead force a deep-neural network to use human-intelligible features as the basis for its decisions. We tested this approach using the natural category domain of rock types. We compared the performance of a black-box implementation of transfer-learning using Resnet50 to that of a network first trained to predict expert-identified features and then forced to use these features to categorise rock images. The performance of this feature-constrained network was virtually identical to that of the unconstrained network. Further, a partially constrained network forced to condense down to a small number of features that was not trained with expert features did not result in these abstracted features being intelligible; nevertheless, an affine transformation of these features could be found that aligned well with expert-intelligible features. These findings show that making an AI intrinsically intelligible need not be at the cost of performance
- …