27 research outputs found
Black-Box Anomaly Attribution
When the prediction of a black-box machine learning model deviates from the
true observation, what can be said about the reason behind that deviation? This
is a fundamental and ubiquitous question that the end user in a business or
industrial AI application often asks. The deviation may be due to a sub-optimal
black-box model, or it may be simply because the sample in question is an
outlier. In either case, one would ideally wish to obtain some form of
attribution score -- a value indicative of the extent to which an input
variable is responsible for the anomaly.
In the present paper we address this task of ``anomaly attribution,''
particularly in the setting in which the model is black-box and the training
data are not available. Specifically, we propose a novel likelihood-based
attribution framework we call the ``likelihood compensation (LC),'' in which
the responsibility score is equated with the correction on each input variable
needed to attain the highest possible likelihood. We begin by showing formally
why mainstream model-agnostic explanation methods, such as the local linear
surrogate modeling and Shapley values, are not designed to explain anomalies.
In particular, we show that they are ``deviation-agnostic,'' namely, that their
explanations are blind to the fact that there is a deviation in the model
prediction for the sample of interest. We do this by positioning these existing
methods under the unified umbrella of a function family we call the
``integrated gradient family.'' We validate the effectiveness of the proposed
LC approach using publicly available data sets. We also conduct a case study
with a real-world building energy prediction task and confirm its usefulness in
practice based on expert feedback
Targeted Advertising on Social Networks Using Online Variational Tensor Regression
This paper is concerned with online targeted advertising on social networks.
The main technical task we address is to estimate the activation probability
for user pairs, which quantifies the influence one user may have on another
towards purchasing decisions. This is a challenging task because one marketing
episode typically involves a multitude of marketing campaigns/strategies of
different products for highly diverse customers. In this paper, we propose what
we believe is the first tensor-based contextual bandit framework for online
targeted advertising. The proposed framework is designed to accommodate any
number of feature vectors in the form of multi-mode tensor, thereby enabling to
capture the heterogeneity that may exist over user preferences, products, and
campaign strategies in a unified manner. To handle inter-dependency of tensor
modes, we introduce an online variational algorithm with a mean-field
approximation. We empirically confirm that the proposed TensorUCB algorithm
achieves a significant improvement in influence maximization tasks over the
benchmarks, which is attributable to its capability of capturing the
user-product heterogeneity.Comment: 18 pages, 7 figure
Anomaly Attribution with Likelihood Compensation
This paper addresses the task of explaining anomalous predictions of a
black-box regression model. When using a black-box model, such as one to
predict building energy consumption from many sensor measurements, we often
have a situation where some observed samples may significantly deviate from
their prediction. It may be due to a sub-optimal black-box model, or simply
because those samples are outliers. In either case, one would ideally want to
compute a ``responsibility score'' indicative of the extent to which an input
variable is responsible for the anomalous output. In this work, we formalize
this task as a statistical inverse problem: Given model deviation from the
expected value, infer the responsibility score of each of the input variables.
We propose a new method called likelihood compensation (LC), which is founded
on the likelihood principle and computes a correction to each input variable.
To the best of our knowledge, this is the first principled framework that
computes a responsibility score for real valued anomalous model deviations. We
apply our approach to a real-world building energy prediction task and confirm
its utility based on expert feedback.Comment: 8 pages, 7 figure
Diagnostic Spatio-temporal Transformer with Faithful Encoding
This paper addresses the task of anomaly diagnosis when the underlying data
generation process has a complex spatio-temporal (ST) dependency. The key
technical challenge is to extract actionable insights from the dependency
tensor characterizing high-order interactions among temporal and spatial
indices. We formalize the problem as supervised dependency discovery, where the
ST dependency is learned as a side product of multivariate time-series
classification. We show that temporal positional encoding used in existing ST
transformer works has a serious limitation in capturing higher frequencies
(short time scales). We propose a new positional encoding with a theoretical
guarantee, based on discrete Fourier transform. We also propose a new ST
dependency discovery framework, which can provide readily consumable diagnostic
information in both spatial and temporal directions. Finally, we demonstrate
the utility of the proposed model, DFStrans (Diagnostic Fourier-based
Spatio-temporal Transformer), in a real industrial application of building
elevator control
Latent Trait Analysis for Risk Management of Complex Information Technology Projects
Abstract-Recent years have seen a major increase in the application of predictive analytics to the service delivery domain as more and more service providers rely on such analytics for proactive risk management. At the pre-contract stage, identifying potential project risks accurately is of vital importance since it allows service providers to avoid profit erosion through proactive risk management. This paper describes a data-driven approach to project failure prediction of complex information technology (IT) projects. We introduce a novel theoretical framework of Latent Trait Analysis (LTA), whose original form was first developed in psychometrics. We take as the input questionnaire data of risk assessment reviews in the quality assurance (QA) process of IT projects before contract signing, and attempt to predict the project health in the delivery phase after contract signing. The idea is to explicitly capture the human cognitive process through LTA, and estimate the latent project failure tendency hidden behind the questionnaire answers collected by QA experts. Using real QA data of an IT service provider, we demonstrate that our approach outperforms existing approaches in project failure prediction while providing practical information on the usefulness of individual question items
Eigenspace-based Anomaly Detection in Computer Systems
We report on an automated runtime anomaly detection method at the application layer of multi-node computer systems. Although several network management systems are available in the market, none of them have sufficient capabilities to detect faults in multi-tier Web-based systems with redundancy. We model a Web-based system as a weighted graph, where each node represents a “service ” and each edge represents a dependency between services. Since the edge weights vary greatly over time, the problem we address is that of anomaly detection from a time sequence of graphs. In our method, we first extract a feature vector from the adjacency matrix that represents the activities of all of the services. The heart of our method is to use the principal eigenvector of the eigenclusters of the graph. Then we derive a probability distribution for an anomaly measure defined for a time-series of directional data derived from the graph sequence. Given a critical probability, the threshold value is adaptively updated using a novel online algorithm. We demonstrate that a fault in a Web application can be automatically detected and the faulty services are identified without using detailed knowledge of the behavior of the system
Theoretical Study on Nonlocal Effects in Resonant X-Ray Emission Spectra of Strongly-Correlated Systems
報告番号: 甲14943 ; 学位授与年月日: 2000-03-29 ; 学位の種別: 課程博士 ; 学位の種類: 博士(理学) ; 学位記番号: 博理第3707号 ; 研究科・専攻: 理学系研究科物理学専
Tensorial Change Analysis Using Probabilistic Tensor Regression
This paper proposes a new method for change detection and analysis using tensor regression. Change detection in our setting is to detect changes in the relationship between the input tensor and the output scalar while change analysis is to compute the responsibility score of individual tensor modes and dimensions for the change detected. We develop a new probabilistic tensor regression method, which can be viewed as a probabilistic generalization of the alternating least squares algorithm. Thanks to the probabilistic formulation, the derived change scores have a clear information-theoretic interpretation. We apply our method to semiconductor manufacturing to demonstrate the utility. To the best of our knowledge, this is the first work of change analysis based on probabilistic tensor regression