44,881 research outputs found
CSNE: Conditional Signed Network Embedding
Signed networks are mathematical structures that encode positive and negative
relations between entities such as friend/foe or trust/distrust. Recently,
several papers studied the construction of useful low-dimensional
representations (embeddings) of these networks for the prediction of missing
relations or signs. Existing embedding methods for sign prediction generally
enforce different notions of status or balance theories in their optimization
function. These theories, however, are often inaccurate or incomplete, which
negatively impacts method performance.
In this context, we introduce conditional signed network embedding (CSNE).
Our probabilistic approach models structural information about the signs in the
network separately from fine-grained detail. Structural information is
represented in the form of a prior, while the embedding itself is used for
capturing fine-grained information. These components are then integrated in a
rigorous manner. CSNE's accuracy depends on the existence of sufficiently
powerful structural priors for modelling signed networks, currently unavailable
in the literature. Thus, as a second main contribution, which we find to be
highly valuable in its own right, we also introduce a novel approach to
construct priors based on the Maximum Entropy (MaxEnt) principle. These priors
can model the \emph{polarity} of nodes (degree to which their links are
positive) as well as signed \emph{triangle counts} (a measure of the degree
structural balance holds to in a network).
Experiments on a variety of real-world networks confirm that CSNE outperforms
the state-of-the-art on the task of sign prediction. Moreover, the MaxEnt
priors on their own, while less accurate than full CSNE, achieve accuracies
competitive with the state-of-the-art at very limited computational cost, thus
providing an excellent runtime-accuracy trade-off in resource-constrained
situations
CSNE : Conditional Signed Network Embedding
Signed networks are mathematical structures that encode positive and negative relations between entities such as friend/foe or trust/distrust. Recently, several papers studied the construction of useful low-dimensional representations (embeddings) of these networks for the prediction of missing relations or signs. Existing embedding methods for sign prediction generally enforce different notions of status or balance theories in their optimization function. These theories, however, are often inaccurate or incomplete, which negatively impacts method performance.
In this context, we introduce conditional signed network embedding (CSNE). Our probabilistic approach models structural information about the signs in the network separately from fine-grained detail. Structural information is represented in the form of a prior, while the embedding itself is used for capturing fine-grained information. These components are then integrated in a rigorous manner. CSNE's accuracy depends on the existence of sufficiently powerful structural priors for modelling signed networks, currently unavailable in the literature. Thus, as a second main contribution, which we find to be highly valuable in its own right, we also introduce a novel approach to construct priors based on the Maximum Entropy (MaxEnt) principle. These priors can model the polarity of nodes (degree to which their links are positive) as well as signed triangle counts (a measure of the degree structural balance holds to in a network).
Experiments on a variety of real-world networks confirm that CSNE outperforms the state-of-the-art on the task of sign prediction. Moreover, the MaxEnt priors on their own, while less accurate than full CSNE, achieve accuracies competitive with the state-of-the-art at very limited computational cost, thus providing an excellent runtime-accuracy trade-off in resource-constrained situations
Social Intelligence Design for Mediated Communication
Without abstract
On the emergent Semantic Web and overlooked issues
The emergent Semantic Web, despite being in its infancy, has already received a lotof attention from academia and industry. This resulted in an abundance of prototype systems and discussion most of which are centred around the underlying infrastructure. However, when we critically review the work done to date we realise that there is little discussion with respect to the vision of the Semantic Web. In particular, there is an observed dearth of discussion on how to deliver knowledge sharing in an environment such as the Semantic Web in effective and efficient manners. There are a lot of overlooked issues, associated with agents and trust to hidden assumptions made with respect to knowledge representation and robust reasoning in a distributed environment. These issues could potentially hinder further development if not considered at the early stages of designing Semantic Web systems. In this perspectives paper, we aim to help engineers and practitioners of the Semantic Web by raising awareness of these issues
The Pragmatic Turn in Explainable Artificial Intelligence (XAI)
In this paper I argue that the search for explainable models and interpretable decisions in AI must be reformulated in terms of the broader project of offering a pragmatic and naturalistic account of understanding in AI. Intuitively, the purpose of providing an explanation of a model or a decision is to make it understandable to its stakeholders. But without a previous grasp of what it means to say that an agent understands a model or a decision, the explanatory strategies will lack a well-defined goal. Aside from providing a clearer objective for XAI, focusing on understanding also allows us to relax the factivity condition on explanation, which is impossible to fulfill in many machine learning models, and to focus instead on the pragmatic conditions that determine the best fit between a model and the methods and devices deployed to understand it. After an examination of the different types of understanding discussed in the philosophical and psychological literature, I conclude that interpretative or approximation models not only provide the best way to achieve the objectual understanding of a machine learning model, but are also a necessary condition to achieve post hoc interpretability. This conclusion is partly based on the shortcomings of the purely functionalist approach to post hoc interpretability that seems to be predominant in most recent literature
- ā¦