26 research outputs found
Disentangling and Operationalizing AI Fairness at LinkedIn
Operationalizing AI fairness at LinkedIn's scale is challenging not only
because there are multiple mutually incompatible definitions of fairness but
also because determining what is fair depends on the specifics and context of
the product where AI is deployed. Moreover, AI practitioners need clarity on
what fairness expectations need to be addressed at the AI level. In this paper,
we present the evolving AI fairness framework used at LinkedIn to address these
three challenges. The framework disentangles AI fairness by separating out
equal treatment and equitable product expectations. Rather than imposing a
trade-off between these two commonly opposing interpretations of fairness, the
framework provides clear guidelines for operationalizing equal AI treatment
complemented with a product equity strategy. This paper focuses on the equal AI
treatment component of LinkedIn's AI fairness framework, shares the principles
that support it, and illustrates their application through a case study. We
hope this paper will encourage other big tech companies to join us in sharing
their approach to operationalizing AI fairness at scale, so that together we
can keep advancing this constantly evolving field
Examining the Role of Mood Patterns in Predicting Self-reported Depressive Symptoms
Depression is the leading cause of disability worldwide. Initial efforts to
detect depression signals from social media posts have shown promising results.
Given the high internal validity, results from such analyses are potentially
beneficial to clinical judgment. The existing models for automatic detection of
depressive symptoms learn proxy diagnostic signals from social media data, such
as help-seeking behavior for mental health or medication names. However, in
reality, individuals with depression typically experience depressed mood, loss
of pleasure nearly in all the activities, feeling of worthlessness or guilt,
and diminished ability to think. Therefore, a lot of the proxy signals used in
these models lack the theoretical underpinnings for depressive symptoms. It is
also reported that social media posts from many patients in the clinical
setting do not contain these signals. Based on this research gap, we propose to
monitor a type of signal that is well-established as a class of symptoms in
affective disorders -- mood. The mood is an experience of feeling that can last
for hours, days, or even weeks. In this work, we attempt to enrich current
technology for detecting symptoms of potential depression by constructing a
'mood profile' for social media users.Comment: Accepted at The Web Science Conference 202
Extended Linear Models with Gaussian Priors
In extended linear models the input space is projected onto a feature space by means of an arbitrary non-linear transformation. A linear model is then applied to the feature space to construct the model output. The dimension of the feature space can be very large, or even infinite, giving the model a very big flexibility. Support Vector Machines (SVM's) and Gaussian processes are two examples of such models. In this technical report I present a model in which the dimension of the feature space remains finite, and where a Bayesian approach is used to train the model with Gaussian priors on the parameters. The Relevance Vector Machine..
Extended Linear Models with Gaussian Priors Extended Linear Models with Gaussian Priors
In extended linear models the input space is projected onto a feature space by means of an arbitrary non-linear transformation. A linear model is then applied to the feature space to construct the model output. The dimension of the feature space can be very large, or even infinite, giving the model a very big flexibility. Support Vector Machines (SVM’s) and Gaussian processes are two examples of such models. In this technical report I present a model in which the dimension of the feature space remains finite, and where a Bayesian approach is used to train the model with Gaussian priors on the parameters. The Relevance Vector Machine, introduced by Tipping Tipping (2001), is a particular case of such a model. I give the detailed derivations of the expectation-maximisation (EM) algorithm used in the training. These derivations are not found in the literature, and might be helpful for newcomers
Large margin non-linear embedding
It is common in classification methods to first place data in a vector space and then learn decision boundaries. We propose reversing that process: for fixed decision boundaries, we “learn ” the location of the data. This way we (i) do not need a metric (or even stronger structure) – pairwise dissimilarities suffice; and additionally (ii) produce low-dimensional embeddings that can be analyzed visually. We achieve this by combining an entropybased embedding method with an entropybased version of semi-supervised logistic regression. We present results for clustering and semi-supervised classification. 1
Healing the relevance vector machine through augmentation
The Relevance Vector Machine (RVM) is a sparse approximate Bayesian kernel method. It provides full predictive distributions for test cases. However, the predictive uncertainties have the unintuitive property, that they get smaller the further you move away from the training cases. We give a thorough analysis. Inspired by the analogy to nondegenerate Gaussian Processes, we suggest augmentation to solve the problem. The purpose of the resulting model, RVM*, is primarily to corroborate the theoretical and experimental analysis. Although RVM * could be used in practical applications, it is no longer a truly sparse model. Experiments show that sparsity comes at the expense of worse predictive distributions. Bayesian inference based on Gaussian Processes (GPs) has become widespread in the machine learning community. However, their naïve applicability is marred by computational constraints. A number of recent publications have addressed this issue by means of sparse approximations, although ideologically sparseness is at variance with Bayesian principles1. In this paper we view sparsity purely as a way to achieve computational convenience and not as under other non-Bayesian paradigms where sparseness itself is seen as a means to ensure good generalization