4 research outputs found
Sesgos de g茅nero en el uso de inteligencia artificial para la gesti贸n de las relaciones laborales: an谩lisis desde el derecho antidiscriminatorio
El an谩lisis jur铆dico de la discriminaci贸n derivada de decisiones automatizadas que puedan provocar un impacto discriminatorio requiere combinar dos campos jur铆dicos: el de la protecci贸n de datos y el derecho antidiscriminatorio. En el primero los derechos reconocidos son accesorios al n煤cleo principal de afectaci贸n: el derecho de intervenci贸n humana y, principalmente, la explicabilidad de los algoritmos, manifestaci贸n de la debida justificaci贸n objetiva y razonable que acompa帽a a las decisiones prima facie discriminatorias para eludir su calificaci贸n como tales. Pero el tratamiento jur铆dico de la discriminaci贸n algor铆tmica requiere, tambi茅n, dar respuesta a problemas de calificaci贸n de los sesgos en los que incurre el aprendizaje autom谩tico como resultado de las infinitas inferencias de datos que perfilan a personas en el contexto del derecho antidiscriminatorio, donde potencian su impacto discriminatorio, como son la discriminaci贸n por asociaci贸n o la discriminaci贸n m煤ltiple o interseccional. The legal analysis of discrimination derived from automated decisions that may have a discriminatory impact requires combining two legal fields: data protection law and anti-discrimination law. The recognized rights of first field are accessory to the nucleus of affectation, and they're the right of human intervention and, mainly, the explicability of the algorithms, which constitutes a manifestation of the due objective and reasonable justification in case of discriminatory prima facie decisions in order to elude their qualification as such. But the legal approach to algorithmic discrimination also requires responding to problems of qualifying the biases derived from machine learning as a result of the infinite data inferences that outline people in the context of anti-discrimination law, where they enhance their discriminatory impact, such as discrimination by association or multiple or intersectional discrimination
Simultaneous discrimination prevention and privacy protection in data publishing and mining
Data mining is an increasingly important technology for extracting useful knowledge hidden
in large collections of data. There are, however, negative social perceptions about data mining,
among which potential privacy violation and potential discrimination. The former is an
unintentional or deliberate disclosure of a user pro le or activity data as part of the output
of a data mining algorithm or as a result of data sharing. For this reason, privacy preserving
data mining has been introduced to trade o the utility of the resulting data/models for
protecting individual privacy. The latter consists of treating people unfairly on the basis
of their belonging to a speci c group. Automated data collection and data mining techniques
such as classi cation have paved the way to making automated decisions, like loan
granting/denial, insurance premium computation, etc. If the training datasets are biased
in what regards discriminatory attributes like gender, race, religion, etc., discriminatory
decisions may ensue. For this reason, anti-discrimination techniques including discrimination
discovery and prevention have been introduced in data mining. Discrimination can be
either direct or indirect. Direct discrimination occurs when decisions are made based on
discriminatory attributes. Indirect discrimination occurs when decisions are made based
on non-discriminatory attributes which are strongly correlated with biased discriminatory
ones.
In the rst part of this thesis, we tackle discrimination prevention in data mining and
propose new techniques applicable for direct or indirect discrimination prevention individually
or both at the same time. We discuss how to clean training datasets and outsourced
datasets in such a way that direct and/or indirect discriminatory decision rules are converted
to legitimate (non-discriminatory) classi cation rules. The experimental evaluations
demonstrate that the proposed techniques are e ective at removing direct and/or indirect
discrimination biases in the original dataset while preserving data quality.
In the second part of this thesis, by presenting samples of privacy violation and potential
discrimination in data mining, we argue that privacy and discrimination risks should
be tackled together. We explore the relationship between privacy preserving data mining
and discrimination prevention in data mining to design holistic approaches capable of addressing
both threats simultaneously during the knowledge discovery process. As part of
this e ort, we have investigated for the rst time the problem of discrimination and privacy
aware frequent pattern discovery, i.e. the sanitization of the collection of patterns mined
from a transaction database in such a way that neither privacy-violating nor discriminatory
inferences can be inferred on the released patterns. Moreover, we investigate the problem
of discrimination and privacy aware data publishing, i.e. transforming the data, instead of
patterns, in order to simultaneously ful ll privacy preservation and discrimination prevention.
In the above cases, it turns out that the impact of our transformation on the quality
of data or patterns is the same or only slightly higher than the impact of achieving just
privacy preservation