4 research outputs found

    Sesgos de g茅nero en el uso de inteligencia artificial para la gesti贸n de las relaciones laborales: an谩lisis desde el derecho antidiscriminatorio

    Get PDF
    El an谩lisis jur铆dico de la discriminaci贸n derivada de decisiones automatizadas que puedan provocar un impacto discriminatorio requiere combinar dos campos jur铆dicos: el de la protecci贸n de datos y el derecho antidiscriminatorio. En el primero los derechos reconocidos son accesorios al n煤cleo principal de afectaci贸n: el derecho de intervenci贸n humana y, principalmente, la explicabilidad de los algoritmos, manifestaci贸n de la debida justificaci贸n objetiva y razonable que acompa帽a a las decisiones prima facie discriminatorias para eludir su calificaci贸n como tales. Pero el tratamiento jur铆dico de la discriminaci贸n algor铆tmica requiere, tambi茅n, dar respuesta a problemas de calificaci贸n de los sesgos en los que incurre el aprendizaje autom谩tico como resultado de las infinitas inferencias de datos que perfilan a personas en el contexto del derecho antidiscriminatorio, donde potencian su impacto discriminatorio, como son la discriminaci贸n por asociaci贸n o la discriminaci贸n m煤ltiple o interseccional. The legal analysis of discrimination derived from automated decisions that may have a discriminatory impact requires combining two legal fields: data protection law and anti-discrimination law. The recognized rights of first field are accessory to the nucleus of affectation, and they're the right of human intervention and, mainly, the explicability of the algorithms, which constitutes a manifestation of the due objective and reasonable justification in case of discriminatory prima facie decisions in order to elude their qualification as such. But the legal approach to algorithmic discrimination also requires responding to problems of qualifying the biases derived from machine learning as a result of the infinite data inferences that outline people in the context of anti-discrimination law, where they enhance their discriminatory impact, such as discrimination by association or multiple or intersectional discrimination

    Simultaneous discrimination prevention and privacy protection in data publishing and mining

    No full text
    Data mining is an increasingly important technology for extracting useful knowledge hidden in large collections of data. There are, however, negative social perceptions about data mining, among which potential privacy violation and potential discrimination. The former is an unintentional or deliberate disclosure of a user pro le or activity data as part of the output of a data mining algorithm or as a result of data sharing. For this reason, privacy preserving data mining has been introduced to trade o the utility of the resulting data/models for protecting individual privacy. The latter consists of treating people unfairly on the basis of their belonging to a speci c group. Automated data collection and data mining techniques such as classi cation have paved the way to making automated decisions, like loan granting/denial, insurance premium computation, etc. If the training datasets are biased in what regards discriminatory attributes like gender, race, religion, etc., discriminatory decisions may ensue. For this reason, anti-discrimination techniques including discrimination discovery and prevention have been introduced in data mining. Discrimination can be either direct or indirect. Direct discrimination occurs when decisions are made based on discriminatory attributes. Indirect discrimination occurs when decisions are made based on non-discriminatory attributes which are strongly correlated with biased discriminatory ones. In the rst part of this thesis, we tackle discrimination prevention in data mining and propose new techniques applicable for direct or indirect discrimination prevention individually or both at the same time. We discuss how to clean training datasets and outsourced datasets in such a way that direct and/or indirect discriminatory decision rules are converted to legitimate (non-discriminatory) classi cation rules. The experimental evaluations demonstrate that the proposed techniques are e ective at removing direct and/or indirect discrimination biases in the original dataset while preserving data quality. In the second part of this thesis, by presenting samples of privacy violation and potential discrimination in data mining, we argue that privacy and discrimination risks should be tackled together. We explore the relationship between privacy preserving data mining and discrimination prevention in data mining to design holistic approaches capable of addressing both threats simultaneously during the knowledge discovery process. As part of this e ort, we have investigated for the rst time the problem of discrimination and privacy aware frequent pattern discovery, i.e. the sanitization of the collection of patterns mined from a transaction database in such a way that neither privacy-violating nor discriminatory inferences can be inferred on the released patterns. Moreover, we investigate the problem of discrimination and privacy aware data publishing, i.e. transforming the data, instead of patterns, in order to simultaneously ful ll privacy preservation and discrimination prevention. In the above cases, it turns out that the impact of our transformation on the quality of data or patterns is the same or only slightly higher than the impact of achieving just privacy preservation
    corecore