2,380 research outputs found
Using Artificial Intelligence to Identify Perpetrators of Technology Facilitated Coercive Control
This study is one of the 21 projects funded by the Home Office for research on perpetrators of domestic abuse. It is interested in a specific form of domestic abuse known as Technology Facilitated Coercive Control (TFCC) and focussed on the digital communication between (alleged) perpetrators and victim/survivors held on mobile phones. The purpose of this feasibility study was twofold,
i. to test the viability of an Artificial Intelligence (AI) programme to identify () perpetrators (including alleged perpetrators) of domestic abuse using digital communications held on mobile phones
ii. to examine police and victim/survivor attitudes towards using AI in police investigations.
Using digital conversations extracted from court transcriptions where TFCC was identified as a factor in the offending, the research team tested data sets built on different methods and techniques of AI. Natural Language Processing (NLP) tools, a subfield of AI, were also tested for their speed and accuracy in recognising abusive communication and identifying and risk assessing perpetrators of TFCC.
Conscious of national concern about policing practices relating to Violence Against Women and Girls and that any AI programme would be futile without the co-operation of both the police and the public, two online surveys were devised to measure opinion. The first sought insight into the attitudes of victim/survivors, viewed as experts in domestic abuse, about using AI in police investigations. The second involved the police and questioned their views of using AI in this way
Using Artificial Intelligence to Identify Perpetrators of Technology Facilitated Coercive Control.
This study investigated the feasibility of using Artificial Intelligence to identify perpetrators of coercive control through digital data held on mobile phones. The research also sought the views of the police and victim/survivors of domestic abuse to using technology in this way
Directions in abusive language training data, a systematic review: Garbage in, garbage out
Data-driven and machine learning based approaches for detecting, categorising and measuring abusive content such as hate speech and harassment have gained traction due to their scalability, robustness and increasingly high performance. Making effective detection systems for abusive content relies on having the right training datasets, reflecting a widely accepted mantra in computer science: Garbage In, Garbage Out. However, creating training datasets which are large, varied, theoretically-informed and that minimize biases is difficult, laborious and requires deep expertise. This paper systematically reviews 63 publicly available training datasets which have been created to train abusive language classifiers. It also reports on creation of a dedicated website for cataloguing abusive language data hatespeechdata.com. We discuss the challenges and opportunities of open science in this field, and argue that although more dataset sharing would bring many benefits it also poses social and ethical risks which need careful consideration. Finally, we provide evidence-based recommendations for practitioners creating new abusive content training datasets
Voucher Abuse Detection with Prompt-based Fine-tuning on Graph Neural Networks
Voucher abuse detection is an important anomaly detection problem in
E-commerce. While many GNN-based solutions have emerged, the supervised
paradigm depends on a large quantity of labeled data. A popular alternative is
to adopt self-supervised pre-training using label-free data, and further
fine-tune on a downstream task with limited labels. Nevertheless, the
"pre-train, fine-tune" paradigm is often plagued by the objective gap between
pre-training and downstream tasks. Hence, we propose VPGNN, a prompt-based
fine-tuning framework on GNNs for voucher abuse detection. We design a novel
graph prompting function to reformulate the downstream task into a similar
template as the pretext task in pre-training, thereby narrowing the objective
gap. Extensive experiments on both proprietary and public datasets demonstrate
the strength of VPGNN in both few-shot and semi-supervised scenarios. Moreover,
an online deployment of VPGNN in a production environment shows a 23.4%
improvement over two existing deployed models.Comment: 7 pages, Accepted by CIKM23 Applied Research Trac
- …