236 research outputs found
SemEval 2023 Task 6: LegalEval -- Understanding Legal Texts
In populous countries, pending legal cases have been growing exponentially.
There is a need for developing NLP-based techniques for processing and
automatically understanding legal documents. To promote research in the area of
Legal NLP we organized the shared task LegalEval - Understanding Legal Texts at
SemEval 2023. LegalEval task has three sub-tasks: Task-A (Rhetorical Roles
Labeling) is about automatically structuring legal documents into semantically
coherent units, Task-B (Legal Named Entity Recognition) deals with identifying
relevant entities in a legal document and Task-C (Court Judgement Prediction
with Explanation) explores the possibility of automatically predicting the
outcome of a legal case along with providing an explanation for the prediction.
In total 26 teams (approx. 100 participants spread across the world) submitted
systems paper. In each of the sub-tasks, the proposed systems outperformed the
baselines; however, there is a lot of scope for improvement. This paper
describes the tasks, and analyzes techniques proposed by various teams.Comment: 13 Pages (9 Pages + References), Accepted at SemEval 202
Multimodal Explainable Artificial Intelligence: A Comprehensive Review of Methodological Advances and Future Research Directions
The current study focuses on systematically analyzing the recent advances in
the field of Multimodal eXplainable Artificial Intelligence (MXAI). In
particular, the relevant primary prediction tasks and publicly available
datasets are initially described. Subsequently, a structured presentation of
the MXAI methods of the literature is provided, taking into account the
following criteria: a) The number of the involved modalities, b) The stage at
which explanations are produced, and c) The type of the adopted methodology
(i.e. mathematical formalism). Then, the metrics used for MXAI evaluation are
discussed. Finally, a comprehensive analysis of current challenges and future
research directions is provided.Comment: 26 pages, 11 figure
Interpreting Neural Networks for and with Natural Language
In the past decade, natural language processing (NLP) systems have come to be built almost exclusively on a backbone of large neural models. As the landscape of feasible tasks has widened due to the capabilities of these models, the space of applications has also widened to include subfields with real-world consequences, such as fact-checking, fake news detection, and medical decision support. The increasing size and nonlinearity of these models results in an opacity that hinders efforts by machine learning practitioners and lay-users alike to understand their internals and derive meaning or trust from their predictions.
The fields of explainable artificial intelligence (XAI) and more specifically explainable NLP (ExNLP) have emerged as an active area for remedying this opacity and for ensuring models' reliability and trustworthiness in high-stakes scenarios, by providing textual explanations meaningful to human users. Models that produce justifications for their individual predictions can be inspected for the purposes of debugging, quantifying bias and fairness, understanding model behavior, and ascertaining robustness and privacy. Textual explanation is a predominant form of explanation in machine learning datasets regardless of task modality. As such, this dissertation covers both explaining tasks with natural language and explaining natural language tasks.
In this dissertation, I propose test suites for evaluating the quality of model explanations under two definitions of meaning: faithfulness and human acceptability. I use these evaluation methods to investigate the utility of two explanation forms and three model architectures. I finally propose two methods to improve explanation quality– one which increases the likelihood of faithful highlight explanations and one which improves the human acceptability of free-text explanations. This work strives to increase the likelihood of positive use and outcomes when AI systems are deployed in practice.Ph.D
Data Augmentation For Text Classification Tasks
Thanks to increases in computing power and the growing availability of large datasets,
neural networks have achieved state of the art results in many natural language process-
ing (NLP) and computer vision (CV) tasks. These models require a large number of
training examples that are balanced between classes, but in many application areas they
rely on training sets that are either small or imbalanced, or both. To address this, data
augmentation has become standard practice in CV. This research is motivated by the ob-
servation that, relative to CV, data augmentation is underused and understudied in NLP.
Three methods of data augmentation are implemented and tested: synonym replacement,
backtranslation, and contextual augmentation. Tests are conducted with two models: a
Recurrent Neural Network (RNN) and Bidirectional Encoder Representations from Trans-
formers (BERT). To develop learning curves and study the ability of augmentation methods
to rebalance datasets, each of three binary classification datasets are made artificially small
and made artificially imbalanced. The results show that these augmentation methods can
offer accuracy improvements of over 1% to models with a baseline accuracy as high as
92%. On the two largest datasets, the accuracy of BERT is usually improved by either
synonym replacement or backtranslation, while the accuracy of the RNN is usually im-
proved by all three augmentation techniques. The augmentation techniques tend to yield
the largest accuracy boost when the datasets are smallest or most imbalanced; the per-
formance benefits appear to converge to 0% as the dataset becomes larger. The optimal
augmentation distance, the extent to which augmented training examples tend to deviate
from their original form, decreases as datasets become more balanced. The results show
that data augmentation is a powerful method of improving performance when training on
datasets with fewer than 10,000 training examples. The accuracy increases that they offer
are reduced by recent advancements in transfer learning schemes, but they are certainly
not eliminated
Knowledge Augmented Machine Learning with Applications in Autonomous Driving: A Survey
The existence of representative datasets is a prerequisite of many successful artificial intelligence and machine learning models. However, the subsequent application of these models often involves scenarios that are inadequately represented in the data used for training. The reasons for this are manifold and range from time and cost constraints to ethical considerations. As a consequence, the reliable use of these models, especially in safety-critical applications, is a huge challenge. Leveraging additional, already existing sources of knowledge is key to overcome the limitations of purely data-driven approaches, and eventually to increase the generalization capability of these models. Furthermore, predictions that conform with knowledge are crucial for making trustworthy and safe decisions even in underrepresented scenarios. This work provides an overview of existing techniques and methods in the literature that combine data-based models with existing knowledge. The identified approaches are structured according to the categories integration, extraction and conformity. Special attention is given to applications in the field of autonomous driving
Tracking the Temporal-Evolution of Supernova Bubbles in Numerical Simulations
The study of low-dimensional, noisy manifolds embedded in a higher dimensional space has been extremely useful in many applications, from the chemical analysis of multi-phase flows to simulations of galactic mergers. Building a probabilistic model of the manifolds has helped in describing their essential properties and how they vary in space. However, when the manifold is evolving through time, a joint spatio-temporal modelling is needed, in order to fully comprehend its nature. We propose a first-order Markovian process that propagates the spatial probabilistic model of a manifold at fixed time, to its adjacent temporal stages. The proposed methodology is demonstrated using a particle simulation of an interacting dwarf galaxy to describe the evolution of a cavity generated by a Supernov
- …