11,152 research outputs found
Experts and Decision Making: First Steps Towards a Unifying Theory of Decision Making in Novices, Intermediates and Experts
Expertise research shows quite ambiguous results on the abilities of experts in judgment and decision making (JDM) classic models cannot account for. This problem becomes even more accentuated if different levels of expertise are considered. We argue that parallel constraint satisfaction models (PCS) might be a useful base to understand the processes underlying expert JDM and the hitherto existing, differentiated results from expertise research. It is outlined how expertise might influence model parameters and mental representations according to PCS. It is discussed how this differential impact of expertise on model parameters relates to empirical results showing quite different courses in the development of expertise; allowing, for example, to predict under which conditions intermediates might outperform experts. Methodological requirements for testing the proposed unifying theory under complex real-world conditions are discussed.Judgment and Decision Making, Expertise, Intermediate Effects, Parallel Constraint Satisfaction, Mental Representation
ALJP: An Arabic Legal Judgment Prediction in Personal Status Cases Using Machine Learning Models
Legal Judgment Prediction (LJP) aims to predict judgment outcomes based on
case description. Several researchers have developed techniques to assist
potential clients by predicting the outcome in the legal profession. However,
none of the proposed techniques were implemented in Arabic, and only a few
attempts were implemented in English, Chinese, and Hindi. In this paper, we
develop a system that utilizes deep learning (DL) and natural language
processing (NLP) techniques to predict the judgment outcome from Arabic case
scripts, especially in cases of custody and annulment of marriage. This system
will assist judges and attorneys in improving their work and time efficiency
while reducing sentencing disparity. In addition, it will help litigants,
lawyers, and law students analyze the probable outcomes of any given case
before trial. We use a different machine and deep learning models such as
Support Vector Machine (SVM), Logistic regression (LR), Long Short Term Memory
(LSTM), and Bidirectional Long Short-Term Memory (BiLSTM) using representation
techniques such as TF-IDF and word2vec on the developed dataset. Experimental
results demonstrate that compared with the five baseline methods, the SVM model
with word2vec and LR with TF-IDF achieve the highest accuracy of 88% and 78% in
predicting the judgment on custody cases and annulment of marriage,
respectively. Furthermore, the LR and SVM with word2vec and BiLSTM model with
TF-IDF achieved the highest accuracy of 88% and 69% in predicting the
probability of outcomes on custody cases and annulment of marriage,
respectively
JEC-QA: A Legal-Domain Question Answering Dataset
We present JEC-QA, the largest question answering dataset in the legal
domain, collected from the National Judicial Examination of China. The
examination is a comprehensive evaluation of professional skills for legal
practitioners. College students are required to pass the examination to be
certified as a lawyer or a judge. The dataset is challenging for existing
question answering methods, because both retrieving relevant materials and
answering questions require the ability of logic reasoning. Due to the high
demand of multiple reasoning abilities to answer legal questions, the
state-of-the-art models can only achieve about 28% accuracy on JEC-QA, while
skilled humans and unskilled humans can reach 81% and 64% accuracy
respectively, which indicates a huge gap between humans and machines on this
task. We will release JEC-QA and our baselines to help improve the reasoning
ability of machine comprehension models. You can access the dataset from
http://jecqa.thunlp.org/.Comment: 9 pages, 2 figures, 10 tables, accepted by AAAI202
Bringing order into the realm of Transformer-based language models for artificial intelligence and law
Transformer-based language models (TLMs) have widely been recognized to be a
cutting-edge technology for the successful development of deep-learning-based
solutions to problems and applications that require natural language processing
and understanding. Like for other textual domains, TLMs have indeed pushed the
state-of-the-art of AI approaches for many tasks of interest in the legal
domain. Despite the first Transformer model being proposed about six years ago,
there has been a rapid progress of this technology at an unprecedented rate,
whereby BERT and related models represent a major reference, also in the legal
domain. This article provides the first systematic overview of TLM-based
methods for AI-driven problems and tasks in the legal sphere. A major goal is
to highlight research advances in this field so as to understand, on the one
hand, how the Transformers have contributed to the success of AI in supporting
legal processes, and on the other hand, what are the current limitations and
opportunities for further research development.Comment: Please refer to the published version: Greco, C.M., Tagarelli, A.
(2023) Bringing order into the realm of Transformer-based language models for
artificial intelligence and law. Artif Intell Law, Springer Nature. November
2023. https://doi.org/10.1007/s10506-023-09374-
SemEval 2023 Task 6: LegalEval -- Understanding Legal Texts
In populous countries, pending legal cases have been growing exponentially.
There is a need for developing NLP-based techniques for processing and
automatically understanding legal documents. To promote research in the area of
Legal NLP we organized the shared task LegalEval - Understanding Legal Texts at
SemEval 2023. LegalEval task has three sub-tasks: Task-A (Rhetorical Roles
Labeling) is about automatically structuring legal documents into semantically
coherent units, Task-B (Legal Named Entity Recognition) deals with identifying
relevant entities in a legal document and Task-C (Court Judgement Prediction
with Explanation) explores the possibility of automatically predicting the
outcome of a legal case along with providing an explanation for the prediction.
In total 26 teams (approx. 100 participants spread across the world) submitted
systems paper. In each of the sub-tasks, the proposed systems outperformed the
baselines; however, there is a lot of scope for improvement. This paper
describes the tasks, and analyzes techniques proposed by various teams.Comment: 13 Pages (9 Pages + References), Accepted at SemEval 202
- …