5 research outputs found

    Legal NERC with ontologies, Wikipedia and curriculum learning

    Get PDF
    International audienceIn this paper, we present a Wikipedia-based approach to develop resources for the legal domain. We establish a mapping between a legal domain ontology, LKIF (Hoekstra et al., 2007), and a Wikipedia-based ontology, YAGO (Suchanek et al., 2007), and through that we populate LKIF. Moreover, we use the mentions of those entities in Wikipedia text to train a specific Named Entity Recognizer and Classifier. We find that this classifier works well in the Wikipedia, but, as could be expected, performance decreases in a corpus of judgments of the European Court of Human Rights. However, this tool will be used as a preprocess for human annotation. We resort to a technique called curriculum learning aimed to overcome problems of overfitting by learning increasingly more complex concepts. However, we find that in this particular setting, the method works best by learning from most specific to most general concepts, not the other way round

    Intelligent Case Assignment Method Based on the Chain of Criminal Behavior Elements

    Get PDF
    The assignment of cases means the court assigns cases to specific judges. The traditional case assignment methods, based on the facts of a case, are weak in the analysis of semantic structure of the case not considering the judges\u27 expertise. By analyzing judges\u27 trial logic, we find that the order of criminal behaviors affects the final judgement. To solve these problems, we regard intelligent case assignment as a text-matching problem, and propose an intelligent case assignment method based on the chain of criminal behavior elements. This method introduces the chain of criminal behavior elements to enhance the structured semantic analysis of the case. We build a BCTA (Bert-Cnn-Transformer-Attention) model to achieve intelligent case assignment. This model integrates a judge\u27s expertise in the judge\u27s presentation, thus recommending the most compatible judge for the case. Comparing the traditional case assignment methods, our BCTA model obtains 84% absolutely considerable improvement under P@1. In addition, comparing other classic text matching models, our BCTA model achieves an absolute considerable improvement of 4% under P@1 and 9% under Macro F1. Experiments conducted on real-world data set demonstrate the superiority of our method

    A Low-cost, High-coverage Legal Named Entity Recognizer, Classifier and Linker

    Get PDF
    International audienceIn this paper, we try to improve Information Extraction in legal texts by creating a legal Named Entity Recognizer, Classifier and Linker. With this tool, we can identify relevant parts of texts and connect them to a structured knowledge representation, the LKIF ontology.More interestingly, this tool has been developed with relatively little effort, by mapping the LKIF ontology to the YAGO ontology and through it, taking advantage of the mentions of entities in the Wikipedia. These mentions are used as manually annotated examples to train the Named Entity Recognizer, Classifier and Linker.We have evaluated the approach on holdout texts from the Wikipedia and also on a small sample of judgments of the European Court of Human Rights, resulting in a very good performance, i.e., around 80% F-measure for different levels of granularity. We present an extensive error analysis to direct further developments, and we expect that this approach can be successfully ported to other legal subdomains, represented by different ontologies
    corecore