85,331 research outputs found
Using Toulmin Argumentation to develop an Online Dispute Resolution Environment
Our goal is to model reasoning in discretionary legal domains. To do so, we use Knowledge Discovery from Database Techniques. However there are obstacles to this approach, including difficulties in generating explanations once conclusions have been inferred, difficulties associated with the collection of sufficient data from past cases and difficulties associated with integrating two vastly different paradigms. Toulminâs treatise on the uses of argument can be gainfully employed to construct legal decision support systems in discretionary domains. We show how we can use Toulminâs approach to build such systems with examples taken from the domains of eligibility for legal aid, evaluation of eyewitness evidence, family law, refugee law and sentencing. We then show how Toulmin Argument Structures can be developed to construct an Online Dispute Resolution environment that allows for determining BATNAs, exchanging opinions and providing advice about tradeoffs
Recommended from our members
An Ontological formalization of the planning task
In this paper we propose a generic task ontology, which formalizes the space of planning problems. Although planning is one of the oldest researched areas in Artificial Intelligence and attempts have been made in the past at developing task ontologies for planning, these formalizations suffer from serious limitations: they do not exhibit the required level of formalization and precision and they usually fail to include some of the key concepts required for specifying planning problems. In con-trast with earlier proposals, our task ontology formalizes the nature of the planning task independently of any planning paradigm, specific domains, or applications and provides a fine-grained, precise and comprehensive characterization of the space of planning problems. Finally, in addition to producing a formal specification we have also operationalized the ontology into a set of executable definitions, which provide a concrete reusable resource for knowledge acquisition and system development in planning applications
New Economic Analysis of Law: Beyond Technocracy and Market Design
This special issue on New Economic Analysis of Law features illuminating syntheses of social science and law. What would law and economics look like if macroeconomics were a concern of scholars now focused entirely on microeconomics? Do emerging online phenomena, such as algorithmic pricing and platform capitalism, promise to perfect economic theories of market equilibrium, or challenge their foundations? How did simplified economic models gain ideological power in policy circles, and how can they be improved or replaced? This issue highlights scholars whose work has made the legal academy more than an âimporterâ of ideas from other disciplinesâand who have, instead, shown that rigorous legal analysis is fundamental to understanding economic affairs.The essays in this issue should help ensure that policymakersâ turn to new economic thinking promotes inclusive prosperity. Listokin, Bayern, and Kwak have identified major aporias in popular applications of law and economics methods. RanchordĂĄs, Stucke, and Ezrachi have demonstrated that technological fixes, ranging from digital ranking and rating systems to artificial intelligence-driven personal assistants, are unlikely to improve matters unless they are wisely regulated. McCluskey and Rahman offer a blueprint for democratic regulation, which shapes the economy in productive ways and alleviates structural inequalities. Taken as a whole, this issue of Critical Analysis of Law shows that legal thinkers are not merely importers of ideas and models from economics, but also active participants, with a great deal to contribute to social science research
Recommended from our members
Some reflections on the legitimacy of international trial justice
This paper addresses a number of interrelated conceptual difficulties that impact adversely on the ability of international criminal trials to deliver outcomes perceived as legitimate by victims and communities in post-conflict states. It begins by exploring the extent to which those moral justifications for punishment espoused by international courts are instrumental in marginalizing the aspirations for justice of victims and victim communities, and suggests how a greater appreciation of the sociological context of punishing international crimes can contribute towards an improved understanding of normative practice. The paper then examines the relationship between perceptions of international crime and punishment, and the broader issue of whether international criminal law provides an appropriate normative structure for giving effect to those universal humanitarian values concerned with punishment in an increasingly pluralistic world. Finally, the paper considers how the theory and practice of punishing international crimes can more effectively satisfy both local and global aspirations for post-conflict justice through enhancing the transformative capacity of international criminal trials
Pushing the bounds of rationality: Argumentation and extended cognition
One of the central tasks of a theory of argumentation is to supply a theory of appraisal: a set of standards and norms according to which argumentation, and the reasoning involved in it, is properly evaluated. In their most general form, these can be understood as rational norms, where the core idea of rationality is that we rightly respond to reasons by according the credence we attach to our doxastic and conversational commitments with the probative strength of the reasons we have for them. Certain kinds of rational failings are so because they are manifestly illogical â for example, maintaining overtly contradictory commitments, violating deductive closure by refusing to accept the logical consequences of oneâs present commitments, or failing to track basing relations by not updating oneâs commitments in view of new, defeating information. Yet, according to the internal and empirical critiques, logic and probability theory fail to supply a fit set of norms for human reasoning and argument. Particularly, theories of bounded rationality have put pressure on argumentation theory to lower the normative standards of rationality for reasoners and arguers on the grounds that we are bounded, finite, and fallible agents incapable of meeting idealized standards. This paper explores the idea that argumentation, as a set of practices, together with the procedures and technologies of argumentation theory, is able to extend cognition such that we are better able to meet these idealized logical standards, thereby extending our responsibilities to adhere to idealized rational norms
Optimization, Path Dependence and the Law: Can Judges Promote Efficiency?
The thesis that judges could (voluntarily or not) promote efficiency through their decisions has largely been discussed since Posner put it forward in the early 1970s. There nonetheless remains a methodological aspect that has never (to our knowledge) been analyzed and that we address in this paper. We thus show that both promoters and critics of the judge-and-efficiency thesis similarly use a definition of optimization in which history, constraints and path-dependency are viewed as obstacles that must be removed to reach the most efficient outcome. This is misleading. Efficiency cannot be defined in absolute terms, as a âglobal idealâ that would mean being free from any constraint, including historically deposited ones. That judges are obliged to refer to the past does not mean that they are unable to make the most efficient decision because constraints are part of the optimization process; or optimization is necessarily path- dependent. Thus, the output of legal systems cannot be efficient or inefficient per se. This is what we argue in this paper.Judicial decision making; Historical inertia; Inefficiency; Adaptationism; Spandrelism; Global ideal; Rationality; Lock-in institutions.
Designing Normative Theories for Ethical and Legal Reasoning: LogiKEy Framework, Methodology, and Tool Support
A framework and methodology---termed LogiKEy---for the design and engineering
of ethical reasoners, normative theories and deontic logics is presented. The
overall motivation is the development of suitable means for the control and
governance of intelligent autonomous systems. LogiKEy's unifying formal
framework is based on semantical embeddings of deontic logics, logic
combinations and ethico-legal domain theories in expressive classic
higher-order logic (HOL). This meta-logical approach enables the provision of
powerful tool support in LogiKEy: off-the-shelf theorem provers and model
finders for HOL are assisting the LogiKEy designer of ethical intelligent
agents to flexibly experiment with underlying logics and their combinations,
with ethico-legal domain theories, and with concrete examples---all at the same
time. Continuous improvements of these off-the-shelf provers, without further
ado, leverage the reasoning performance in LogiKEy. Case studies, in which the
LogiKEy framework and methodology has been applied and tested, give evidence
that HOL's undecidability often does not hinder efficient experimentation.Comment: 50 pages; 10 figure
- âŠ