548 research outputs found
The Emergence of Successful Export Activities in Mexico: Three Case Studies
This paper consists of three case studies of the emergence of three successful export activities in Mexico: avocado production, the manufacture of catheters, and call center outsourcing. Each case study discusses how companies, associations, and governments at various levels have addressed market failures and facilitated the provision of public goods necessary for each activity. The case studies additionally profile first movers in each activity and describe the positive externalities they provide to imitators, particularly diffusion of export knowledge. Also include in each case study is a counterfactual case of a less successful activity (mangos, stem cell banking, and other types of business process outsourcing, respectively) and a section on policy implications.Agriculture, Exports, Manufacturing, Services, Mexico
Multidisciplinary perspectives on Artificial Intelligence and the law
This open access book presents an interdisciplinary, multi-authored, edited collection of chapters on Artificial Intelligence (‘AI’) and the Law. AI technology has come to play a central role in the modern data economy. Through a combination of increased computing power, the growing availability of data and the advancement of algorithms, AI has now become an umbrella term for some of the most transformational technological breakthroughs of this age. The importance of AI stems from both the opportunities that it offers and the challenges that it entails. While AI applications hold the promise of economic growth and efficiency gains, they also create significant risks and uncertainty. The potential and perils of AI have thus come to dominate modern discussions of technology and ethics – and although AI was initially allowed to largely develop without guidelines or rules, few would deny that the law is set to play a fundamental role in shaping the future of AI. As the debate over AI is far from over, the need for rigorous analysis has never been greater. This book thus brings together contributors from different fields and backgrounds to explore how the law might provide answers to some of the most pressing questions raised by AI. An outcome of the Católica Research Centre for the Future of Law and its interdisciplinary working group on Law and Artificial Intelligence, it includes contributions by leading scholars in the fields of technology, ethics and the law.info:eu-repo/semantics/publishedVersio
Research in the Language, Information and Computation Laboratory of the University of Pennsylvania
This report takes its name from the Computational Linguistics Feedback Forum (CLiFF), an informal discussion group for students and faculty. However the scope of the research covered in this report is broader than the title might suggest; this is the yearly report of the LINC Lab, the Language, Information and Computation Laboratory of the University of Pennsylvania.
It may at first be hard to see the threads that bind together the work presented here, work by faculty, graduate students and postdocs in the Computer Science and Linguistics Departments, and the Institute for Research in Cognitive Science. It includes prototypical Natural Language fields such as: Combinatorial Categorial Grammars, Tree Adjoining Grammars, syntactic parsing and the syntax-semantics interface; but it extends to statistical methods, plan inference, instruction understanding, intonation, causal reasoning, free word order languages, geometric reasoning, medical informatics, connectionism, and language acquisition.
Naturally, this introduction cannot spell out all the connections between these abstracts; we invite you to explore them on your own. In fact, with this issue it’s easier than ever to do so: this document is accessible on the “information superhighway”. Just call up http://www.cis.upenn.edu/~cliff-group/94/cliffnotes.html
In addition, you can find many of the papers referenced in the CLiFF Notes on the net. Most can be obtained by following links from the authors’ abstracts in the web version of this report.
The abstracts describe the researchers’ many areas of investigation, explain their shared concerns, and present some interesting work in Cognitive Science. We hope its new online format makes the CLiFF Notes a more useful and interesting guide to Computational Linguistics activity at Penn
Transdisciplinary AI Observatory -- Retrospective Analyses and Future-Oriented Contradistinctions
In the last years, AI safety gained international recognition in the light of
heterogeneous safety-critical and ethical issues that risk overshadowing the
broad beneficial impacts of AI. In this context, the implementation of AI
observatory endeavors represents one key research direction. This paper
motivates the need for an inherently transdisciplinary AI observatory approach
integrating diverse retrospective and counterfactual views. We delineate aims
and limitations while providing hands-on-advice utilizing concrete practical
examples. Distinguishing between unintentionally and intentionally triggered AI
risks with diverse socio-psycho-technological impacts, we exemplify a
retrospective descriptive analysis followed by a retrospective counterfactual
risk analysis. Building on these AI observatory tools, we present near-term
transdisciplinary guidelines for AI safety. As further contribution, we discuss
differentiated and tailored long-term directions through the lens of two
disparate modern AI safety paradigms. For simplicity, we refer to these two
different paradigms with the terms artificial stupidity (AS) and eternal
creativity (EC) respectively. While both AS and EC acknowledge the need for a
hybrid cognitive-affective approach to AI safety and overlap with regard to
many short-term considerations, they differ fundamentally in the nature of
multiple envisaged long-term solution patterns. By compiling relevant
underlying contradistinctions, we aim to provide future-oriented incentives for
constructive dialectics in practical and theoretical AI safety research
Neuroeconomics: How Neuroscience Can Inform Economics
Neuroeconomics uses knowledge about brain mechanisms to inform economic analysis, and roots economics in biology. It opens up the "black box" of the brain, much as organizational economics adds detail to the theory of the firm. Neuroscientists use many tools— including brain imaging, behavior of patients with localized brain lesions, animal behavior, and recording single neuron activity. The key insight for economics is that the brain is composed of multiple systems which interact. Controlled systems ("executive function") interrupt automatic ones. Emotions and cognition both guide decisions. Just as prices and allocations emerge from the interaction of two processes—supply and demand— individual decisions can be modeled as the result of two (or more) processes interacting. Indeed, "dual-process" models of this sort are better rooted in neuroscientific fact, and more empirically accurate, than single-process models (such as utility-maximization). We discuss how brain evidence complicates standard assumptions about basic preference, to include homeostasis and other kinds of state-dependence. We also discuss applications to intertemporal choice, risk and decision making, and game theory. Intertemporal choice appears to be domain-specific and heavily influenced by emotion. The simplified ß-d of quasi-hyperbolic discounting is supported by activation in distinct regions of limbic and cortical systems. In risky decision, imaging data tentatively support the idea that gains and losses are coded separately, and that ambiguity is distinct from risk, because it activates fear and discomfort regions. (Ironically, lesion patients who do not receive fear signals in prefrontal cortex are "rationally" neutral toward ambiguity.) Game theory studies show the effect of brain regions implicated in "theory of mind", correlates of strategic skill, and effects of hormones and other biological variables. Finally, economics can contribute to neuroscience because simple rational-choice models are useful for understanding highly-evolved behavior like motor actions that earn rewards, and Bayesian integration of sensorimotor information
Responsible and Regulatory Conform Machine Learning for Medicine: A Survey of Challenges and Solutions
Machine learning is expected to fuel significant improvements in medical
care. To ensure that fundamental principles such as beneficence, respect for
human autonomy, prevention of harm, justice, privacy, and transparency are
respected, medical machine learning systems must be developed responsibly. Many
high-level declarations of ethical principles have been put forth for this
purpose, but there is a severe lack of technical guidelines explicating the
practical consequences for medical machine learning. Similarly, there is
currently considerable uncertainty regarding the exact regulatory requirements
placed upon medical machine learning systems. This survey provides an overview
of the technical and procedural challenges involved in creating medical machine
learning systems responsibly and in conformity with existing regulations, as
well as possible solutions to address these challenges. First, a brief review
of existing regulations affecting medical machine learning is provided, showing
that properties such as safety, robustness, reliability, privacy, security,
transparency, explainability, and nondiscrimination are all demanded already by
existing law and regulations - albeit, in many cases, to an uncertain degree.
Next, the key technical obstacles to achieving these desirable properties are
discussed, as well as important techniques to overcome these obstacles in the
medical context. We notice that distribution shift, spurious correlations,
model underspecification, uncertainty quantification, and data scarcity
represent severe challenges in the medical context. Promising solution
approaches include the use of large and representative datasets and federated
learning as a means to that end, the careful exploitation of domain knowledge,
the use of inherently transparent models, comprehensive out-of-distribution
model testing and verification, as well as algorithmic impact assessments
Counterfactual reasoning in strategy context : a theoretical investigation of the role of hindsight in strategic foresight
The purpose of this doctoral thesis is to deepen theoretical understanding of the role that hindsight plays in foresight. The thesis argues that the past is not an isolated static state, but one that is intimately connected with the future. However, there are several biases that influence our perceptions and conceptions of the past. These biases act as constraints on strategic learning by limiting our ability to understand the driving forces that emerge from the past, play out through the present and become critical uncertainties in the future. They can result in misperceptions about events or processes, and as such, may impair foresight methodologies such as scenario thinking. Such foresightful thinking flaws are characterised by a combination of hindsight biases and creeping determinism, which result in searching for information that corresponds to people's views about both the past and the future, logical path-dependencies, misaligned dominant logics, routines, recipes and paradigms, and over-confidence and defensive pessimism. Drawing on received research in psychology, the role of counter-to-factual reasoning as a heuristic is discussed and analysed as a possible antidote to foresightful thinking flaws. The judicious use of such a heuristic device as counterfactual reasoning, both as a sense-making process and as an analytical reasoning tool applied to the analysis of historical data, the thesis concludes, is a method for investigating and discovering the past and fortifying foresightful strategic thinking
Automatically Correcting Large Language Models: Surveying the landscape of diverse self-correction strategies
Large language models (LLMs) have demonstrated remarkable performance across
a wide array of NLP tasks. However, their efficacy is undermined by undesired
and inconsistent behaviors, including hallucination, unfaithful reasoning, and
toxic content. A promising approach to rectify these flaws is self-correction,
where the LLM itself is prompted or guided to fix problems in its own output.
Techniques leveraging automated feedback -- either produced by the LLM itself
or some external system -- are of particular interest as they are a promising
way to make LLM-based solutions more practical and deployable with minimal
human feedback. This paper presents a comprehensive review of this emerging
class of techniques. We analyze and taxonomize a wide array of recent work
utilizing these strategies, including training-time, generation-time, and
post-hoc correction. We also summarize the major applications of this strategy
and conclude by discussing future directions and challenges.Comment: Work in Progress. Version
- …