54,617 research outputs found
Some Computational Aspects of Essential Properties of Evolution and Life
While evolution has inspired algorithmic methods of heuristic optimisation, little has been done in the way of using concepts of computation to advance our understanding of salient aspects of biological evolution. We argue that under reasonable assumptions, interesting conclusions can be drawn that are of relevance to behavioural evolution. We will focus on two important features of life--robustness and fitness optimisation--which, we will argue, are related to algorithmic probability and to the thermodynamics of computation, subjects that may be capable of explaining and modelling key features of living organisms, and which can be used in understanding and formulating algorithms of evolutionary computation
The Algorithmic Origins of Life
Although it has been notoriously difficult to pin down precisely what it is
that makes life so distinctive and remarkable, there is general agreement that
its informational aspect is one key property, perhaps the key property. The
unique informational narrative of living systems suggests that life may be
characterized by context-dependent causal influences, and in particular, that
top-down (or downward) causation -- where higher-levels influence and constrain
the dynamics of lower-levels in organizational hierarchies -- may be a major
contributor to the hierarchal structure of living systems. Here we propose that
the origin of life may correspond to a physical transition associated with a
shift in causal structure, where information gains direct, and
context-dependent causal efficacy over the matter it is instantiated in. Such a
transition may be akin to more traditional physical transitions (e.g.
thermodynamic phase transitions), with the crucial distinction that determining
which phase (non-life or life) a given system is in requires dynamical
information and therefore can only be inferred by identifying causal
architecture. We discuss some potential novel research directions based on this
hypothesis, including potential measures of such a transition that may be
amenable to laboratory study, and how the proposed mechanism corresponds to the
onset of the unique mode of (algorithmic) information processing characteristic
of living systems.Comment: 13 pages, 1 tabl
Algorithmic Fairness from a Non-ideal Perspective
Inspired by recent breakthroughs in predictive modeling, practitioners in both industry and government have turned to machine learning with hopes of operationalizing predictions to drive automated decisions. Unfortunately, many social desiderata concerning consequential decisions, such as justice or fairness, have no natural formulation within a purely predictive framework. In efforts to mitigate these problems, researchers have proposed a variety of metrics for quantifying deviations from various statistical parities that we might expect to observe in a fair world and offered a variety of algorithms in attempts to satisfy subsets of these parities or to trade o the degree to which they are satised against utility. In this paper, we connect this approach to fair machine learning to the literature on ideal and non-ideal methodological approaches in political philosophy. The ideal approach requires positing the principles according to which a just world would operate. In the most straightforward application of ideal theory, one supports a proposed policy by arguing that it closes a discrepancy between the real and the perfectly just world. However, by failing to account for the mechanisms by which our non-ideal world arose, the responsibilities of various decision-makers, and the impacts of proposed policies, naive applications of ideal thinking can lead to misguided interventions. In this paper, we demonstrate a connection between the fair machine learning literature and the ideal approach in political philosophy, and argue that the increasingly apparent shortcomings of proposed fair machine learning algorithms reflect broader troubles
faced by the ideal approach. We conclude with a critical discussion of the harms of misguided solutions, a
reinterpretation of impossibility results, and directions for future researc
Automatic differentiation in machine learning: a survey
Derivatives, mostly in the form of gradients and Hessians, are ubiquitous in
machine learning. Automatic differentiation (AD), also called algorithmic
differentiation or simply "autodiff", is a family of techniques similar to but
more general than backpropagation for efficiently and accurately evaluating
derivatives of numeric functions expressed as computer programs. AD is a small
but established field with applications in areas including computational fluid
dynamics, atmospheric sciences, and engineering design optimization. Until very
recently, the fields of machine learning and AD have largely been unaware of
each other and, in some cases, have independently discovered each other's
results. Despite its relevance, general-purpose AD has been missing from the
machine learning toolbox, a situation slowly changing with its ongoing adoption
under the names "dynamic computational graphs" and "differentiable
programming". We survey the intersection of AD and machine learning, cover
applications where AD has direct relevance, and address the main implementation
techniques. By precisely defining the main differentiation techniques and their
interrelationships, we aim to bring clarity to the usage of the terms
"autodiff", "automatic differentiation", and "symbolic differentiation" as
these are encountered more and more in machine learning settings.Comment: 43 pages, 5 figure
- …