110,494 research outputs found
On the Algorithmic Nature of the World
We propose a test based on the theory of algorithmic complexity and an
experimental evaluation of Levin's universal distribution to identify evidence
in support of or in contravention of the claim that the world is algorithmic in
nature. To this end we have undertaken a statistical comparison of the
frequency distributions of data from physical sources on the one
hand--repositories of information such as images, data stored in a hard drive,
computer programs and DNA sequences--and the frequency distributions generated
by purely algorithmic means on the other--by running abstract computing devices
such as Turing machines, cellular automata and Post Tag systems. Statistical
correlations were found and their significance measured.Comment: Book chapter in Gordana Dodig-Crnkovic and Mark Burgin (eds.)
Information and Computation by World Scientific, 2010.
(http://www.idt.mdh.se/ECAP-2005/INFOCOMPBOOK/). Paper website:
http://www.mathrix.org/experimentalAIT
Picturing algorithmic surveillance: the politics of facial recognition systems
This paper opens up for scrutiny the politics of algorithmic surveillance through an examination of Facial Recognition Systems (FRS's) in video surveillance, showing that seemingly mundane design decisions may have important political consequences that ought to be subject to scrutiny. It first focuses on the politics of technology and algorithmic surveillance systems in particular: considering the broad politics of technology; the nature of algorithmic surveillance and biometrics, claiming that software algorithms are a particularly important domain of techno-politics; and finally considering both the growth of algorithmic biometric surveillance and the potential problems with such systems. Secondly, it gives an account of FRS's, the
algorithms upon which they are based, and the biases embedded therein. In the third part, the ways in which these biases may manifest itself in real world implementation of FRS’s are outlined. Finally, some policy suggestions for the future development of FRS’s are made; it is noted that the most common critiques of such systems are based on notions of privacy which seem increasingly at odds with the world of automated systems
Reframing the Horizon within the Algorithmic Landscape of Northern Britain
Emerging from the artist’s constructed photographs and walking projects in the north, this paper considers the tension between the photograph as a fixed composition of the world and the dynamic image constructed from data. Whereas arguably, the traditional photograph exhibits a stable relationship between the world and the image, the constructed photograph shifts the focus onto the underlying algorithmic processes of production. This focus on the relational nature of the constructed photograph shifts our gaze from the horizon to the underlying systems in operation as we consider the relational nature of data as a photograph
Is Evolution Algorithmic?
In Darwin’s Dangerous Idea, Daniel Dennett claims that evolution is algorithmic. On Dennett’s analysis, evolutionary processes are trivially algorithmic because he assumes that all natural processes are algorithmic. I will argue that there are more robust ways to understand algorithmic processes that make the claim that evolution is algorithmic empirical and not conceptual. While laws of nature can be seen as compression algorithms of information about the world, it does not follow logically that they are implemented as algorithms by physical processes. For that to be true, the processes have to be part of computational systems. The basic difference between mere simulation and real computing is having proper causal structure. I will show what kind of requirements this poses for natural evolutionary processes if they are to be computational
Impact Remediation: Optimal Interventions to Reduce Inequality
A significant body of research in the data sciences considers unfair
discrimination against social categories such as race or gender that could
occur or be amplified as a result of algorithmic decisions. Simultaneously,
real-world disparities continue to exist, even before algorithmic decisions are
made. In this work, we draw on insights from the social sciences and humanistic
studies brought into the realm of causal modeling and constrained optimization,
and develop a novel algorithmic framework for tackling pre-existing real-world
disparities. The purpose of our framework, which we call the "impact
remediation framework," is to measure real-world disparities and discover the
optimal intervention policies that could help improve equity or access to
opportunity for those who are underserved with respect to an outcome of
interest. We develop a disaggregated approach to tackling pre-existing
disparities that relaxes the typical set of assumptions required for the use of
social categories in structural causal models. Our approach flexibly
incorporates counterfactuals and is compatible with various ontological
assumptions about the nature of social categories. We demonstrate impact
remediation with a real-world case study and compare our disaggregated approach
to an existing state-of-the-art approach, comparing its structure and resulting
policy recommendations. In contrast to most work on optimal policy learning, we
explore disparity reduction itself as an objective, explicitly focusing the
power of algorithms on reducing inequality
Disaggregated interventions to reduce inequality
A significant body of research in the data sciences considers unfair discrimination against social categories such as race or gender that could occur or be amplified as a result of algorithmic decisions. Simultaneously, real-world disparities continue to exist, even before algorithmic decisions are made. In this work, we draw on insights from the social sciences brought into the realm of causal modeling and constrained optimization, and develop a novel algorithmic framework for tackling pre-existing real-world disparities. The purpose of our framework, which we call the "impact remediation framework,"is to measure real-world disparities and discover the optimal intervention policies that could help improve equity or access to opportunity for those who are underserved with respect to an outcome of interest. We develop a disaggregated approach to tackling pre-existing disparities that relaxes the typical set of assumptions required for the use of social categories in structural causal models. Our approach flexibly incorporates counterfactuals and is compatible with various ontological assumptions about the nature of social categories. We demonstrate impact remediation with a hypothetical case study and compare our disaggregated approach to an existing state-of-the-art approach, comparing its structure and resulting policy recommendations. In contrast to most work on optimal policy learning, we explore disparity reduction itself as an objective, explicitly focusing the power of algorithms on reducing inequality
Minimal Algorithmic Information Loss Methods for Dimension Reduction, Feature Selection and Network Sparsification
We introduce a family of unsupervised, domain-free, and (asymptotically)
model-independent algorithms based on the principles of algorithmic probability
and information theory designed to minimize the loss of algorithmic
information, including a lossless-compression-based lossy compression
algorithm. The methods can select and coarse-grain data in an
algorithmic-complexity fashion (without the use of popular compression
algorithms) by collapsing regions that may procedurally be regenerated from a
computable candidate model. We show that the method can preserve the salient
properties of objects and perform dimension reduction, denoising, feature
selection, and network sparsification. As validation case, we demonstrate that
the method preserves all the graph-theoretic indices measured on a well-known
set of synthetic and real-world networks of very different nature, ranging from
degree distribution and clustering coefficient to edge betweenness and degree
and eigenvector centralities, achieving equal or significantly better results
than other data reduction and some of the leading network sparsification
methods. The methods (InfoRank, MILS) can also be applied to applications such
as image segmentation based on algorithmic probability.Comment: 23 pages in double column including Appendix, online implementation
at http://complexitycalculator.com/MILS
- …