36,647 research outputs found
Apperceptive patterning: Artefaction, extensional beliefs and cognitive scaffolding
In âPsychopower and Ordinary Madnessâ my ambition, as it relates to Bernard Stieglerâs recent literature, was twofold: 1) critiquing Stieglerâs work on exosomatization and artefactual posthumanismâor, more specifically, nonhumanismâto problematize approaches to media archaeology that rely upon technical exteriorization; 2) challenging how Stiegler engages with Giuseppe Longo and Francis Baillyâs conception of negative entropy. These efforts were directed by a prevalent techno-cultural qualifier: the rise of Synthetic Intelligence (including neural nets, deep learning, predictive processing and Bayesian models of cognition). This paper continues this project but first directs a critical analytic lens at the Derridean practice of the ontologization of grammatization from which Stiegler emerges while also distinguishing how metalanguages operate in relation to object-oriented environmental interaction by way of inferentialism. Stalking continental (Kapp, Simondon, Leroi-Gourhan, etc.) and analytic traditions (e.g., Carnap, Chalmers, Clark, Sutton, Novaes, etc.), we move from artefacts to AI and Predictive Processing so as to link theories related to technicity with philosophy of mind. Simultaneously drawing forth Robert Brandomâs conceptualization of the roles that commitments play in retrospectively reconstructing the social experiences that lead to our endorsement(s) of norms, we compliment this account with Reza Negarestaniâs deprivatized account of intelligence while analyzing the equipollent role between language and media (both digital and analog)
An Introduction to Mechanized Reasoning
Mechanized reasoning uses computers to verify proofs and to help discover new
theorems. Computer scientists have applied mechanized reasoning to economic
problems but -- to date -- this work has not yet been properly presented in
economics journals. We introduce mechanized reasoning to economists in three
ways. First, we introduce mechanized reasoning in general, describing both the
techniques and their successful applications. Second, we explain how mechanized
reasoning has been applied to economic problems, concentrating on the two
domains that have attracted the most attention: social choice theory and
auction theory. Finally, we present a detailed example of mechanized reasoning
in practice by means of a proof of Vickrey's familiar theorem on second-price
auctions
Creationism and evolution
In Tower of Babel, Robert Pennock wrote that
âdefenders of evolution would help their case
immeasurably if they would reassure their
audience that morality, purpose, and meaning are
not lost by accepting the truth of evolution.â We
first consider the thesis that the creationistsâ
movement exploits moral concerns to spread its
ideas against the theory of evolution. We analyze
their arguments and possible reasons why they are
easily accepted. Creationists usually employ two
contradictive strategies to expose the purported
moral degradation that comes with accepting the
theory of evolution. On the one hand they claim
that evolutionary theory is immoral. On the other
hand creationists think of evolutionary theory as
amoral. Both objections come naturally in a
monotheistic view. But we can find similar
conclusions about the supposed moral aspects of
evolution in non-religiously inspired discussions.
Meanwhile, the creationism-evolution debate
mainly focuses â understandably â on what
constitutes good science. We consider the need for
moral reassurance and analyze reassuring
arguments from philosophers. Philosophers may
stress that science does not prescribe and is
therefore not immoral, but this reaction opens the
door for the objection of amorality that evolution
â as a naturalistic world view at least â
supposedly endorses. We consider that the topic of
morality and its relation to the acceptance of
evolution may need more empirical research
Empiricism without Magic: Transformational Abstraction in Deep Convolutional Neural Networks
In artificial intelligence, recent research has demonstrated the remarkable potential of Deep Convolutional Neural Networks (DCNNs), which seem to exceed state-of-the-art performance in new domains weekly, especially on the sorts of very difficult perceptual discrimination tasks that skeptics thought would remain beyond the reach of artificial intelligence. However, it has proven difficult to explain why DCNNs perform so well. In philosophy of mind, empiricists have long suggested that complex cognition is based on information derived from sensory experience, often appealing to a faculty of abstraction. Rationalists have frequently complained, however, that empiricists never adequately explained how this faculty of abstraction actually works. In this paper, I tie these two questions together, to the mutual benefit of both disciplines. I argue that the architectural features that distinguish DCNNs from earlier neural networks allow them to implement a form of hierarchical processing that I call âtransformational abstractionâ. Transformational abstraction iteratively converts sensory-based representations of category exemplars into new formats that are increasingly tolerant to ânuisance variationâ in input. Reflecting upon the way that DCNNs leverage a combination of linear and non-linear processing to efficiently accomplish this feat allows us to understand how the brain is capable of bi-directional travel between exemplars and abstractions, addressing longstanding problems in empiricist philosophy of mind. I end by considering the prospects for future research on DCNNs, arguing that rather than simply implementing 80s connectionism with more brute-force computation, transformational abstraction counts as a qualitatively distinct form of processing ripe with philosophical and psychological significance, because it is significantly better suited to depict the generic mechanism responsible for this important kind of psychological processing in the brain
The Dismissal of âSubstanceâ and âBeingâ in Peirceâs Regenerated Logic
After introducing the debate between substance philosophy and process philosophy, and clarifying the relevance of the category of âsubstanceâ in Peirceâs thought, the present paper reconstructs the role of âsubstanceâ and âbeingâ from Peirceâs early works to his theory of the proposition, provided after his studies on the logic of relatives. If those two categories apparently disappear in Peirceâs writings from the mid-1890s onwards, the account of âsubjectâ and âcopulaâ in Peirceâs analysis of the proposition allows one to grasp the reasons why Peirce omits âsubstanceâ and âbeingâ in favor of his three categories (Firstness, Secondness, Thirdness), and to understand why his philosophy cannot be considered as a substance philosophy
The propositional nature of human associative learning
The past 50 years have seen an accumulation of evidence suggesting that associative learning depends oil high-level cognitive processes that give rise to propositional knowledge. Yet, many learning theorists maintain a belief in a learning mechanism in which links between mental representations are formed automatically. We characterize and highlight the differences between the propositional and link approaches, and review the relevant empirical evidence. We conclude that learning is the consequence of propositional reasoning processes that cooperate with the unconscious processes involved in memory retrieval and perception. We argue that this new conceptual framework allows many of the important recent advances in associative learning research to be retained, but recast in a model that provides a firmer foundation for both immediate application and future research
Pass it on: towards a political economy of propensity
The paper argues that the work of Gabriel Tarde on imitation provides a fertile means of understanding how capitalism is forging a new affective technology which conforms to a logic of propensity rather than to means-end reasoning. This it does by drawing together a biological understanding of semiconscious cognition with various practical geometric arts so as to re-stage the world as a series of susceptible situations which can be ridden rather than rigidly controlled. The paper examines the advent of technologies which attend to the variable geometry of so-called animal spirits in the realm of business and then, using Tarde's work as a springboard, considers some alternative means of understanding imitative rays which have less instrumental undertones. The paper is an illustration of the way in which biology and culture have increasingly become intertwined
- âŠ