4,557 research outputs found

    From conditioning to learning communities: Implications of fifty years of research in e‐learning interaction design

    Get PDF
    This paper will consider e‐learning in terms of the underlying learning processes and interactions that are stimulated, supported or favoured by new media and the contexts or communities in which it is used. We will review and critique a selection of research and development from the past fifty years that has linked pedagogical and learning theory to the design of innovative e‐learning systems and activities, and discuss their implications. It will include approaches that are, essentially, behaviourist (Skinner and Gagné), cognitivist (Pask, Piaget and Papert), situated (Lave, Wenger and Seely‐Brown), socio‐constructivist (Vygotsky), socio‐cultural (Nardi and Engestrom) and community‐based (Wenger and Preece). Emerging from this review is the argument that effective e‐learning usually requires, or involves, high‐quality educational discourse, that leads to, at the least, improved knowledge, and at the best, conceptual development and improved understanding. To achieve this I argue that we need to adopt a more holistic approach to design that synthesizes features of the included approaches, leading to a framework that emphasizes the relationships between cognitive changes, dialogue processes and the communities, or contexts for e‐learning

    ‘Putting apes (body and language) together again’, a review article of Savage-Rumbaugh, S., Taylor, T. J., and Shanker, S. G. Apes, Language, and the Human Mind (Oxford: 1999) and Clark, A. Being There: Putting Brain, Body, and World Together Again (MIT: 1997)

    Get PDF
    It is argued that the account of Savage-Rumbaugh’s ape language research in Savage-Rumbaugh, Shanker and Taylor (1998. Apes, Language and the Human Mind. Oxford University Press, Oxford) is profitably read in the terms of the theoretical perspective developed in Clark (1997. Being There, Putting Brain, Body and World Together Again. MIT Press, Cambridge, MA). The former work details some striking results concerning chimpanzee and bonobo subjects, trained to make use of keyboards containing ‘lexigram’ symbols. The authors, though, make heavy going of a critique of what they take to be standard approaches to understanding language and cognition in animals, and fail to offer a worthwhile theoretical position from which to make sense of their own data. It is suggested that the achievements of Savage-Rumbaugh’s non-human subjects suggest that language ability need not be explained by reference to specialised brain capacities. The contribution made by Clark’s work is to show the range of ways in which cognition exploits bodily and environmental resources. This model of ‘distributed’ cognition helps makes sense of the lexigram activity of Savage-Rumbaugh’s subjects, and points to a re-evaluation of the language behaviour of humans

    USER-AWARENESS AND ADAPTATION IN CONVERSATIONAL AGENTS

    Get PDF
    This paper considers the research question of developing user-aware and adaptive conversational agents. The conversational agent is a system which is user-aware to the extent that it recognizes the user identity and his/her emotional states that are relevant in a given interaction domain. The conversational agent is user-adaptive to the extent that it dynamically adapts its dialogue behavior according to the user and his/her emotional state. The paper summarizes some aspects of our previous work and presents work-in-progress in the field of speech-based human-machine interaction. It focuses particularly on the development of speech recognition modules in cooperation with both modules for emotion recognition and speaker recognition, as well as the dialogue management module. Finally, it proposes an architecture of a conversational agent that integrates those modules and improves each of them based on some kind of synergies among themselves

    Conceptual Spaces for Cognitive Architectures: A Lingua Franca for Different Levels of Representation

    Get PDF
    During the last decades, many cognitive architectures (CAs) have been realized adopting different assumptions about the organization and the representation of their knowledge level. Some of them (e.g. SOAR [35]) adopt a classical symbolic approach, some (e.g. LEABRA[ 48]) are based on a purely connectionist model, while others (e.g. CLARION [59]) adopt a hybrid approach combining connectionist and symbolic representational levels. Additionally, some attempts (e.g. biSOAR) trying to extend the representational capacities of CAs by integrating diagrammatical representations and reasoning are also available [34]. In this paper we propose a reflection on the role that Conceptual Spaces, a framework developed by Peter G¨ardenfors [24] more than fifteen years ago, can play in the current development of the Knowledge Level in Cognitive Systems and Architectures. In particular, we claim that Conceptual Spaces offer a lingua franca that allows to unify and generalize many aspects of the symbolic, sub-symbolic and diagrammatic approaches (by overcoming some of their typical problems) and to integrate them on a common ground. In doing so we extend and detail some of the arguments explored by G¨ardenfors [23] for defending the need of a conceptual, intermediate, representation level between the symbolic and the sub-symbolic one. In particular we focus on the advantages offered by Conceptual Spaces (w.r.t. symbolic and sub-symbolic approaches) in dealing with the problem of compositionality of representations based on typicality traits. Additionally, we argue that Conceptual Spaces could offer a unifying framework for interpreting many kinds of diagrammatic and analogical representations. As a consequence, their adoption could also favor the integration of diagrammatical representation and reasoning in CAs

    Apperceptive patterning: Artefaction, extensional beliefs and cognitive scaffolding

    Get PDF
    In “Psychopower and Ordinary Madness” my ambition, as it relates to Bernard Stiegler’s recent literature, was twofold: 1) critiquing Stiegler’s work on exosomatization and artefactual posthumanism—or, more specifically, nonhumanism—to problematize approaches to media archaeology that rely upon technical exteriorization; 2) challenging how Stiegler engages with Giuseppe Longo and Francis Bailly’s conception of negative entropy. These efforts were directed by a prevalent techno-cultural qualifier: the rise of Synthetic Intelligence (including neural nets, deep learning, predictive processing and Bayesian models of cognition). This paper continues this project but first directs a critical analytic lens at the Derridean practice of the ontologization of grammatization from which Stiegler emerges while also distinguishing how metalanguages operate in relation to object-oriented environmental interaction by way of inferentialism. Stalking continental (Kapp, Simondon, Leroi-Gourhan, etc.) and analytic traditions (e.g., Carnap, Chalmers, Clark, Sutton, Novaes, etc.), we move from artefacts to AI and Predictive Processing so as to link theories related to technicity with philosophy of mind. Simultaneously drawing forth Robert Brandom’s conceptualization of the roles that commitments play in retrospectively reconstructing the social experiences that lead to our endorsement(s) of norms, we compliment this account with Reza Negarestani’s deprivatized account of intelligence while analyzing the equipollent role between language and media (both digital and analog)

    Explaining the Qualitative Dimension of Consciousness: Prescission Instead of Reification: Dialogue

    Get PDF
    This paper suggests that it is largely a want of notional distinctions which fosters the “explanatory gap” that has beset the study of consciousness since T. Nagel’s revival of the topic. Modifying Ned Block’s controversial claim that we should countenance a “phenomenal-consciousness” which exists in its own right, we argue that there is a way to recuperate the intuitions he appeals to without engaging in an onerous reification of the facet in question. By renewing with the full type/token/tone trichotomy developed by C. S. Peirce, we think the distinctness Block calls attention to can be seen as stemming not from any separate module lurking within the mind, but rather from our ability to prescind qualities from occurrences

    Non classical concept representation and reasoning in formal ontologies

    Get PDF
    Formal ontologies are nowadays widely considered a standard tool for knowledge representation and reasoning in the Semantic Web. In this context, they are expected to play an important role in helping automated processes to access information. Namely: they are expected to provide a formal structure able to explicate the relationships between different concepts/terms, thus allowing intelligent agents to interpret, correctly, the semantics of the web resources improving the performances of the search technologies. Here we take into account a problem regarding Knowledge Representation in general, and ontology based representations in particular; namely: the fact that knowledge modeling seems to be constrained between conflicting requirements, such as compositionality, on the one hand and the need to represent prototypical information on the other. In particular, most common sense concepts seem not to be captured by the stringent semantics expressed by such formalisms as, for example, Description Logics (which are the formalisms on which the ontology languages have been built). The aim of this work is to analyse this problem, suggesting a possible solution suitable for formal ontologies and semantic web representations. The questions guiding this research, in fact, have been: is it possible to provide a formal representational framework which, for the same concept, combines both the classical modelling view (accounting for compositional information) and defeasible, prototypical knowledge ? Is it possible to propose a modelling architecture able to provide different type of reasoning (e.g. classical deductive reasoning for the compositional component and a non monotonic reasoning for the prototypical one)? We suggest a possible answer to these questions proposing a modelling framework able to represent, within the semantic web languages, a multilevel representation of conceptual information, integrating both classical and non classical (typicality based) information. Within this framework we hypothesise, at least in principle, the coexistence of multiple reasoning processes involving the different levels of representation

    Complexity over Uncertainty in Generalized Representational\ud Information Theory (GRIT): A Structure-Sensitive General\ud Theory of Information

    Get PDF
    What is information? Although researchers have used the construct of information liberally to refer to pertinent forms of domain-specific knowledge, relatively few have attempted to generalize and standardize the construct. Shannon and Weaver(1949)offered the best known attempt at a quantitative generalization in terms of the number of discriminable symbols required to communicate the state of an uncertain event. This idea, although useful, does not capture the role that structural context and complexity play in the process of understanding an event as being informative. In what follows, we discuss the limitations and futility of any generalization (and particularly, Shannon’s) that is not based on the way that agents extract patterns from their environment. More specifically, we shall argue that agent concept acquisition, and not the communication of\ud states of uncertainty, lie at the heart of generalized information, and that the best way of characterizing information is via the relative gain or loss in concept complexity that is experienced when a set of known entities (regardless of their nature or domain of origin) changes. We show that Representational Information Theory perfectly captures this crucial aspect of information and conclude with the first generalization of Representational Information Theory (RIT) to continuous domains
    corecore