2,116 research outputs found

    Learning and Games

    Get PDF
    Part of the Volume on the Ecology of Games: Connecting Youth, Games, and Learning In this chapter, I argue that good video games recruit good learning and that a game's design is inherently connected to designing good learning for players. I start with a perspective on learning now common in the Learning Sciences that argues that people primarily think and learn through experiences they have had, not through abstract calculations and generalizations. People store these experiences in memory -- and human long-term memory is now viewed as nearly limitless -- and use them to run simulations in their minds to prepare for problem solving in new situations. These simulations help them to form hypotheses about how to proceed in the new situation based on past experiences. The chapter also discusses the conditions experience must meet if it is to be optimal for learning and shows how good video games can deliver such optimal learning experiences. Some of the issues covered include: identity and learning; models and model-based thinking; the control of avatars and "empathy for a complex system"; distributed intelligence and cross-functional teams for learning; motivation, and ownership; emotion in learning; and situated meaning, that is, the ways in which games represent verbal meaning through images, actions, and dialogue, not just other words and definitions

    Integration of Action and Language Knowledge: A Roadmap for Developmental Robotics

    Get PDF
    “This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder." “Copyright IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.”This position paper proposes that the study of embodied cognitive agents, such as humanoid robots, can advance our understanding of the cognitive development of complex sensorimotor, linguistic, and social learning skills. This in turn will benefit the design of cognitive robots capable of learning to handle and manipulate objects and tools autonomously, to cooperate and communicate with other robots and humans, and to adapt their abilities to changing internal, environmental, and social conditions. Four key areas of research challenges are discussed, specifically for the issues related to the understanding of: 1) how agents learn and represent compositional actions; 2) how agents learn and represent compositional lexica; 3) the dynamics of social interaction and learning; and 4) how compositional action and language representations are integrated to bootstrap the cognitive system. The review of specific issues and progress in these areas is then translated into a practical roadmap based on a series of milestones. These milestones provide a possible set of cognitive robotics goals and test scenarios, thus acting as a research roadmap for future work on cognitive developmental robotics.Peer reviewe

    m-Reading: Fiction reading from mobile phones

    Get PDF
    Mobile phones are reportedly the most rapidly expanding e-reading device worldwide. However, the embodied, cognitive and affective implications of smartphone-supported fiction reading for leisure (m-reading) have yet to be investigated empirically. Revisiting the theoretical work of digitization scholar Anne Mangen, we argue that the digital reading experience is not only contingent on patterns of embodied reader–device interaction (Mangen, 2008 and later) but also embedded in the immediate environment and broader situational context. We call this the situation constraint. Its application to Mangen’s general framework enables us to identify four novel research areas, wherein m-reading should be investigated with regard to its unique affordances. The areas are reader–device affectivity, situated embodiment, attention training and long-term immersion

    Embodied learning environments for graphing motion : a systematic literature review

    Get PDF
    Embodied learning environments have a substantial share in teaching interventions and research for enhancing learning in science, technology, engineering, and mathematics (STEM) education. In these learning environments, students’ bodily experiences are an essential part of the learning activities and hence, of the learning. In this systematic review, we focused on embodied learning environments supporting students’ understanding of graphing change in the context of modeling motion. Our goal was to deepen the theoretical understanding of what aspects of these embodied learning environments are important for teaching and learning. We specified four embodied configurations by juxtaposing embodied learning environments on the degree of bodily involvement (own and others/objects’ motion) and immediacy (immediate and non-immediate) resulting in four classes of embodied learning environments. Our review included 44 articles (comprising 62 learning environments) and uncovered eight mediating factors, as described by the authors of the reviewed articles: real-world context, multimodality, linking motion to graph, multiple representations, semiotics, student control, attention capturing, and cognitive conflict. Different combinations of mediating factors were identified in each class of embodied learning environments. Additionally, we found that learning environments making use of students’ own motion immediately linked to its representation were most effective in terms of learning outcomes. Implications of this review for future research and the design of embodied learning environments are discussed.publishedVersionPaid Open Acces

    Merging multi-modal information and cross-modal learning in artificial cognitive systems

    Get PDF
    Cross-modal binding is the ability to merge two or more modal representations of the same entity into a single shared representation. This ability is one of the fundamental properties of any cognitive system operating in a complex environment. In order to adapt successfully to changes in a dynamic environment the binding mechanism has to be supplemented with cross-modal learning. But perhaps the most difficult task is the integration of both mechanisms into a cognitive system. Their role in such a system is two-fold: to bridge the semantic gap between modalities, and to mediate between the lower-level mechanisms for processing the sensory data, and the higher-level cognitive processes, such as motivation and planning. In this master thesis, we present an approach to probabilistic merging of multi-modal information in cognitive systems. By this approach, we formulate a model of binding and cross-modal learning in Markov logic networks, and describe the principles of its integration into a cognitive architecture. We implement a prototype of the model and evaluate it with off-line experiments that simulate a cognitive architecture with three modalities. Based on our approach, we design, implement and integrate the belief layer -- a subsystem that bridges the semantic gap in a prototype cognitive system named George. George is an intelligent robot that is able to detect and recognise objects in its surroundings, and learn about their properties in a situated dialogue with a human tutor. Its main purpose is to validate various paradigms of interactive learning. To this end, we have developed and performed on-line experiments that evaluate the mechanisms of robot's behaviour. With these experiments, we were also able to test and evaluate our approach to merging multi-modal information as part of a functional cognitive system
    • …
    corecore