287,383 research outputs found

    Weblogs in Higher Education - Why Do Students (Not) Blog?

    Get PDF
    Positive impacts on learning through blogging, such as active knowledge construction and reflective writing, have been reported. However, not many students use weblogs in informal contexts, even when appropriate facilities are offered by their universities. While motivations for blogging have been subject to empirical studies, little research has addressed the issue of why students choose not to blog. This paper presents an empirical study undertaken to gain insights into the decision making process of students when deciding whether to keep a blog or not. A better understanding of students' motivations for (not) blogging may help decision makers at universities in the process of selecting, introducing, and maintaining similar services. As informal learning gains increased recognition, results of this study can help to advance appropriate designs of informal learning contexts in Higher Education. The method of ethnographic decision tree modelling was applied in an empirical study conducted at the Vienna University of Technology, Austria. Since 2004, the university has been offering free weblog accounts for all students and staff members upon entering school, not bound to any course or exam. Qualitative, open interviews were held with 3 active bloggers, 3 former bloggers, and 3 non‑ bloggers to elicit their decision criteria. Decision tree models were developed out of the interviews. It turned out that the modelling worked best when splitting the decision process into two parts: one model representing decisions on whether to start a weblog at all, and a second model representing criteria on whether to continue with a weblog once it was set up. The models were tested for their validity through questionnaires developed out of the decision tree models. 30 questionnaires have been distributed to bloggers, former bloggers and non‑ bloggers. Results show that the main reasons for students not to keep a weblog include a preference for direct (online) communication, and concerns about the loss of privacy through blogging. Furthermore, the results indicate that intrinsic motivation factors keep students blogging, whereas stopping a weblog is mostly attributable to external factors

    Digital or Diligent? Web 2.0's challenge to formal schooling

    Get PDF
    This paper explores the tensions that arise for young people as both 'digital kids' and 'diligent students'. It does so by drawing on a study conducted in an elite private school, where the tensions between 'going digital' and 'being diligent' are exacerbated by the high value the school places on academic achievement, and on learning through digital innovation. At the school under study, high levels of intellectual and technological resourcing bring with them an equally high level of expectation to excel in traditional academic tasks and high-stakes assessment. The students, under constant pressure to perform well in standardised tests, need to make decisions about the extent to which they take up school-sanctioned digitally enhanced learning opportunities that do not explicitly address academic performance. The paper examines this conundrum by investigating student preparedness to engage with a new learning innovation – a student-led media centre – in the context of the traditional pedagogical culture that is relatively untouched by such digital innovation. The paper presents an analysis of findings from a survey of 481 students in the school. The survey results were subjected to quantitative regression tree modelling to flesh out how different student learning dispositions, social and technological factors influence the extent to which students engage with a specific digital learning opportunity in the form of the Web 2.0 Student Media Centre (SMC) designed to engage the senior school community in flexible digital-networked learning. What emerges from the study is that peer support, perceived ease of use and usefulness, learning goals and cognitive playfulness are significant predictors of the choices that students make to negotiate the fundamental tensions of being digital and/or diligent. In scrutinising the tensions around a digital or a diligent student identity in this way, the paper contributes new empirical evidence to understanding the problematic relationship between student-led learning using new digital media tools and formal schooling

    Understanding urban gentrification through machine learning

    Get PDF
    Recent developments in the field of machine learning offer new ways of modelling complex socio-spatial processes, allowing us to make predictions about how and where they might manifest in the future. Drawing on earlier empirical and theoretical attempts to understand gentrification and urban change, this paper shows it is possible to analyse existing patterns and processes of neighbourhood change to identify areas likely to experience change in the future. This is evidenced through an analysis of socio-economic transition in London neighbourhoods (based on 2001 and 2011 Census variables) which is used to predict those areas most likely to demonstrate ‘uplift’ or ‘decline’ by 2021. The paper concludes with a discussion of the implications of such modelling for the understanding of gentrification processes, noting that if qualitative work on gentrification and neighbourhood change is to offer more than a rigorous post-mortem then intensive, qualitative case studies must be confronted with – and complemented by – predictions stemming from other, more extensive approaches. As a demonstration of the capabilities of machine learning, this paper underlines the continuing value of quantitative approaches in understanding complex urban processes such as gentrification

    Empirical modelling principles to support learning in a cultural context

    Get PDF
    Much research on pedagogy stresses the need for a broad perspective on learning. Such a perspective might take account (for instance) of the experience that informs knowledge and understanding [Tur91], the situation in which the learning activity takes place [Lav88], and the influence of multiple intelligences [Gar83]. Educational technology appears to hold great promise in this connection. Computer-related technologies such as new media, the internet, virtual reality and brain-mediated communication afford access to a range of learning resources that grows ever wider in its scope and supports ever more sophisticated interactions. Whether educational technology is fulfilling its potential in broadening the horizons for learning activity is more controversial. Though some see the successful development of radically new educational resources as merely a matter of time, investment and engineering, there are also many critics of the trends in computer-based learning who see little evidence of the greater degree of human engagement to which new technologies aspire [Tal95]. This paper reviews the potential application to educational technology of principles and tools for computer-based modelling that have been developed under the auspices of the Empirical Modelling (EM) project at Warwick [EMweb]. This theme was first addressed at length in a previous paper [Bey97], and is here revisited in the light of new practical developments in EM both in respect of tools and of model-building that has been targetted at education at various levels. Our central thesis is that the problems of educational technology stem from the limitations of current conceptual frameworks and tool support for the essential cognitive model building activity, and that tackling these problems requires a radical shift in philosophical perspective on the nature and role of empirical knowledge that has significant practical implications. The paper is in two main sections. The first discusses the limitations of the classical computer science perspective where educational technology to support situated learning is concerned, and relates the learning activities that are most closely associated with a cultural context to the empiricist perspective on learning introduced in [Bey97]. The second outlines the principles of EM and describes and illustrates features of its practical application that are particularly well-suited to learning in a cultural setting

    Quality-aware model-driven service engineering

    Get PDF
    Service engineering and service-oriented architecture as an integration and platform technology is a recent approach to software systems integration. Quality aspects ranging from interoperability to maintainability to performance are of central importance for the integration of heterogeneous, distributed service-based systems. Architecture models can substantially influence quality attributes of the implemented software systems. Besides the benefits of explicit architectures on maintainability and reuse, architectural constraints such as styles, reference architectures and architectural patterns can influence observable software properties such as performance. Empirical performance evaluation is a process of measuring and evaluating the performance of implemented software. We present an approach for addressing the quality of services and service-based systems at the model-level in the context of model-driven service engineering. The focus on architecture-level models is a consequence of the black-box character of services

    Cause-Effect Inference in Location-Scale Noise Models: Maximum Likelihood vs. Independence Testing

    Full text link
    A fundamental problem of causal discovery is cause-effect inference, learning the correct causal direction between two random variables. Significant progress has been made through modelling the effect as a function of its cause and a noise term, which allows us to leverage assumptions about the generating function class. The recently introduced heteroscedastic location-scale noise functional models (LSNMs) combine expressive power with identifiability guarantees. LSNM model selection based on maximizing likelihood achieves state-of-the-art accuracy, when the noise distributions are correctly specified. However, through an extensive empirical evaluation, we demonstrate that the accuracy deteriorates sharply when the form of the noise distribution is misspecified by the user. Our analysis shows that the failure occurs mainly when the conditional variance in the anti-causal direction is smaller than that in the causal direction. As an alternative, we find that causal model selection through residual independence testing is much more robust to noise misspecification and misleading conditional variance.Comment: preprin

    A distributed framework for semi-automatically developing architectures of brain and mind

    Get PDF
    Developing comprehensive theories of low-level neuronal brain processes and high-level cognitive behaviours, as well as integrating them, is an ambitious challenge that requires new conceptual, computational, and empirical tools. Given the complexities of these theories, they will almost certainly be expressed as computational systems. Here, we propose to use recent developments in grid technology to develop a system of evolutionary scientific discovery, which will (a) enable empirical researchers to make their data widely available for use in developing and testing theories, and (b) enable theorists to semi-automatically develop computational theories. We illustrate these ideas with a case study taken from the domain of categorisation

    Understanding from Machine Learning Models

    Get PDF
    Simple idealized models seem to provide more understanding than opaque, complex, and hyper-realistic models. However, an increasing number of scientists are going in the opposite direction by utilizing opaque machine learning models to make predictions and draw inferences, suggesting that scientists are opting for models that have less potential for understanding. Are scientists trading understanding for some other epistemic or pragmatic good when they choose a machine learning model? Or are the assumptions behind why minimal models provide understanding misguided? In this paper, using the case of deep neural networks, I argue that it is not the complexity or black box nature of a model that limits how much understanding the model provides. Instead, it is a lack of scientific and empirical evidence supporting the link that connects a model to the target phenomenon that primarily prohibits understanding
    corecore