438 research outputs found
The cat's cradle network
In this paper we will argue that the representation of context in knowledge management is appropriately served by the representation of the knowledge networks in an historicised form. Characterising context as essentially extra to any particular knowledge representation, we argue that another dimension to these be modelled, rather than simply elaborating a form in its own terms. We present the formalism of the cat's cradle network, and show how it can be represented by an extension of the Pathfinder associative network that includes this temporal dimension, and allows evolutions of understandings to be traced. Grounding its semantics in communities of practice ensures utility and cohesiveness, which is lost when mere externalities of a representation are communicated in fully fledged forms. The scheme is general and subsumes other formalisms for knowledge representation. The cat's cradle network enables us to model such community-based social constructs as pattern languages, shared memory and patterns of trust and reliance, by placing their establishment in a structure that shows their essential temporality
Just below the surface: developing knowledge management systems using the paradigm of the noetic prism
In this paper we examine how the principles embodied in the paradigm of the noetic prism can illuminate the construction of knowledge management systems. We draw on the formalism of the prism to examine three successful tools: frames, spreadsheets and databases, and show how their power and also their shortcomings arise from their domain representation, and how any organisational system based on integration of these tools and conversion between them is inevitably lossy. We suggest how a late-binding, hybrid knowledge based management system (KBMS) could be designed that draws on the lessons learnt from these tools, by maintaining noetica at an atomic level and storing the combinatory processes necessary to create higher level structure as the need arises. We outline the âjust-below-the-surfaceâ systems design, and describe its implementation in an enterprise-wide knowledge-based system that has all of the conventional office automation features
The Noetic Prism
Definitions of âknowledgeâ and its relationships with âdataâ and âinformationâ are varied, inconsistent and often contradictory. In particular the traditional hierarchy of data-information-knowledge and its various revisions do not stand up to close scrutiny. We suggest that the problem lies in a flawed analysis that sees data, information and knowledge as separable concepts that are transformed into one another through processing. We propose instead that we can describe collectively all of the materials of computation as ânoeticaâ, and that the terms data, information and knowledge can be reconceptualised as late-binding, purpose-determined aspects of the same body of material. Changes in complexity of noetica occur due to value-adding through the imposition of three different principles: increase in aggregation (granularity), increase in set relatedness (shape), and increase in contextualisation through the formation of networks (scope). We present a new model in which granularity, shape and scope are seen as the three vertices of a triangular prism, and show that all value-adding through computation can be seen as movement within the prism space. We show how the conceptual framework of the noetic prism provides a new and comprehensive analysis of the foundations of computing and information systems, and how it can provide a fresh analysis of many of the common problems in the management of intellectual resources
A knowledge development lifecycle for reflective practice
Reflective practice is valuable because of its potential for continuous improvement through feedback and learning. Conventional models of knowledge practice however do not explicitly include reflection as part of the practice, nor locate it in a developmental cycle. They focus on modelling in a knowledge plane which itself is contextualised by active knowing processes, and ignore the influence of power in their activity models. Further, many models focus on either an artefact or a process view, resulting from a conceptual disconnect between knowledge and knowing, and failure to relate passive to active views. Using the idea of higher order loops that govern knowledge development processes, in this paper we propose a conceptualisation of a reflective Knowledge Development Life Cycle (KDLC). This explicitly includes the investigator and the organisation itself as dynamic components of a systemic process and is suited to either a constructivist or realist epistemological stance. We describe the stages required in the KDLC and discuss their significance. Finally we show how incorporation of reflection into process enables dynamic interplay between the knowing and the knowledge in the organisation
User developed applications and information systems success: A test of DeLone and McLean's model
DeLone and McLeanâs (1992) model of information systems success has received much attention amongst researchers. This study provides the first empirical test of an adaptation of DeLone and McLeanâs model in the user-developed application domain. The model tested was only partially supported by the data. Of the nine hypothesized relationships tested, four were found to be significant and the remainder not significant. The model provided strong support for the relationships between perceived system quality and user satisfaction, perceived information quality and user satisfaction, user satisfaction and intended use, and user satisfaction and perceived individual impact. This study indicates that user perceptions of information systems success play a significant role in the user-developed application domain. There was, however, no relationship between user developersâ perceptions of system quality and independent expertsâ evaluations, and user ratings of individual impact were not associated with organizational impact measured as company performance in a business simulation. Further research is required to understand the relationship between user perceptions of IS success and objective measures of success, and to provide a model of IS success appropriate to end user development
Developing a self-paced interactive package for introductory computer programming
Teaching introductory programming can be a challenging task. Students can become too concerned with learning syntax at the expense of more general conceptual understanding. Distance education students in particular often have problems as the difficulty of the course content is compounded by the problems of isolation from other students and their tutor. Although instructional strategies for emphasising conceptual knowledge are well known for face-to-face instruction, there has been little attempt to apply them to the external situation. This paper describes the development of a supplementary package for external students designed to address some of these problems
User satisfaction as a measure of success in end user application development: An empirical investigation
End user application development to support organisational tasks is a phenomenon that is becoming increasingly important to organisations, yet there is little empirical research evaluating its effectiveness. Measures of user satisfaction, which are commonly used to evaluate organisational systems, may be inappropriate when end users assess systems they have developed themselves. This study compared the satisfaction ratings of applications which were evaluated by their end user developers with ratings of the same application by other end users. End users were significantly more satisfied with applications they had developed themselves. Possible implications of this finding are discussed
Search for direct production of charginos and neutralinos in events with three leptons and missing transverse momentum in âs = 7 TeV pp collisions with the ATLAS detector
A search for the direct production of charginos and neutralinos in final states with three electrons or muons and missing transverse momentum is presented. The analysis is based on 4.7 fbâ1 of protonâproton collision data delivered by the Large Hadron Collider and recorded with the ATLAS detector. Observations are consistent with Standard Model expectations in three signal regions that are either depleted or enriched in Z-boson decays. Upper limits at 95% confidence level are set in R-parity conserving phenomenological minimal supersymmetric models and in simplified models, significantly extending previous results
Jet size dependence of single jet suppression in lead-lead collisions at sqrt(s(NN)) = 2.76 TeV with the ATLAS detector at the LHC
Measurements of inclusive jet suppression in heavy ion collisions at the LHC
provide direct sensitivity to the physics of jet quenching. In a sample of
lead-lead collisions at sqrt(s) = 2.76 TeV corresponding to an integrated
luminosity of approximately 7 inverse microbarns, ATLAS has measured jets with
a calorimeter over the pseudorapidity interval |eta| < 2.1 and over the
transverse momentum range 38 < pT < 210 GeV. Jets were reconstructed using the
anti-kt algorithm with values for the distance parameter that determines the
nominal jet radius of R = 0.2, 0.3, 0.4 and 0.5. The centrality dependence of
the jet yield is characterized by the jet "central-to-peripheral ratio," Rcp.
Jet production is found to be suppressed by approximately a factor of two in
the 10% most central collisions relative to peripheral collisions. Rcp varies
smoothly with centrality as characterized by the number of participating
nucleons. The observed suppression is only weakly dependent on jet radius and
transverse momentum. These results provide the first direct measurement of
inclusive jet suppression in heavy ion collisions and complement previous
measurements of dijet transverse energy imbalance at the LHC.Comment: 15 pages plus author list (30 pages total), 8 figures, 2 tables,
submitted to Physics Letters B. All figures including auxiliary figures are
available at
http://atlas.web.cern.ch/Atlas/GROUPS/PHYSICS/PAPERS/HION-2011-02
Measurement of the polarisation of W bosons produced with large transverse momentum in pp collisions at sqrt(s) = 7 TeV with the ATLAS experiment
This paper describes an analysis of the angular distribution of W->enu and
W->munu decays, using data from pp collisions at sqrt(s) = 7 TeV recorded with
the ATLAS detector at the LHC in 2010, corresponding to an integrated
luminosity of about 35 pb^-1. Using the decay lepton transverse momentum and
the missing transverse energy, the W decay angular distribution projected onto
the transverse plane is obtained and analysed in terms of helicity fractions
f0, fL and fR over two ranges of W transverse momentum (ptw): 35 < ptw < 50 GeV
and ptw > 50 GeV. Good agreement is found with theoretical predictions. For ptw
> 50 GeV, the values of f0 and fL-fR, averaged over charge and lepton flavour,
are measured to be : f0 = 0.127 +/- 0.030 +/- 0.108 and fL-fR = 0.252 +/- 0.017
+/- 0.030, where the first uncertainties are statistical, and the second
include all systematic effects.Comment: 19 pages plus author list (34 pages total), 9 figures, 11 tables,
revised author list, matches European Journal of Physics C versio
- âŠ