56,085 research outputs found
Architecture and Design of Medical Processor Units for Medical Networks
This paper introduces analogical and deductive methodologies for the design
medical processor units (MPUs). From the study of evolution of numerous earlier
processors, we derive the basis for the architecture of MPUs. These specialized
processors perform unique medical functions encoded as medical operational
codes (mopcs). From a pragmatic perspective, MPUs function very close to CPUs.
Both processors have unique operation codes that command the hardware to
perform a distinct chain of subprocesses upon operands and generate a specific
result unique to the opcode and the operand(s). In medical environments, MPU
decodes the mopcs and executes a series of medical sub-processes and sends out
secondary commands to the medical machine. Whereas operands in a typical
computer system are numerical and logical entities, the operands in medical
machine are objects such as such as patients, blood samples, tissues, operating
rooms, medical staff, medical bills, patient payments, etc. We follow the
functional overlap between the two processes and evolve the design of medical
computer systems and networks.Comment: 17 page
An agent-based architecture for managing the provision of community care - the INCA (Intelligent Community Alarm) experience
Community Care is an area that requires extensive cooperation
between independent agencies, each of which needs to meet its own objectives and targets. None are engaged solely in the delivery of community care, and need to integrate the service with their other responsibilities in a coherent and efficient manner. Agent technology provides the means by which effective cooperation can take place without compromising the essential security of both the client and the
agencies involved as the appropriate set of responses can be generated through negotiation between the parties without the need for access to the main information repositories that would be necessary with conventional collaboration models. The autonomous nature of agents also means that a variety of agents can cooperate
together with various local capabilities, so long as they conform to the relevant messaging requirements. This allows a variety of agents, with capabilities tailored to the carers to which they are attached to be developed so that cost-effective solutions can be provided.
</p
Recommended from our members
Excellence in teaching and learning: A review of the literature for the Higher Education Academy
The Higher Education Academy commissioned a review of the literature on excellence in learning and teaching in higher education to enhance the higher education sectorâs understanding of the varied conceptualisations and usages of the term âexcellenceâ and consider the implications for future policy and practice in relation to promoting and developing excellence.
The literature searched included published research in the form of journal articles; books; reports from UK policy bodies and other agencies; as well as âgreyâ literature. It covered conceptual studies, academic critiques and research studies on learning and teaching, as well as policy documents.
Within a diverse and expanding system of higher education, such as in the UK, discourse on teaching and student learning highlights the tensions between different notions of excellence â for example, excellence as a positional good for students, as an aspirational target for continuous quality enhancement, as a form of reputational advantage for higher education institutions or as a means for achieving governmental economic and social goals.
The review addressed questions of conceptualisations and usage at different (but interlinked) levels: system-wide; institutional; departmental; individual, and from two different perspectives, teaching and student learning
Expert Elicitation for Reliable System Design
This paper reviews the role of expert judgement to support reliability
assessments within the systems engineering design process. Generic design
processes are described to give the context and a discussion is given about the
nature of the reliability assessments required in the different systems
engineering phases. It is argued that, as far as meeting reliability
requirements is concerned, the whole design process is more akin to a
statistical control process than to a straightforward statistical problem of
assessing an unknown distribution. This leads to features of the expert
judgement problem in the design context which are substantially different from
those seen, for example, in risk assessment. In particular, the role of experts
in problem structuring and in developing failure mitigation options is much
more prominent, and there is a need to take into account the reliability
potential for future mitigation measures downstream in the system life cycle.
An overview is given of the stakeholders typically involved in large scale
systems engineering design projects, and this is used to argue the need for
methods that expose potential judgemental biases in order to generate analyses
that can be said to provide rational consensus about uncertainties. Finally, a
number of key points are developed with the aim of moving toward a framework
that provides a holistic method for tracking reliability assessment through the
design process.Comment: This paper commented in: [arXiv:0708.0285], [arXiv:0708.0287],
[arXiv:0708.0288]. Rejoinder in [arXiv:0708.0293]. Published at
http://dx.doi.org/10.1214/088342306000000510 in the Statistical Science
(http://www.imstat.org/sts/) by the Institute of Mathematical Statistics
(http://www.imstat.org
Changing power relations in work based learning: Collaborative and contested relations between tutors, learners and employers
This is the author's pdf pre-print of a book chapter due to be published in 2011.This book chapter discusses some of the implications for the role of university tutors and the centrality of educational objectives in circumstances where there is a 'cultural shift' towards meeting the needs of learners and employers. The work based and integrative studies (WBIS) programme at the University of Chester is used as a case study to examine the changing power relations between university tutors, learners, employers and the university, compared to relations on traditional programmes
Artificial intelligence approaches to software engineering
Artificial intelligence approaches to software engineering are examined. The software development life cycle is a sequence of not so well-defined phases. Improved techniques for developing systems have been formulated over the past 15 years, but pressure continues to attempt to reduce current costs. Software development technology seems to be standing still. The primary objective of the knowledge-based approach to software development presented in this paper is to avoid problem areas that lead to schedule slippages, cost overruns, or software products that fall short of their desired goals. Identifying and resolving software problems early, often in the phase in which they first occur, has been shown to contribute significantly to reducing risks in software development. Software development is not a mechanical process but a basic human activity. It requires clear thinking, work, and rework to be successful. The artificial intelligence approaches to software engineering presented support the software development life cycle through the use of software development techniques and methodologies in terms of changing current practices and methods. These should be replaced by better techniques that that improve the process of of software development and the quality of the resulting products. The software development process can be structured into well-defined steps, of which the interfaces are standardized, supported and checked by automated procedures that provide error detection, production of the documentation and ultimately support the actual design of complex programs
Software reliability and dependability: a roadmap
Shifting the focus from software reliability to user-centred measures of dependability in complete software-based systems. Influencing design practice to facilitate dependability assessment. Propagating awareness of dependability issues and the use of existing, useful methods. Injecting some rigour in the use of process-related evidence for dependability assessment. Better understanding issues of diversity and variation as drivers of dependability. Bev Littlewood is founder-Director of the Centre for Software Reliability, and Professor of Software Engineering at City University, London. Prof Littlewood has worked for many years on problems associated with the modelling and evaluation of the dependability of software-based systems; he has published many papers in international journals and conference proceedings and has edited several books. Much of this work has been carried out in collaborative projects, including the successful EC-funded projects SHIP, PDCS, PDCS2, DeVa. He has been employed as a consultant t
Overview of methodologies for building ontologies
A few research groups are now proposing a series of steps and methodologies for developing ontologies. However, mainly due to the fact that Ontological Engineering is still a relatively immature discipline, each work group employs its own methodology. Our goal is to present the most representative methodologies used in ontology development and to perform an analysis of such methodologies against the same framework of reference. So, the goal of this paper is not to provide new insights about methodologies, but to put it all in one place and help people to select which methodology to use
- âŠ