1,036 research outputs found
BAYESIAN BELIEF NETWORK AND FUZZY LOGIC ADAPTIVE MODELING OF DYNAMIC SYSTEM: EXTENSION AND COMPARISON
The purpose of this thesis is to develop, expand, compare and contrast two methodologies, namely BBN and FLM, which are used in the modeling of the dynamics of physical system behavior and are instrumental in a better understanding on the POF. The paper begins with an introduction of the proposed approaches in the modeling of complex physical systems, followed by a quick literature review of FLM and BBN. This thesis uses an existing pump system [3] as a case study, where the resulting NPSHA data obtained from the applications of BBN and FLM are compared with the outputs derived from the implementation of a Mathematical Model. Based on these findings, discussions and analyses are made, including the identification of the respective strengths and weaknesses posed by the two methodologies. Last but not least, further extensions and improvements towards this research are discussed at the end of this paper
Stepping Beyond the Newtonian Paradigm in Biology. Towards an Integrable Model of Life: Accelerating Discovery in the Biological Foundations of Science
The INBIOSA project brings together a group of experts across many disciplines
who believe that science requires a revolutionary transformative
step in order to address many of the vexing challenges presented by the
world. It is INBIOSAâs purpose to enable the focused collaboration of an
interdisciplinary community of original thinkers.
This paper sets out the case for support for this effort. The focus of the
transformative research program proposal is biology-centric. We admit
that biology to date has been more fact-oriented and less theoretical than
physics. However, the key leverageable idea is that careful extension of the
science of living systems can be more effectively applied to some of our
most vexing modern problems than the prevailing scheme, derived from
abstractions in physics. While these have some universal application and
demonstrate computational advantages, they are not theoretically mandated
for the living. A new set of mathematical abstractions derived from biology
can now be similarly extended. This is made possible by leveraging
new formal tools to understand abstraction and enable computability. [The
latter has a much expanded meaning in our context from the one known
and used in computer science and biology today, that is "by rote algorithmic
means", since it is not known if a living system is computable in this
sense (Mossio et al., 2009).] Two major challenges constitute the effort.
The first challenge is to design an original general system of abstractions
within the biological domain. The initial issue is descriptive leading to the
explanatory. There has not yet been a serious formal examination of the
abstractions of the biological domain. What is used today is an amalgam;
much is inherited from physics (via the bridging abstractions of chemistry)
and there are many new abstractions from advances in mathematics (incentivized
by the need for more capable computational analyses). Interspersed
are abstractions, concepts and underlying assumptions ânativeâ to biology
and distinct from the mechanical language of physics and computation as
we know them. A pressing agenda should be to single out the most concrete
and at the same time the most fundamental process-units in biology
and to recruit them into the descriptive domain. Therefore, the first challenge
is to build a coherent formal system of abstractions and operations
that is truly native to living systems.
Nothing will be thrown away, but many common methods will be philosophically
recast, just as in physics relativity subsumed and reinterpreted
Newtonian mechanics.
This step is required because we need a comprehensible, formal system to
apply in many domains. Emphasis should be placed on the distinction between
multi-perspective analysis and synthesis and on what could be the
basic terms or tools needed.
The second challenge is relatively simple: the actual application of this set
of biology-centric ways and means to cross-disciplinary problems. In its
early stages, this will seem to be a ânew scienceâ.
This White Paper sets out the case of continuing support of Information
and Communication Technology (ICT) for transformative research in biology
and information processing centered on paradigm changes in the epistemological,
ontological, mathematical and computational bases of the science
of living systems. Today, curiously, living systems cannot be said to
be anything more than dissipative structures organized internally by genetic
information. There is not anything substantially different from abiotic
systems other than the empirical nature of their robustness. We believe that
there are other new and unique properties and patterns comprehensible at
this bio-logical level. The report lays out a fundamental set of approaches
to articulate these properties and patterns, and is composed as follows.
Sections 1 through 4 (preamble, introduction, motivation and major biomathematical
problems) are incipient. Section 5 describes the issues affecting
Integral Biomathics and Section 6 -- the aspects of the Grand Challenge
we face with this project. Section 7 contemplates the effort to
formalize a General Theory of Living Systems (GTLS) from what we have
today. The goal is to have a formal system, equivalent to that which exists
in the physics community. Here we define how to perceive the role of time
in biology. Section 8 describes the initial efforts to apply this general theory
of living systems in many domains, with special emphasis on crossdisciplinary
problems and multiple domains spanning both âhardâ and
âsoftâ sciences. The expected result is a coherent collection of integrated
mathematical techniques. Section 9 discusses the first two test cases, project
proposals, of our approach. They are designed to demonstrate the ability
of our approach to address âwicked problemsâ which span across physics,
chemistry, biology, societies and societal dynamics. The solutions
require integrated measurable results at multiple levels known as âgrand
challengesâ to existing methods. Finally, Section 10 adheres to an appeal
for action, advocating the necessity for further long-term support of the
INBIOSA program.
The report is concluded with preliminary non-exclusive list of challenging
research themes to address, as well as required administrative actions. The
efforts described in the ten sections of this White Paper will proceed concurrently.
Collectively, they describe a program that can be managed and
measured as it progresses
Optimising outcomes for potentially resectable pancreatic cancer through personalised predictive medicine : the application of complexity theory to probabilistic statistical modeling
Survival outcomes for pancreatic cancer remain poor. Surgical resection with adjuvant therapy is the only potentially curative treatment, but for many people surgery is of limited benefit. Neoadjuvant therapy has emerged as an alternative treatment pathway however the evidence base surrounding the treatment of potentially resectable pancreatic cancer is highly heterogeneous and fraught with uncertainty and controversy.
This research seeks to engage with conjunctive theorising by avoiding simplification and abstraction to draw on different kinds of data from multiple sources to move research towards a theory that can build a rich picture of pancreatic cancer management pathways as a complex system. The overall aim is to move research towards personalised realistic medicine by using personalised predictive modeling to facilitate better decision making to achieve the optimisation of outcomes.
This research is theory driven and empirically focused from a complexity perspective. Combining operational and healthcare research methodology, and drawing on influences from complementary paradigms of critical realism and systems theory, then enhancing their impact by using Cilliersâ complexity theory âlean ontologyâ, an open-world ontology is held and both epistemic reality and judgmental relativity are accepted. The use of imperfect data within statistical simulation models is explored to attempt to expand our capabilities for handling the emergent and uncertainty and to find other ways of relating to complexity within the field of pancreatic cancer research.
Markov and discrete-event simulation modelling uncovered new insights and added a further dimension to the current debate by demonstrating that superior treatment pathway selection depended on individual patient and tumour factors. A Bayesian Belief Network was developed that modelled the dynamic nature of this complex system to make personalised prognostic predictions across competing treatments pathways throughout the patient journey to facilitate better shared clinical decision making with an accuracy exceeding existing predictive models.Survival outcomes for pancreatic cancer remain poor. Surgical resection with adjuvant therapy is the only potentially curative treatment, but for many people surgery is of limited benefit. Neoadjuvant therapy has emerged as an alternative treatment pathway however the evidence base surrounding the treatment of potentially resectable pancreatic cancer is highly heterogeneous and fraught with uncertainty and controversy.
This research seeks to engage with conjunctive theorising by avoiding simplification and abstraction to draw on different kinds of data from multiple sources to move research towards a theory that can build a rich picture of pancreatic cancer management pathways as a complex system. The overall aim is to move research towards personalised realistic medicine by using personalised predictive modeling to facilitate better decision making to achieve the optimisation of outcomes.
This research is theory driven and empirically focused from a complexity perspective. Combining operational and healthcare research methodology, and drawing on influences from complementary paradigms of critical realism and systems theory, then enhancing their impact by using Cilliersâ complexity theory âlean ontologyâ, an open-world ontology is held and both epistemic reality and judgmental relativity are accepted. The use of imperfect data within statistical simulation models is explored to attempt to expand our capabilities for handling the emergent and uncertainty and to find other ways of relating to complexity within the field of pancreatic cancer research.
Markov and discrete-event simulation modelling uncovered new insights and added a further dimension to the current debate by demonstrating that superior treatment pathway selection depended on individual patient and tumour factors. A Bayesian Belief Network was developed that modelled the dynamic nature of this complex system to make personalised prognostic predictions across competing treatments pathways throughout the patient journey to facilitate better shared clinical decision making with an accuracy exceeding existing predictive models
Digital Image Access & Retrieval
The 33th Annual Clinic on Library Applications of Data Processing, held at the University of Illinois at Urbana-Champaign in March of 1996, addressed the theme of "Digital Image Access & Retrieval." The papers from this conference cover a wide range of topics concerning digital imaging technology for visual resource collections. Papers covered three general areas: (1) systems, planning, and implementation; (2) automatic and semi-automatic indexing; and (3) preservation with the bulk of the conference focusing on indexing and retrieval.published or submitted for publicatio
Lean Thinking For Lead-Time Reduction And Efficient Knowledge Creation In Product Development
There are many distinct differences between manufacturing process and Product Development (PD) process, so lean tools have to be customized to deliver results in the later domain. The main focus of this dissertation is to extend them to manage and improve the PD process in order to develop the product faster while improving or at least maintaining the level of performance and quality. For aforesaid purpose, value stream mapping (VSM) method is used to explore the wastes, inefficiencies, non-valued added steps in a single, definable process out of complete PD process. Besides numerous intangible benefits, VSM framework will help the development team to reduce the lead-time by over 50%. Next, a set of ten lean tools and methods is proposed in order to support and improve efficiency of the knowledge creation (KC) process. The approach establishes a KC framework in PD environment, and systematically demonstrates how these lean tools and methods conceptually fit into and play a significant role in enhancing the performance of KC process. Following this, each of them is analysed and appropriately positioned in a SECI (socialization-externalization-combination-internalization) mode depending on the best fit. Quick and correct KC at the right time aids in further improving the development lead-time and product quality.
Such successful innovation is often associated with adoption and execution of all SECI modes within any PD phase. This dissertation attempts to argue with this general notion and to distinguish different PD phases\u27 affinity corresponding to distinct SECI mode. In this regard, an extended Fuzzy Analytic Hierarchy Process (EFAHP) approach to determine the ranking in which any PD phase is influenced from SECI modes is proposed. In the EFAHP approach, the complex problem of KC is first itemized into a simple hierarchical structure for pairwise comparisons. Next, a triangular fuzzy number concept is applied to capture the inherent vagueness in linguistic terms of a decision-maker. This dissertation recommends mapping the triangular fuzzy numbers (TFNs) with normal distributions about X-axis when the pessimistic value of one TFN is less than the optimistic value of other TFN (t23 †t11). This allows us to develop a mathematical formulation to estimate the degree of possibility of two criteria as opposed to zero resulted by the use of the current technique in the literature. In order to demonstrate the applicability and usefulness of the proposed EFAHP in ranking the SECI modes, an empirical study of development phase is considered. After stringent analysis, we found that the combination mode was the mode that highly influenced the development phase
North American Fuzzy Logic Processing Society (NAFIPS 1992), volume 2
This document contains papers presented at the NAFIPS '92 North American Fuzzy Information Processing Society Conference. More than 75 papers were presented at this Conference, which was sponsored by NAFIPS in cooperation with NASA, the Instituto Tecnologico de Morelia, the Indian Society for Fuzzy Mathematics and Information Processing (ISFUMIP), the Instituto Tecnologico de Estudios Superiores de Monterrey (ITESM), the International Fuzzy Systems Association (IFSA), the Japan Society for Fuzzy Theory and Systems, and the Microelectronics and Computer Technology Corporation (MCC). The fuzzy set theory has led to a large number of diverse applications. Recently, interesting applications have been developed which involve the integration of fuzzy systems with adaptive processes such a neural networks and genetic algorithms. NAFIPS '92 was directed toward the advancement, commercialization, and engineering development of these technologies
Mixing Methods: Practical Insights from the Humanities in the Digital Age
The digital transformation is accompanied by two simultaneous processes: digital humanities challenging the humanities, their theories, methodologies and disciplinary identities, and pushing computer science to get involved in new fields. But how can qualitative and quantitative methods be usefully combined in one research project? What are the theoretical and methodological principles across all disciplinary digital approaches? This volume focusses on driving innovation and conceptualising the humanities in the 21st century. Building on the results of 10 research projects, it serves as a useful tool for designing cutting-edge research that goes beyond conventional strategies
Computer Vision and Architectural History at Eye Level:Mixed Methods for Linking Research in the Humanities and in Information Technology
Information on the history of architecture is embedded in our daily surroundings, in vernacular and heritage buildings and in physical objects, photographs and plans. Historians study these tangible and intangible artefacts and the communities that built and used them. Thus valuableinsights are gained into the past and the present as they also provide a foundation for designing the future. Given that our understanding of the past is limited by the inadequate availability of data, the article demonstrates that advanced computer tools can help gain more and well-linked data from the past. Computer vision can make a decisive contribution to the identification of image content in historical photographs. This application is particularly interesting for architectural history, where visual sources play an essential role in understanding the built environment of the past, yet lack of reliable metadata often hinders the use of materials. The automated recognition contributes to making a variety of image sources usable forresearch.<br/
- âŠ