40 research outputs found
Proceedings of the 1st Doctoral Consortium at the European Conference on Artificial Intelligence (DC-ECAI 2020)
1st Doctoral Consortium at the European Conference on
Artificial Intelligence (DC-ECAI 2020), 29-30 August, 2020
Santiago de Compostela, SpainThe DC-ECAI 2020 provides a unique opportunity for PhD students, who are close to finishing their doctorate research, to interact with experienced researchers in the field. Senior members of the community are assigned as mentors for each group of students based on the student’s research or similarity of research interests. The DC-ECAI 2020, which is held virtually this year, allows students from all over the world to present their research and discuss their ongoing research and career plans with their mentor, to do networking with other participants, and to receive training and mentoring about career planning and career option
Biomedical applications of belief networks
Biomedicine is an area in which computers have long been expected to play a significant
role. Although many of the early claims have proved unrealistic, computers are gradually
becoming accepted in the biomedical, clinical and research environment. Within these
application areas, expert systems appear to have met with the most resistance, especially
when applied to image interpretation.In order to improve the acceptance of computerised decision support systems it is
necessary to provide the information needed to make rational judgements concerning
the inferences the system has made. This entails an explanation of what inferences
were made, how the inferences were made and how the results of the inference are to
be interpreted. Furthermore there must be a consistent approach to the combining of
information from low level computational processes through to high level expert analyses.nformation from low level computational processes through to high level expert analyses.
Until recently ad hoc formalisms were seen as the only tractable approach to reasoning
under uncertainty. A review of some of these formalisms suggests that they are less
than ideal for the purposes of decision making. Belief networks provide a tractable way
of utilising probability theory as an inference formalism by combining the theoretical
consistency of probability for inference and decision making, with the ability to use the
knowledge of domain experts.nowledge of domain experts.
The potential of belief networks in biomedical applications has already been recog¬
nised and there has been substantial research into the use of belief networks for medical
diagnosis and methods for handling large, interconnected networks. In this thesis the use
of belief networks is extended to include detailed image model matching to show how,
in principle, feature measurement can be undertaken in a fully probabilistic way. The
belief networks employed are usually cyclic and have strong influences between adjacent
nodes, so new techniques for probabilistic updating based on a model of the matching
process have been developed.An object-orientated inference shell called FLAPNet has been implemented and used
to apply the belief network formalism to two application domains. The first application is
model-based matching in fetal ultrasound images. The imaging modality and biological
variation in the subject make model matching a highly uncertain process. A dynamic,
deformable model, similar to active contour models, is used. A belief network combines
constraints derived from local evidence in the image, with global constraints derived from
trained models, to control the iterative refinement of an initial model cue.In the second application a belief network is used for the incremental aggregation of
evidence occurring during the classification of objects on a cervical smear slide as part of
an automated pre-screening system. A belief network provides both an explicit domain
model and a mechanism for the incremental aggregation of evidence, two attributes
important in pre-screening systems.Overall it is argued that belief networks combine the necessary quantitative features
required of a decision support system with desirable qualitative features that will lead
to improved acceptability of expert systems in the biomedical domain
Development and application of artificial intelligence strategies to solve infrared spectroscopic problems
The ever-increasing power of modern infrared instrumentation, coupled with the decreasing number of experienced spectroscopists has created an imbalance between information generation and interpretation capabilities. At the same time, digital computers are being developed which continue to grow in storage and processing capabilities, and shrink in cost. Clearly, the computer may serve as a valuable tool to aid the analytical chemist in interpreting spectroscopic information. This dissertation deals with the development of new approaches to exploiting computer technology to interpret infrared spectroscopic data.
A large existing expert system for functional group analysis, PAIRS, has been modified to transfer the maximum amount of information to the chemist. Two closely coupled knowledge based systems, IRBASE and MIXIR, have been created to identify major components of condensed phase mixtures. A second version of MIXIR has been developed to identify major components of vapor phase mixtures. Finally, a neural network approach to peak detection in analytical data has been developed
LOGICAL AND PSYCHOLOGICAL PARTITIONING OF MIND: DEPICTING THE SAME MAP?
The aim of this paper is to demonstrate that empirically delimited structures of mind are also differentiable by means of systematic logical analysis. In the sake of this aim, the paper first summarizes Demetriou's theory of cognitive organization and growth. This theory assumes that the mind is a multistructural entity that develops across three fronts: the processing system that constrains processing potentials, a set of specialized structural systems (SSSs) that guide processing within different reality and knowledge domains, and a hypecognitive system that monitors and controls the functioning of all other systems. In the second part the paper focuses on the SSSs, which are the target of our logical analysis, and it summarizes a series of empirical studies demonstrating their autonomous operation. The third part develops the logical proof showing that each SSS involves a kernel element that cannot be reduced to standard logic or to any other SSS. The implications of this analysis for the general theory of knowledge and cognitive development are discussed in the concluding part of the paper
Natively probabilistic computation
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Brain and Cognitive Sciences, 2009.Includes bibliographical references (leaves 129-135).I introduce a new set of natively probabilistic computing abstractions, including probabilistic generalizations of Boolean circuits, backtracking search and pure Lisp. I show how these tools let one compactly specify probabilistic generative models, generalize and parallelize widely used sampling algorithms like rejection sampling and Markov chain Monte Carlo, and solve difficult Bayesian inference problems. I first introduce Church, a probabilistic programming language for describing probabilistic generative processes that induce distributions, which generalizes Lisp, a language for describing deterministic procedures that induce functions. I highlight the ways randomness meshes with the reflectiveness of Lisp to support the representation of structured, uncertain knowledge, including nonparametric Bayesian models from the current literature, programs for decision making under uncertainty, and programs that learn very simple programs from data. I then introduce systematic stochastic search, a recursive algorithm for exact and approximate sampling that generalizes a popular form of backtracking search to the broader setting of stochastic simulation and recovers widely used particle filters as a special case. I use it to solve probabilistic reasoning problems from statistical physics, causal reasoning and stereo vision. Finally, I introduce stochastic digital circuits that model the probability algebra just as traditional Boolean circuits model the Boolean algebra.(cont.) I show how these circuits can be used to build massively parallel, fault-tolerant machines for sampling and allow one to efficiently run Markov chain Monte Carlo methods on models with hundreds of thousands of variables in real time. I emphasize the ways in which these ideas fit together into a coherent software and hardware stack for natively probabilistic computing, organized around distributions and samplers rather than deterministic functions. I argue that by building uncertainty and randomness into the foundations of our programming languages and computing machines, we may arrive at ones that are more powerful, flexible and efficient than deterministic designs, and are in better alignment with the needs of computational science, statistics and artificial intelligence.by Vikash Kumar Mansinghka.Ph.D
Resource Rationality
Theories of rational decision making often abstract away from computational and other resource limitations faced by real agents. An alternative approach known as resource rationality puts such matters front and center, grounding choice and decision in the rational use of finite resources. Anticipated by earlier work in economics and in computer science, this approach has recently seen rapid development and application in the cognitive sciences. Here, the theory of rationality plays a dual role, both as a framework for normative assessment and as a source of scientific hypotheses about how mental processes in fact work. The latter project, often called rational analysis, depends for its success on a fine-grained characterization of the computational problem facing a decision maker, which may in turn depend on realistic assumptions about what the relevant agent is like. As a consequence, resource rationality involves a delicate, but often fruitful interplay between the normative and the descriptive
Artificial general intelligence: Proceedings of the Second Conference on Artificial General Intelligence, AGI 2009, Arlington, Virginia, USA, March 6-9, 2009
Artificial General Intelligence (AGI) research focuses on the original and ultimate goal of AI – to create broad human-like and transhuman intelligence, by exploring all available paths, including theoretical and experimental computer science, cognitive science, neuroscience, and innovative interdisciplinary methodologies. Due to the difficulty of this task, for the last few decades the majority of AI researchers have focused on what has been called narrow AI – the production of AI systems displaying intelligence regarding specific, highly constrained tasks. In
recent years, however, more and more researchers have recognized the necessity – and feasibility – of returning to the original goals of the field. Increasingly, there is a call for a transition back to confronting the more difficult issues of human level intelligence and more broadly artificial general intelligence