18 research outputs found
Applications of interval analysis to selected topics in statistical computing
In interval analysis, an interval is treated not only as a set of numbers, but as a number in and of itself. The development of interval analysis is closely connected to the development of electronic digital computers. Conventional electronic computation is typically performed using a fixed-precision, floating-point processor. This approach is a finite approximation to calculations with real numbers of infinite precision. The finite approximation leads to errors of various types. While the fundamental operations of addition, subtraction, multiplication and division are typically accurate to one-half unit-last-place in floating-point computations, the effect of cumulative error in repeated calculations is usually unknown and too-frequently ignored. Using interval analysis, an interval is constructed which (after each computation) is guaranteed to contain the true value. By seeking ways to keep the interval narrow, it is possible to obtain results which are of guaranteed accuracy;This dissertation uses interval analysis in topics of statistical computing. Two major topics are addressed: bounding computational errors and global optimization;For bounding computational errors, series are used which yield a bound on the truncation error which results from a finite series approximation to an infinite series. By evaluating the series with intervals to bound rounding errors and by using the bound on the truncation error, an interval is obtained which is guaranteed to contain the true value. For some series, interval numerical quadrature rules are also employed. These ideas are applied to the computation of tail probabilities and critical points of several statistical distributions such as Bivariate Chi-Square and Bivariate F distributions;As regards to global optimization, the EM algorithm is one tool frequently used for optimization in Statistics and Probability; The EM algorithm is fairly flexible and is able to handle missing data. However, as with most optimization algorithms, there is no guarantee of finding a global optimum. Interval analysis can be used to compute an enclosure of the range of a function over a specified domain. By enclosing the range of the gradient of the loglikelihood, those parts of the parameter space where the gradient is nonzero can be eliminated as not containing stationary points. An algorithm proceeds by repeatedly bisecting an initial region into smaller regions which are evaluated for the possibility of the gradient being nonzero. Upon termination, all stationary points of the loglikelihood are contained in the remaining regions
Dagstuhl News January - December 2008
"Dagstuhl News" is a publication edited especially for the members of the Foundation "Informatikzentrum Schloss Dagstuhl" to thank them for their support. The News give a summary of the scientific work being done in Dagstuhl. Each Dagstuhl Seminar is presented by a small abstract describing the contents and scientific highlights of the seminar as well as the perspectives or challenges of the research topic
Tuning the Computational Effort: An Adaptive Accuracy-aware Approach Across System Layers
This thesis introduces a novel methodology to realize accuracy-aware systems, which will help designers integrate accuracy awareness into their systems. It proposes an adaptive accuracy-aware approach across system layers that addresses current challenges in that domain, combining and tuning accuracy-aware methods on different system layers. To widen the scope of accuracy-aware computing including approximate computing for other domains, this thesis presents innovative accuracy-aware methods and techniques for different system layers.
The required tuning of the accuracy-aware methods is integrated into a configuration layer that tunes the available knobs of the accuracy-aware methods integrated into a system
First Annual Workshop on Space Operations Automation and Robotics (SOAR 87)
Several topics relative to automation and robotics technology are discussed. Automation of checkout, ground support, and logistics; automated software development; man-machine interfaces; neural networks; systems engineering and distributed/parallel processing architectures; and artificial intelligence/expert systems are among the topics covered
Recommended from our members
Integrating Data Mining and Social Network Techniques into the Development of a Web-based Adaptive Play-based Assessment tool for School Readiness.
A major challenge that faces most families is effectively anticipating how ready to
start school a given child is. Traditional tests are not very effective as they depend on
the skills of the expert conducting the test. It is argued that automated tools are more
attractive especially when they are extended with games capabilities that would be
the most attractive for the children to be seriously involved in the test. The first part
of this thesis reviews the school readiness approaches applied in various countries.
This motivated the development of the sophisticated system described in the thesis.
Extensive research was conducted to enrich the system with features that consider
machine learning and social network aspects. A modified genetic algorithm was
integrated into a web-based stealth assessment tool for school readiness. The
research goal is to create a web-based stealth assessment tool that can learn the user's
skills and adjust the assessment tests accordingly. The user plays various sessions
from various games, while the Genetic Algorithm (GA) selects the upcoming session
or group of sessions to be presented to the user according to his/her skills and status.
The modified GA and the learning procedure were described. A penalizing system
and a fitness heuristic for best choice selection were integrated into the GA. Two
methods for learning were presented, namely a memory system and a no-memory
system. Several methods were presented for the improvement of the speed of
learning. In addition, learning mechanisms were introduced in the social network
aspect to address further usage of stealth assessment automation. The effect of the
relatives and friends on the readiness of the child was studied by investigating the
social communities to which the child belongs and how the trend in these
communities will reflect on to the child under investigation.
The plan is to develop this framework further by incorporating more information
related to social network construction and analysis. Also, it is planned to turn the
framework into a self adaptive one by utilizing the feedback from the usage patterns
to learn and adjust the evaluation process accordingly
xxAI - Beyond Explainable AI
This is an open access book.
Statistical machine learning (ML) has triggered a renaissance of artificial intelligence (AI). While the most successful ML models, including Deep Neural Networks (DNN), have developed better predictivity, they have become increasingly complex, at the expense of human interpretability (correlation vs. causality). The field of explainable AI (xAI) has emerged with the goal of creating tools and models that are both predictive and interpretable and understandable for humans.
Explainable AI is receiving huge interest in the machine learning and AI research communities, across academia, industry, and government, and there is now an excellent opportunity to push towards successful explainable AI applications. This volume will help the research community to accelerate this process, to promote a more systematic use of explainable AI to improve models in diverse applications, and ultimately to better understand how current explainable AI methods need to be improved and what kind of theory of explainable AI is needed.
After overviews of current methods and challenges, the editors include chapters that describe new developments in explainable AI. The contributions are from leading researchers in the field, drawn from both academia and industry, and many of the chapters take a clear interdisciplinary approach to problem-solving. The concepts discussed include explainability, causability, and AI interfaces with humans, and the applications include image processing, natural language, law, fairness, and climate science.https://digitalcommons.unomaha.edu/isqafacbooks/1000/thumbnail.jp
xxAI - Beyond Explainable AI
This is an open access book. Statistical machine learning (ML) has triggered a renaissance of artificial intelligence (AI). While the most successful ML models, including Deep Neural Networks (DNN), have developed better predictivity, they have become increasingly complex, at the expense of human interpretability (correlation vs. causality). The field of explainable AI (xAI) has emerged with the goal of creating tools and models that are both predictive and interpretable and understandable for humans. Explainable AI is receiving huge interest in the machine learning and AI research communities, across academia, industry, and government, and there is now an excellent opportunity to push towards successful explainable AI applications. This volume will help the research community to accelerate this process, to promote a more systematic use of explainable AI to improve models in diverse applications, and ultimately to better understand how current explainable AI methods need to be improved and what kind of theory of explainable AI is needed. After overviews of current methods and challenges, the editors include chapters that describe new developments in explainable AI. The contributions are from leading researchers in the field, drawn from both academia and industry, and many of the chapters take a clear interdisciplinary approach to problem-solving. The concepts discussed include explainability, causability, and AI interfaces with humans, and the applications include image processing, natural language, law, fairness, and climate science
xxAI - Beyond Explainable AI
This is an open access book. Statistical machine learning (ML) has triggered a renaissance of artificial intelligence (AI). While the most successful ML models, including Deep Neural Networks (DNN), have developed better predictivity, they have become increasingly complex, at the expense of human interpretability (correlation vs. causality). The field of explainable AI (xAI) has emerged with the goal of creating tools and models that are both predictive and interpretable and understandable for humans. Explainable AI is receiving huge interest in the machine learning and AI research communities, across academia, industry, and government, and there is now an excellent opportunity to push towards successful explainable AI applications. This volume will help the research community to accelerate this process, to promote a more systematic use of explainable AI to improve models in diverse applications, and ultimately to better understand how current explainable AI methods need to be improved and what kind of theory of explainable AI is needed. After overviews of current methods and challenges, the editors include chapters that describe new developments in explainable AI. The contributions are from leading researchers in the field, drawn from both academia and industry, and many of the chapters take a clear interdisciplinary approach to problem-solving. The concepts discussed include explainability, causability, and AI interfaces with humans, and the applications include image processing, natural language, law, fairness, and climate science