78 research outputs found
Scale-invariant segmentation of dynamic contrast-enhanced perfusion MR-images with inherent scale selection
Selection of the best set of scales is problematic when developing signaldriven
approaches for pixel-based image segmentation. Often, different
possibly conflicting criteria need to be fulfilled in order to obtain the best tradeoff
between uncertainty (variance) and location accuracy. The optimal set of
scales depends on several factors: the noise level present in the image material,
the prior distribution of the different types of segments, the class-conditional
distributions associated with each type of segment as well as the actual size of
the (connected) segments. We analyse, theoretically and through experiments,
the possibility of using the overall and class-conditional error rates as criteria
for selecting the optimal sampling of the linear and morphological scale spaces.
It is shown that the overall error rate is optimised by taking the prior class
distribution in the image material into account. However, a uniform (ignorant)
prior distribution ensures constant class-conditional error rates. Consequently,
we advocate for a uniform prior class distribution when an uncommitted, scaleinvariant
segmentation approach is desired.
Experiments with a neural net classifier developed for segmentation of
dynamic MR images, acquired with a paramagnetic tracer, support the
theoretical results. Furthermore, the experiments show that the addition of
spatial features to the classifier, extracted from the linear or morphological
scale spaces, improves the segmentation result compared to a signal-driven
approach based solely on the dynamic MR signal. The segmentation results
obtained from the two types of features are compared using two novel quality
measures that characterise spatial properties of labelled images
Risk-based neuro-grid architecture for multimodal biometrics
Recent research indicates that multimodal biometrics is the way forward for a highly reliable adoption of biometric identification systems in various applications, such as banks, businesses, government
Genome-wide Copy Number Profiling on High-density Bacterial Artificial Chromosomes, Single-nucleotide Polymorphisms, and Oligonucleotide Microarrays: A Platform Comparison based on Statistical Power Analysis
Recently, comparative genomic hybridization onto bacterial artificial chromosome (BAC) arrays (array-based comparative genomic hybridization) has proved to be successful for the detection of submicroscopic DNA copy-number variations in health and disease. Technological improvements to achieve a higher resolution have resulted in the generation of additional microarray platforms encompassing larger numbers of shorter DNA targets (oligonucleotides). Here, we present a novel method to estimate the ability of a microarray to detect genomic copy-number variations of different sizes and types (i.e. deletions or duplications). We applied our method, which is based on statistical power analysis, to four widely used high-density genomic microarray platforms. By doing so, we found that the high-density oligonucleotide platforms are superior to the BAC platform for the genome-wide detection of copy-number variations smaller than 1 Mb. The capacity to reliably detect single copy-number variations below 100 kb, however, appeared to be limited for all platforms tested. In addition, our analysis revealed an unexpected platform-dependent difference in sensitivity to detect a single copy-number loss and a single copy-number gain. These analyses provide a first objective insight into the true capacities and limitations of different genomic microarrays to detect and define DNA copy-number variations
Learning, Memory, and the Role of Neural Network Architecture
The performance of information processing systems, from artificial neural networks to natural neuronal ensembles, depends heavily on the underlying system architecture. In this study, we compare the performance of parallel and layered network architectures during sequential tasks that require both acquisition and retention of information, thereby identifying tradeoffs between learning and memory processes. During the task of supervised, sequential function approximation, networks produce and adapt representations of external information. Performance is evaluated by statistically analyzing the error in these representations while varying the initial network state, the structure of the external information, and the time given to learn the information. We link performance to complexity in network architecture by characterizing local error landscape curvature. We find that variations in error landscape structure give rise to tradeoffs in performance; these include the ability of the network to maximize accuracy versus minimize inaccuracy and produce specific versus generalizable representations of information. Parallel networks generate smooth error landscapes with deep, narrow minima, enabling them to find highly specific representations given sufficient time. While accurate, however, these representations are difficult to generalize. In contrast, layered networks generate rough error landscapes with a variety of local minima, allowing them to quickly find coarse representations. Although less accurate, these representations are easily adaptable. The presence of measurable performance tradeoffs in both layered and parallel networks has implications for understanding the behavior of a wide variety of natural and artificial learning systems
Specification and assessment of methods supporting the development of neural networks in medicine
Assessment of the influence of preoperative chemotherapy in patients with osteosarcoma by dynamic contrast-enhanced MRI using pharmacokinetic modeling
Confidence intervals for probabilistic network classifiers
Probabilistic networks (Bayesian networks) are suited as statistical pattern classifiers when the feature variables are discrete. It is argued that their white-box character makes them transparent, a requirement in various applications such as, e.g., credit scoring. In addition, the exact error rate of a probabilistic network classifier can be computed without a dataset. First, the exact error rate for probabilistic network classifiers is specified. Secondly, the exact sampling distribution for the conditional probability estimates in a probabilistic network classifier is derived. Each conditional probability is distributed according to the bivariate binomial distribution. Subsequently, an approach for computing the sampling distribution and hence confidence intervals for the posterior probability in a probabilistic network classifier is derived. Our approach results in parametric bootstrap confidence intervals. Experiments with general probabilistic network classifiers, the Naive Bayes classifier and tree augmented Naive Bayes classifiers (TANs) show that our approximation performs well. Also simulations performed with the Alarm network show good results for large training sets. The amount of computation required is exponential in the number of feature variables. For medium and large-scale classification problems, our approach is well suited for quick simulations. A running example from the domain of credit scoring illustrates how to actually compute the sampling distribution of the posterior probability
- …