762 research outputs found
The Effect of Therapeutic Horseback Riding on Balance and Self-Efficacy in Children with Developmental Disabilities
The prevalence of developmental disabilities in children in the United States is a serious problem. Since children with developmental disabilities often show decreased self-efficacy and balance, researchers have studied the effects of interventions in this population. The purpose of this study is to determine the effect of a 10-week THR session on balance and task-specific self-efficacy in children with physical disabilities ages 6 to 18 years old. Bandura’s social cognitive theory and The Physical Stress Theory will guide the quasi-experimental study. A pre-test post-test design will be implemented over a 12 week span at 3 different riding centers in the Midwest United States. Data collection will begin after an approval from the university review board and obtaining a signed informed consent and assent. Data will be collected through surveys and analyzed with same-sample t-tests
Pre-endoscopy SARS-CoV-2 testing strategy during COVID-19 pandemic: the care must go on
Background: In response to the COVID-19 pandemic, endoscopic societies initially recommended reduction of
endoscopic procedures. In particular non-urgent endoscopies should be postponed. However, this might lead to
unnecessary delay in diagnosing gastrointestinal conditions.
Methods: Retrospectively we analysed the gastrointestinal endoscopies performed at the Central Endoscopy Unit
of Saarland University Medical Center during seven weeks from 23 March to 10 May 2020 and present our real-world
single-centre experience with an individualized rtPCR-based pre-endoscopy SARS-CoV-2 testing strategy. We also
present our experience with this strategy in 2021.
Results: Altogether 359 gastrointestinal endoscopies were performed in the initial period. The testing strategy enabled us to conservatively handle endoscopy programme reduction (44% reduction as compared 2019) during the frst
wave of the COVID-19 pandemic. The results of COVID-19 rtPCR from nasopharyngeal swabs were available in 89%
of patients prior to endoscopies. Apart from six patients with known COVID-19, all other tested patients were negative. The frequencies of endoscopic therapies and clinically signifcant fndings did not difer between patients with
or without SARS-CoV-2 tests. In 2021 we were able to unrestrictedly perform all requested endoscopic procedures
(>5000 procedures) by applying the rtPCR-based pre-endoscopy SARS-CoV-2 testing strategy, regardless of next
waves of COVID-19. Only two out-patients (1893 out-patient procedures) were tested positive in the year 2021.
Conclusion: A structured pre-endoscopy SARS-CoV-2 testing strategy is feasible in the clinical routine of an endoscopy unit. rtPCR-based pre-endoscopy SARS-CoV-2 testing safely allowed unrestricted continuation of endoscopic
procedures even in the presence of high incidence rates of COVID-19. Given the low frequency of positive tests, the
absolute efect of pre-endoscopy testing on viral transmission may be low when FFP-2 masks are regularly used
A Kernel Method for the Two-sample Problem
We propose a framework for analyzing and comparing distributions, allowing us to design statistical tests to determine if two samples are drawn from different distributions. Our test statistic is the largest difference in expectations over functions in the unit ball of a reproducing kernel Hilbert space (RKHS). We present two tests based on large deviation bounds for the test statistic, while a third is based on the asymptotic distribution of this statistic. The test statistic can be computed in quadratic time, although efficient linear time approximations are available. Several classical metrics on distributions are recovered when the function space used to compute the difference in expectations is allowed to be more general (eg.~a Banach space). We apply our two-sample tests to a variety of problems, including attribute matching for databases using the Hungarian marriage method, where they perform strongly. Excellent performance is also obtained when comparing distributions over graphs, for which these are the first such tests
A framework for space-efficient string kernels
String kernels are typically used to compare genome-scale sequences whose
length makes alignment impractical, yet their computation is based on data
structures that are either space-inefficient, or incur large slowdowns. We show
that a number of exact string kernels, like the -mer kernel, the substrings
kernels, a number of length-weighted kernels, the minimal absent words kernel,
and kernels with Markovian corrections, can all be computed in time and
in bits of space in addition to the input, using just a
data structure on the Burrows-Wheeler transform of the
input strings, which takes time per element in its output. The same
bounds hold for a number of measures of compositional complexity based on
multiple value of , like the -mer profile and the -th order empirical
entropy, and for calibrating the value of using the data
Statistical Mechanical Development of a Sparse Bayesian Classifier
The demand for extracting rules from high dimensional real world data is
increasing in various fields. However, the possible redundancy of such data
sometimes makes it difficult to obtain a good generalization ability for novel
samples. To resolve this problem, we provide a scheme that reduces the
effective dimensions of data by pruning redundant components for bicategorical
classification based on the Bayesian framework. First, the potential of the
proposed method is confirmed in ideal situations using the replica method.
Unfortunately, performing the scheme exactly is computationally difficult. So,
we next develop a tractable approximation algorithm, which turns out to offer
nearly optimal performance in ideal cases when the system size is large.
Finally, the efficacy of the developed classifier is experimentally examined
for a real world problem of colon cancer classification, which shows that the
developed method can be practically useful.Comment: 13 pages, 6 figure
On landmark selection and sampling in high-dimensional data analysis
In recent years, the spectral analysis of appropriately defined kernel
matrices has emerged as a principled way to extract the low-dimensional
structure often prevalent in high-dimensional data. Here we provide an
introduction to spectral methods for linear and nonlinear dimension reduction,
emphasizing ways to overcome the computational limitations currently faced by
practitioners with massive datasets. In particular, a data subsampling or
landmark selection process is often employed to construct a kernel based on
partial information, followed by an approximate spectral analysis termed the
Nystrom extension. We provide a quantitative framework to analyse this
procedure, and use it to demonstrate algorithmic performance bounds on a range
of practical approaches designed to optimize the landmark selection process. We
compare the practical implications of these bounds by way of real-world
examples drawn from the field of computer vision, whereby low-dimensional
manifold structure is shown to emerge from high-dimensional video data streams.Comment: 18 pages, 6 figures, submitted for publicatio
Hilbert Space Representations of Probability Distributions
Many problems in unsupervised learning require the analysis of features of probability distributions. At the most fundamental level, we might wish to determine whether two distributions are the same, based on samples from each - this is known as the two-sample or homogeneity problem. We use kernel methods to address this problem, by mapping probability distributions to elements in a reproducing kernel Hilbert space (RKHS). Given a sufficiently rich RKHS, these representations are unique: thus comparing feature space representations allows us to compare distributions without ambiguity. Applications include testing whether cancer subtypes are distinguishable on the basis of DNA microarray data, and whether low frequency oscillations measured at an electrode in the cortex have a different distribution during a neural spike. A more difficult problem is to discover whether two random variables drawn from a joint distribution are independent. It turns out that any dependence between pairs of random variables can be encoded in a cross-covariance operator between appropriate RKHS representations of the variables, and we may test independence by looking at a norm of the operator. We demonstrate this independence test by establishing dependence between an English text and its French translation, as opposed to French text on the same topic but otherwise unrelated. Finally, we show that this operator norm is itself a difference in feature means
The devices, experimental scaffolds, and biomaterials ontology (DEB): a tool for mapping, annotation, and analysis of biomaterials' data
The size and complexity of the biomaterials literature makes systematic data analysis an excruciating manual task. A practical solution is creating databases and information resources. Implant design and biomaterials research can greatly benefit from an open database for systematic data retrieval. Ontologies are pivotal to knowledge base creation, serving to represent and organize domain knowledge. To name but two examples, GO, the gene ontology, and CheBI, Chemical Entities of Biological Interest ontology and their associated databases are central resources to their respective research communities. The creation of the devices, experimental scaffolds, and biomaterials ontology (DEB), an open resource for organizing information about biomaterials, their design, manufacture, and biological testing, is described. It is developed using text analysis for identifying ontology terms from a biomaterials gold standard corpus, systematically curated to represent the domain's lexicon. Topics covered are validated by members of the biomaterials research community. The ontology may be used for searching terms, performing annotations for machine learning applications, standardized meta-data indexing, and other cross-disciplinary data exploitation. The input of the biomaterials community to this effort to create data-driven open-access research tools is encouraged and welcomed.Preprin
Knot selection in sparse Gaussian processes with a variational objective function
Sparse, knot‐based Gaussian processes have enjoyed considerable success as scalable approximations of full Gaussian processes. Certain sparse models can be derived through specific variational approximations to the true posterior, and knots can be selected to minimize the Kullback‐Leibler divergence between the approximate and true posterior. While this has been a successful approach, simultaneous optimization of knots can be slow due to the number of parameters being optimized. Furthermore, there have been few proposed methods for selecting the number of knots, and no experimental results exist in the literature. We propose using a one‐at‐a‐time knot selection algorithm based on Bayesian optimization to select the number and locations of knots. We showcase the competitive performance of this method relative to optimization of knots simultaneously on three benchmark datasets, but at a fraction of the computational cost
- …