365 research outputs found
Illustrating answers: an evaluation of automatically retrieved illustrations of answers to medical questions
In this paper we discuss and evaluate a method for automatic text illustration, applied to answers to medical questions. Our method for selecting illustrations is based on the idea that similarities between the answers and picture-related text (the pictureâs caption or the section/paragraph that includes the picture) can be used as evidence that the picture would be appropriate to illustrate the answer.In a user study, participants rated answer presentations consisting of a textual component and a picture. The textual component was a manually written reference answer; the picture was automatically retrieved by measuring the similarity between the text and either the pictureâs caption or its section. The caption-based selection method resulted in more attractive presentations than the section-based method; the caption-based method was also more consistent in selecting informative pictures and showed a greater correlation between user-rated informativeness and the confidence of relevance of the system.When compared to manually selected pictures, we found that automatically selected pictures were rated similarly to decorative pictures, but worse than informative pictures
Quantization and Compressive Sensing
Quantization is an essential step in digitizing signals, and, therefore, an
indispensable component of any modern acquisition system. This book chapter
explores the interaction of quantization and compressive sensing and examines
practical quantization strategies for compressive acquisition systems.
Specifically, we first provide a brief overview of quantization and examine
fundamental performance bounds applicable to any quantization approach. Next,
we consider several forms of scalar quantizers, namely uniform, non-uniform,
and 1-bit. We provide performance bounds and fundamental analysis, as well as
practical quantizer designs and reconstruction algorithms that account for
quantization. Furthermore, we provide an overview of Sigma-Delta
() quantization in the compressed sensing context, and also
discuss implementation issues, recovery algorithms and performance bounds. As
we demonstrate, proper accounting for quantization and careful quantizer design
has significant impact in the performance of a compressive acquisition system.Comment: 35 pages, 20 figures, to appear in Springer book "Compressed Sensing
and Its Applications", 201
Cornerstones of Sampling of Operator Theory
This paper reviews some results on the identifiability of classes of
operators whose Kohn-Nirenberg symbols are band-limited (called band-limited
operators), which we refer to as sampling of operators. We trace the motivation
and history of the subject back to the original work of the third-named author
in the late 1950s and early 1960s, and to the innovations in spread-spectrum
communications that preceded that work. We give a brief overview of the NOMAC
(Noise Modulation and Correlation) and Rake receivers, which were early
implementations of spread-spectrum multi-path wireless communication systems.
We examine in detail the original proof of the third-named author
characterizing identifiability of channels in terms of the maximum time and
Doppler spread of the channel, and do the same for the subsequent
generalization of that work by Bello.
The mathematical limitations inherent in the proofs of Bello and the third
author are removed by using mathematical tools unavailable at the time. We
survey more recent advances in sampling of operators and discuss the
implications of the use of periodically-weighted delta-trains as identifiers
for operator classes that satisfy Bello's criterion for identifiability,
leading to new insights into the theory of finite-dimensional Gabor systems. We
present novel results on operator sampling in higher dimensions, and review
implications and generalizations of the results to stochastic operators, MIMO
systems, and operators with unknown spreading domains
Comparing cancer survivors in population-based samples with those in online cancer communities:Cross-sectional questionnaire study
BACKGROUND: Most Western countries have websites that provide information on cancer and the opportunity to participate in online cancer communities (OCCs). The number of patients with cancer that participate in these OCCs is growing. These patients are relatively easy to approach for research purposes. OBJECTIVE: The objective of this study is to determine the differences and similarities between survivors of cancer in population-based samples and survivors participating in OCCs who use the internet in relation to their illness. METHODS: In 2017, we drew a sample of 539 population-based patients and 531 OCC patients. The population-based patients were sent a paper-based questionnaire, and the OCC patients were sent the same questionnaire on the web. In the questionnaire, we asked patients about their sociodemographics, internet use, sources of information, media use, and wishes regarding future internet use for health careârelated purposes, and the effect of internet use on their health care consumption. RESULTS: The response rate of population-based internet users was 47% (233/496), and that of the OCC group was 40.3% (214/531). The OCC group had a significantly higher education level (P<.001), was younger (P<.001), had more survivors that were employed (P<.001), and attached greater importance to the internet (171/214, 79.9% vs 126/233, 54.1%; P<.001) and fellow survivors (107/214, 50% vs 60/233, 25.8%; P<.001). Compared with the population-based group, the OCC group reported more intensive internet use immediately after diagnosis, during treatment, and during follow-up (P<.001 in each case). There were similarities in terms of the relative importance that survivors attach to the various sources of information, the topics on which they seek information, and their wishes for future eHealth possibilities. The OCC group reported a greater need to participate in a web-based class or chat with others (92/214, 43% vs 44/233, 18.9%). CONCLUSIONS: We conclude that survivors who are members of an OCC are not representative of survivors of cancer in general. There are significant differences in sociodemographic characteristics, internet use during their treatment journey, internet search frequency during their cancer journey, and participation wishes. Using web-based information and communication can support shared decision-making and may facilitate the active participation of patients during their treatment. For research purposes, it is important to take the bias in OCC groups into account
On Deterministic Sketching and Streaming for Sparse Recovery and Norm Estimation
We study classic streaming and sparse recovery problems using deterministic
linear sketches, including l1/l1 and linf/l1 sparse recovery problems (the
latter also being known as l1-heavy hitters), norm estimation, and approximate
inner product. We focus on devising a fixed matrix A in R^{m x n} and a
deterministic recovery/estimation procedure which work for all possible input
vectors simultaneously. Our results improve upon existing work, the following
being our main contributions:
* A proof that linf/l1 sparse recovery and inner product estimation are
equivalent, and that incoherent matrices can be used to solve both problems.
Our upper bound for the number of measurements is m=O(eps^{-2}*min{log n, (log
n / log(1/eps))^2}). We can also obtain fast sketching and recovery algorithms
by making use of the Fast Johnson-Lindenstrauss transform. Both our running
times and number of measurements improve upon previous work. We can also obtain
better error guarantees than previous work in terms of a smaller tail of the
input vector.
* A new lower bound for the number of linear measurements required to solve
l1/l1 sparse recovery. We show Omega(k/eps^2 + klog(n/k)/eps) measurements are
required to recover an x' with |x - x'|_1 <= (1+eps)|x_{tail(k)}|_1, where
x_{tail(k)} is x projected onto all but its largest k coordinates in magnitude.
* A tight bound of m = Theta(eps^{-2}log(eps^2 n)) on the number of
measurements required to solve deterministic norm estimation, i.e., to recover
|x|_2 +/- eps|x|_1.
For all the problems we study, tight bounds are already known for the
randomized complexity from previous work, except in the case of l1/l1 sparse
recovery, where a nearly tight bound is known. Our work thus aims to study the
deterministic complexities of these problems
Restricted Isometries for Partial Random Circulant Matrices
In the theory of compressed sensing, restricted isometry analysis has become
a standard tool for studying how efficiently a measurement matrix acquires
information about sparse and compressible signals. Many recovery algorithms are
known to succeed when the restricted isometry constants of the sampling matrix
are small. Many potential applications of compressed sensing involve a
data-acquisition process that proceeds by convolution with a random pulse
followed by (nonrandom) subsampling. At present, the theoretical analysis of
this measurement technique is lacking. This paper demonstrates that the th
order restricted isometry constant is small when the number of samples
satisfies , where is the length of the pulse.
This bound improves on previous estimates, which exhibit quadratic scaling
qpMerge: Merging different peptide isoforms using a motif centric strategy
Accurate quantification and enumeration of peptide motifs is hampered by redundancy in peptide identification. A single phosphorylation motif may be split across charge states, alternative modifications (e.g. acetylation and oxidation), and multiple miss-cleavage sites which render the biological interpretation of MS data a challenge. In addition motif redundancy can affect quantitative and statistical analysis and prevent a realistic comparison of peptide numbers between datasets. In this study, we present a merging tool set developed for the Galaxy workflow environment to achieve a non-redundant set of quantifications for phospho-motifs. We present a Galaxy workflow to merge three exemplar dataset, and observe reduced phospho-motif redundancy and decreased replicate variation. The qpMerge tools provide a straightforward and reusable approach to facilitating phospho-motif analysis.
The source-code and wiki documentation is publically available at http://sourceforge.net/projects/ppmerge. The galaxy pipeline used in the exemplar analysis can be found at http://www.myexperiment.org/workflows/4186
Label-free quantitative analysis of the casein kinase 2-responsive phosphoproteome of the marine minimal model species Ostreococcus tauri
Casein kinase 2 (CK2) is a protein kinase that phosphorylates a plethora of cellular target proteins involved in processes including DNA repair, cell cycle control, and circadian timekeeping. CK2 is functionally conserved across eukaryotes, although the substrate proteins identified in a range of complex tissues are often different. The marine alga Ostreococcus tauri is a unicellular eukaryotic model organism ideally suited to efficiently study generic roles of CK2 in the cellular circadian clock. Overexpression of CK2 leads to a slow circadian rhythm, verifying functional conservation of CK2 in timekeeping. The proteome was analysed in wildâtype and CK2âoverexpressing algae at dawn and dusk, revealing that differential abundance of the global proteome across the day is largely unaffected by overexpression. However, CK2 activity contributed more strongly to timekeeping at dusk than at dawn. The phosphoproteome of a CK2 overexpression line and cells treated with CK2 inhibitor was therefore analysed and compared to control cells at dusk. We report an extensive catalogue of 447 unique CK2âresponsive differential phosphopeptide motifs to inform future studies into CK2 activity in the circadian clock of more complex tissues. All MS data have been deposited in the ProteomeXchange with identifier PXD000975 (http://proteomecentral.proteomexchange.org/dataset/PXD000975)
- âŚ