353 research outputs found

    Quantifying Privacy: A Novel Entropy-Based Measure of Disclosure Risk

    Full text link
    It is well recognised that data mining and statistical analysis pose a serious treat to privacy. This is true for financial, medical, criminal and marketing research. Numerous techniques have been proposed to protect privacy, including restriction and data modification. Recently proposed privacy models such as differential privacy and k-anonymity received a lot of attention and for the latter there are now several improvements of the original scheme, each removing some security shortcomings of the previous one. However, the challenge lies in evaluating and comparing privacy provided by various techniques. In this paper we propose a novel entropy based security measure that can be applied to any generalisation, restriction or data modification technique. We use our measure to empirically evaluate and compare a few popular methods, namely query restriction, sampling and noise addition.Comment: 20 pages, 4 figure

    A global optimisation approach to range-restricted survey calibration

    Get PDF
    Survey calibration methods modify minimally unit-level sample weights to fit domain-level benchmark constraints (BC). This allows exploitation of auxiliary information, e.g. census totals, to improve the representativeness of sample data (addressing coverage limitations, non-response) and the quality of estimates of population parameters. Calibration methods may fail with samples presenting small/zero counts for some benchmark groups or when range restrictions (RR), such as positivity, are imposed to avoid unrealistic or extreme weights. User-defined modifications of BC/RR performed after encountering non-convergence allow little control on the solution, and penalization approaches modelling infeasibility may not guarantee convergence. Paradoxically, this has led to underuse in calibration of highly disaggregated information, when available. We present an always-convergent flexible two-step Global Optimisation (GO) survey calibration approach. The feasibility of the calibration problem is assessed, and automatically controlled minimum errors in BC or changes in RR are allowed to guarantee convergence in advance, while preserving the good properties of calibration estimators. Modelling alternatives under different scenarios, using various error/change and distance measures are formulated and discussed. The GO approach is validated by calibrating the weights of the 2012 Health Survey for England to a fine age-gender-region cross-tabulation (378 counts) from the 2011 Census in England and Wales

    A global optimisation approach to range-restricted survey calibration

    Get PDF
    Survey calibration methods modify minimally unit-level sample weights to fit domain-level benchmark constraints (BC). This allows exploitation of auxiliary information, e.g. census totals, to improve the representativeness of sample data (addressing coverage limitations, non-response) and the quality of estimates of population parameters. Calibration methods may fail with samples presenting small/zero counts for some benchmark groups or when range restrictions (RR), such as positivity, are imposed to avoid unrealistic or extreme weights. User-defined modifications of BC/RR performed after encountering non-convergence allow little control on the solution, and penalization approaches modelling infeasibility may not guarantee convergence. Paradoxically, this has led to underuse in calibration of highly disaggregated information, when available. We present an always-convergent flexible two-step Global Optimisation (GO) survey calibration approach. The feasibility of the calibration problem is assessed, and automatically controlled minimum errors in BC or changes in RR are allowed to guarantee convergence in advance, while preserving the good properties of calibration estimators. Modelling alternatives under different scenarios, using various error/change and distance measures are formulated and discussed. The GO approach is validated by calibrating the weights of the 2012 Health Survey for England to a fine age-gender-region cross-tabulation (378 counts) from the 2011 Census in England and Wales

    Body fat measurement by bioelectrical impedance and air displacement plethysmography: a cross-validation study to design bioelectrical impedance equations in Mexican adults

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>The study of body composition in specific populations by techniques such as bio-impedance analysis (BIA) requires validation based on standard reference methods. The aim of this study was to develop and cross-validate a predictive equation for bioelectrical impedance using air displacement plethysmography (ADP) as standard method to measure body composition in Mexican adult men and women.</p> <p>Methods</p> <p>This study included 155 male and female subjects from northern Mexico, 20–50 years of age, from low, middle, and upper income levels. Body composition was measured by ADP. Body weight (BW, kg) and height (Ht, cm) were obtained by standard anthropometric techniques. Resistance, R (ohms) and reactance, Xc (ohms) were also measured. A random-split method was used to obtain two samples: one was used to derive the equation by the "all possible regressions" procedure and was cross-validated in the other sample to test predicted versus measured values of fat-free mass (FFM).</p> <p>Results and Discussion</p> <p>The final model was: FFM (kg) = 0.7374 * (Ht<sup>2 </sup>/R) + 0.1763 * (BW) - 0.1773 * (Age) + 0.1198 * (Xc) - 2.4658. R<sup>2 </sup>was 0.97; the square root of the mean square error (SRMSE) was 1.99 kg, and the pure error (PE) was 2.96. There was no difference between FFM predicted by the new equation (48.57 ± 10.9 kg) and that measured by ADP (48.43 ± 11.3 kg). The new equation did not differ from the line of identity, had a high R<sup>2 </sup>and a low SRMSE, and showed no significant bias (0.87 ± 2.84 kg).</p> <p>Conclusion</p> <p>The new bioelectrical impedance equation based on the two-compartment model (2C) was accurate, precise, and free of bias. This equation can be used to assess body composition and nutritional status in populations similar in anthropometric and physical characteristics to this sample.</p

    The Evolution of Compact Binary Star Systems

    Get PDF
    We review the formation and evolution of compact binary stars consisting of white dwarfs (WDs), neutron stars (NSs), and black holes (BHs). Binary NSs and BHs are thought to be the primary astrophysical sources of gravitational waves (GWs) within the frequency band of ground-based detectors, while compact binaries of WDs are important sources of GWs at lower frequencies to be covered by space interferometers (LISA). Major uncertainties in the current understanding of properties of NSs and BHs most relevant to the GW studies are discussed, including the treatment of the natal kicks which compact stellar remnants acquire during the core collapse of massive stars and the common envelope phase of binary evolution. We discuss the coalescence rates of binary NSs and BHs and prospects for their detections, the formation and evolution of binary WDs and their observational manifestations. Special attention is given to AM CVn-stars -- compact binaries in which the Roche lobe is filled by another WD or a low-mass partially degenerate helium-star, as these stars are thought to be the best LISA verification binary GW sources.Comment: 105 pages, 18 figure

    Is human blood a good surrogate for brain tissue in transcriptional studies?

    Get PDF
    Abstract Background Since human brain tissue is often unavailable for transcriptional profiling studies, blood expression data is frequently used as a substitute. The underlying hypothesis in such studies is that genes expressed in brain tissue leave a transcriptional footprint in blood. We tested this hypothesis by relating three human brain expression data sets (from cortex, cerebellum and caudate nucleus) to two large human blood expression data sets (comprised of 1463 individuals). Results We found mean expression levels were weakly correlated between the brain and blood data (r range: [0.24,0.32]). Further, we tested whether co-expression relationships were preserved between the three brain regions and blood. Only a handful of brain co-expression modules showed strong evidence of preservation and these modules could be combined into a single large blood module. We also identified highly connected intramodular "hub" genes inside preserved modules. These preserved intramodular hub genes had the following properties: first, their expression levels tended to be significantly more heritable than those from non-preserved intramodular hub genes (p &lt; 10-90); second, they had highly significant positive correlations with the following cluster of differentiation genes: CD58, CD47, CD48, CD53 and CD164; third, a significant number of them were known to be involved in infection mechanisms, post-transcriptional and post-translational modification and other basic processes. Conclusions Overall, we find transcriptome organization is poorly preserved between brain and blood. However, the subset of preserved co-expression relationships characterized here may aid future efforts to identify blood biomarkers for neurological and neuropsychiatric diseases when brain tissue samples are unavailable

    Insights into the Molecular Basis of L-Form Formation and Survival in Escherichia coli

    Get PDF
    L-forms have been shown to occur among many species of bacteria and are suspected to be involved in persistent infections. Since their discovery in 1935, numerous studies characterizing L-form morphology, growth, and pathogenic potential have been conducted. However, the molecular mechanisms underlying the formation and survival of L-forms remain unknown. Using unstable L-form colonies of Escherichia coli as a model, we performed genome-wide transcriptome analysis and screened a deletion mutant library to study the molecular mechanisms involved in formation and survival of L-forms. Microarray analysis of L-form versus classical colonies revealed many up-regulated genes of unknown function as well as multiple over-expressed stress pathways shared in common with persister cells and biofilms. Mutant screens identified three groups of mutants which displayed varying degrees of defects in L-form colony formation. Group 1 mutants, which showed the strongest defect in L-form colony formation, belonged to pathways involved in cell envelope stress, DNA repair, iron homeostasis, outer membrane biogenesis, and drug efflux/ABC transporters. Four (Group 1) mutants, rcsB, a positive response regulator of colanic acid capsule synthesis, ruvA, a recombinational junction binding protein, fur, a ferric uptake regulator and smpA a small membrane lipoprotein were selected for complementation. Complementation of the mutants using a high-copy overexpression vector failed, while utilization of a low-copy inducible vector successfully restored L-form formation. This work represents the first systematic genetic evaluation of genes and pathways involved in the formation and survival of unstable L-form bacteria. Our findings provide new insights into the molecular mechanisms underlying L-form formation and survival and have implications for understanding the emergence of antibiotic resistance, bacterial persistence and latent infections and designing novel drugs and vaccines

    How Noisy Does a Noisy Miner Have to Be? Amplitude Adjustments of Alarm Calls in an Avian Urban ‘Adapter’

    Get PDF
    Background: Urban environments generate constant loud noise, which creates a formidable challenge for many animals relying on acoustic communication. Some birds make vocal adjustments that reduce auditory masking by altering, for example, the frequency (kHz) or timing of vocalizations. Another adjustment, well documented for birds under laboratory and natural field conditions, is a noise level-dependent change in sound signal amplitude (the ‘Lombard effect’). To date, however, field research on amplitude adjustments in urban environments has focused exclusively on bird song. Methods: We investigated amplitude regulation of alarm calls using, as our model, a successful urban ‘adapter ’ species, the Noisy miner, Manorina melanocephala. We compared several different alarm calls under contrasting noise conditions. Results: Individuals at noisier locations (arterial roads) alarm called significantly more loudly than those at quieter locations (residential streets). Other mechanisms known to improve sound signal transmission in ‘noise’, namely use of higher perches and in-flight calling, did not differ between site types. Intriguingly, the observed preferential use of different alarm calls by Noisy miners inhabiting arterial roads and residential streets was unlikely to have constituted a vocal modification made in response to sound-masking in the urban environment because the calls involved fell within the main frequency range of background anthropogenic noise. Conclusions: The results of our study suggest that a species, which has the ability to adjust the amplitude of its signals

    Protein expression based multimarker analysis of breast cancer samples

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Tissue microarray (TMA) data are commonly used to validate the prognostic accuracy of tumor markers. For example, breast cancer TMA data have led to the identification of several promising prognostic markers of survival time. Several studies have shown that TMA data can also be used to cluster patients into clinically distinct groups. Here we use breast cancer TMA data to cluster patients into distinct prognostic groups.</p> <p>Methods</p> <p>We apply weighted correlation network analysis (WGCNA) to TMA data consisting of 26 putative tumor biomarkers measured on 82 breast cancer patients. Based on this analysis we identify three groups of patients with low (5.4%), moderate (22%) and high (50%) mortality rates, respectively. We then develop a simple threshold rule using a subset of three markers (p53, Na-KATPase-β1, and TGF β receptor II) that can approximately define these mortality groups. We compare the results of this correlation network analysis with results from a standard Cox regression analysis.</p> <p>Results</p> <p>We find that the rule-based grouping variable (referred to as WGCNA*) is an independent predictor of survival time. While WGCNA* is based on protein measurements (TMA data), it validated in two independent Affymetrix microarray gene expression data (which measure mRNA abundance). We find that the WGCNA patient groups differed by 35% from mortality groups defined by a more conventional stepwise Cox regression analysis approach.</p> <p>Conclusions</p> <p>We show that correlation network methods, which are primarily used to analyze the relationships between gene products, are also useful for analyzing the relationships between patients and for defining distinct patient groups based on TMA data. We identify a rule based on three tumor markers for predicting breast cancer survival outcomes.</p

    Decreased Functional Diversity and Biological Pest Control in Conventional Compared to Organic Crop Fields

    Get PDF
    Organic farming is one of the most successful agri-environmental schemes, as humans benefit from high quality food, farmers from higher prices for their products and it often successfully protects biodiversity. However there is little knowledge if organic farming also increases ecosystem services like pest control. We assessed 30 triticale fields (15 organic vs. 15 conventional) and recorded vascular plants, pollinators, aphids and their predators. Further, five conventional fields which were treated with insecticides were compared with 10 non-treated conventional fields. Organic fields had five times higher plant species richness and about twenty times higher pollinator species richness compared to conventional fields. Abundance of pollinators was even more than one-hundred times higher on organic fields. In contrast, the abundance of cereal aphids was five times lower in organic fields, while predator abundances were three times higher and predator-prey ratios twenty times higher in organic fields, indicating a significantly higher potential for biological pest control in organic fields. Insecticide treatment in conventional fields had only a short-term effect on aphid densities while later in the season aphid abundances were even higher and predator abundances lower in treated compared to untreated conventional fields. Our data indicate that insecticide treatment kept aphid predators at low abundances throughout the season, thereby significantly reducing top-down control of aphid populations. Plant and pollinator species richness as well as predator abundances and predator-prey ratios were higher at field edges compared to field centres, highlighting the importance of field edges for ecosystem services. In conclusion organic farming increases biodiversity, including important functional groups like plants, pollinators and predators which enhance natural pest control. Preventative insecticide application in conventional fields has only short-term effects on aphid densities but long-term negative effects on biological pest control. Therefore conventional farmers should restrict insecticide applications to situations where thresholds for pest densities are reached
    • …
    corecore