8,332 research outputs found

    Assessing luminosity correlations via cluster analysis: Evidence for dual tracks in the radio/X-ray domain of black hole X-ray binaries

    Full text link
    [abridged] The radio:X-ray correlation for hard and quiescent state black hole X-ray binaries is critically investigated in this paper. New observations of known sources, along with newly discovered ones, have resulted in an increasingly large number of outliers lying well outside the scatter about the quoted best-fit relation. Here, we employ and compare state of the art data clustering techniques in order to identify and characterize different data groupings within the radio:X-ray luminosity plane for 18 hard and quiescent state black hole X-ray binaries with nearly simultaneous multi-wavelength coverage. Linear regression is then carried out on the clustered data to infer the parameters of a relationship of the form {ell}_{r}=alpha+beta {ell}_x through a Bayesian approach (where {ell} denotes log lum). We conclude that the two cluster model, with independent linear fits, is a significant improvement over fitting all points as a single cluster. While the upper track slope (0.63\pm0.03) is consistent, within the errors, with the fitted slope for the 2003 relation (0.7\pm0.1), the lower track slope (0.98\pm0.08) is not consistent with the upper track, nor it is with the widely adopted value of ~1.4 for the neutron stars. The two luminosity tracks do not reflect systematic differences in black hole spins as estimated either from reflection, or continuum fitting method. These results are insensitive to the selection of sub-samples, accuracy in the distances, and to the treatment of upper limits. Besides introducing a further level of complexity in understanding the interplay between synchrotron and Comptonised emission from black hole X-ray binaries, the existence of two tracks in the radio:X-ray domain underscores that a high level of caution must be exercised when employing black hole luminosity relations for the purpose of estimating a third parameter, such as distance or mass.Comment: MNRAS, in press (10 pages, 7 figures

    Similarities and differences in lipidomics profiles among healthy monozygotic twin pairs.

    Get PDF
    Differences in genetic background and/or environmental exposure among individuals are expected to give rise to differences in measurable characteristics, or phenotypes. Consequently, genetic resemblance and similarities in environment should manifest as similarities in phenotypes. The metabolome reflects many of the system properties, and is therefore an important part of the phenotype. Nevertheless, it has not yet been examined to what extent individuals sharing part of their genome and/or environment indeed have similar metabolomes. Here we present the results of hierarchical clustering of blood plasma lipid profile data obtained by liquid chromatographymass spectrometry from 23 healthy, 18-year-old twin pairs, of which 21 pairs were monozygotic, and 8 of their siblings. For 13 monozygotic twin pairs, within-pair similarities in relative concentrations of the detected lipids were indeed larger than the similarities with any other study participant. We demonstrate such high coclustering to be unexpected on basis of chance. The similarities between dizygotic twins and between nontwin siblings, as well as between nonfamilial participants, were less pronounced. In a number of twin pairs, within-pair dissimilarity of lipid profiles positively correlated with increased blood plasma concentrations of C-reactive protein in one twin. In conclusion, this study demonstrates that in healthy individuals, the individual genetic background contributes to the blood plasma lipid profile. Furthermore, lipid profiling may prove useful in monitoring health status, for example, in the context of personalized medicine. © 2008 Mary Ann Liebert, Inc. Chemicals / CAS: C-Reactive Protein, 9007-41-4; Lipid

    Complex Independent Component Analysis of Frequency-Domain Electroencephalographic Data

    Full text link
    Independent component analysis (ICA) has proven useful for modeling brain and electroencephalographic (EEG) data. Here, we present a new, generalized method to better capture the dynamics of brain signals than previous ICA algorithms. We regard EEG sources as eliciting spatio-temporal activity patterns, corresponding to, e.g., trajectories of activation propagating across cortex. This leads to a model of convolutive signal superposition, in contrast with the commonly used instantaneous mixing model. In the frequency-domain, convolutive mixing is equivalent to multiplicative mixing of complex signal sources within distinct spectral bands. We decompose the recorded spectral-domain signals into independent components by a complex infomax ICA algorithm. First results from a visual attention EEG experiment exhibit (1) sources of spatio-temporal dynamics in the data, (2) links to subject behavior, (3) sources with a limited spectral extent, and (4) a higher degree of independence compared to sources derived by standard ICA.Comment: 21 pages, 11 figures. Added final journal reference, fixed minor typo

    The Accounting Network: how financial institutions react to systemic crisis

    Get PDF
    The role of Network Theory in the study of the financial crisis has been widely spotted in the latest years. It has been shown how the network topology and the dynamics running on top of it can trigger the outbreak of large systemic crisis. Following this methodological perspective we introduce here the Accounting Network, i.e. the network we can extract through vector similarities techniques from companies' financial statements. We build the Accounting Network on a large database of worldwide banks in the period 2001-2013, covering the onset of the global financial crisis of mid-2007. After a careful data cleaning, we apply a quality check in the construction of the network, introducing a parameter (the Quality Ratio) capable of trading off the size of the sample (coverage) and the representativeness of the financial statements (accuracy). We compute several basic network statistics and check, with the Louvain community detection algorithm, for emerging communities of banks. Remarkably enough sensible regional aggregations show up with the Japanese and the US clusters dominating the community structure, although the presence of a geographically mixed community points to a gradual convergence of banks into similar supranational practices. Finally, a Principal Component Analysis procedure reveals the main economic components that influence communities' heterogeneity. Even using the most basic vector similarity hypotheses on the composition of the financial statements, the signature of the financial crisis clearly arises across the years around 2008. We finally discuss how the Accounting Networks can be improved to reflect the best practices in the financial statement analysis

    Sequences of purchases in credit card data reveal life styles in urban populations

    Full text link
    Zipf-like distributions characterize a wide set of phenomena in physics, biology, economics and social sciences. In human activities, Zipf-laws describe for example the frequency of words appearance in a text or the purchases types in shopping patterns. In the latter, the uneven distribution of transaction types is bound with the temporal sequences of purchases of individual choices. In this work, we define a framework using a text compression technique on the sequences of credit card purchases to detect ubiquitous patterns of collective behavior. Clustering the consumers by their similarity in purchases sequences, we detect five consumer groups. Remarkably, post checking, individuals in each group are also similar in their age, total expenditure, gender, and the diversity of their social and mobility networks extracted by their mobile phone records. By properly deconstructing transaction data with Zipf-like distributions, this method uncovers sets of significant sequences that reveal insights on collective human behavior.Comment: 30 pages, 26 figure

    Clustering by compression

    Full text link
    We present a new method for clustering based on compression. The method doesn't use subject-specific features or background knowledge, and works as follows: First, we determine a universal similarity distance, the normalized compression distance or NCD, computed from the lengths of compressed data files (singly and in pairwise concatenation). Second, we apply a hierarchical clustering method. The NCD is universal in that it is not restricted to a specific application area, and works across application area boundaries. A theoretical precursor, the normalized information distance, co-developed by one of the authors, is provably optimal but uses the non-computable notion of Kolmogorov complexity. We propose precise notions of similarity metric, normal compressor, and show that the NCD based on a normal compressor is a similarity metric that approximates universality. To extract a hierarchy of clusters from the distance matrix, we determine a dendrogram (binary tree) by a new quartet method and a fast heuristic to implement it. The method is implemented and available as public software, and is robust under choice of different compressors. To substantiate our claims of universality and robustness, we report evidence of successful application in areas as diverse as genomics, virology, languages, literature, music, handwritten digits, astronomy, and combinations of objects from completely different domains, using statistical, dictionary, and block sorting compressors. In genomics we presented new evidence for major questions in Mammalian evolution, based on whole-mitochondrial genomic analysis: the Eutherian orders and the Marsupionta hypothesis against the Theria hypothesis.Comment: LaTeX, 27 pages, 20 figure

    Mapping Science Based on Research Content Similarity

    Get PDF
    Maps of science representing the structure of science help us understand science and technology development. Thus, research in scientometrics has developed techniques for analyzing research activities and for measuring their relationships; however, navigating the recent scientific landscape is still challenging, since conventional inter-citation and co-citation analysis has difficulty in applying to recently published articles and ongoing projects. Therefore, to characterize what is being attempted in the current scientific landscape, this article proposes a content-based method of locating research articles/projects in a multi-dimensional space using word/paragraph embedding. Specifically, for addressing an unclustered problem, we introduced cluster vectors based on the information entropies of technical concepts. The experimental results showed that our method formed a clustered map from approx. 300 k IEEE articles and NSF projects from 2012 to 2016. Finally, we confirmed that formation of specific research areas can be captured as changes in the network structure

    A quantitative method to analyse an open-ended questionnaire: A case study about the Boltzmann Factor

    Get PDF
    This paper describes a quantitative method to analyse an openended questionnaire. Student responses to a specially designed written questionnaire are quantitatively analysed by not hierarchical clustering called k-means method. Through this we can characterise behaviour students with respect their expertise to formulate explanations for phenomena or processes and/or use a given model in the different context. The physics topic is about the Boltzmann Factor, which allows the students to have a unifying view of different phenomena in different contexts
    corecore