541 research outputs found

    Negative tunneling magnetoresistance by canted magnetization in MgO/NiO tunnel barriers

    Full text link
    The influence of insertion of an ultra-thin NiO layer between the MgO barrier and ferromagnetic electrode in magnetic tunnel junctions has been investigated by measuring the tunneling magnetoresistance and the X-ray magnetic circular dichroism (XMCD). The magnetoresistance shows a high asymmetry with respect to bias voltage, giving rise to a negative value of -16% at 2.8 K. We attribute this to the formation of non-collinear spin structures in the NiO layer as observed by XMCD. The magnetic moments of the interface Ni atoms tilt from the easy axis due to exchange interaction and the tilting angle decreases with increasing the NiO thickness. The experimental observations are further support by non-collinear spin density functional theory

    Prediction of ‘Nules Clementine’ mandarin susceptibility to rind breakdown disorder using Vis/NIR spectroscopy

    Get PDF
    The use of diffuse reflectance visible and near infrared (Vis/NIR) spectroscopy was explored as a non-destructive technique to predict ‘Nules Clementine’ mandarin fruit susceptibility to rind breakdown (RBD) disorder by detecting rind physico-chemical properties of 80 intact fruit harvested from different canopy positions. Vis/NIR spectra were obtained using a LabSpec® spectrophotometer. Reference physico-chemical data of the fruit were obtained after 8 weeks of storage at 8 °C using conventional methods and included RBD, hue angle, colour index, mass loss, rind dry matter, as well as carbohydrates (sucrose, glucose, fructose, total carbohydrates), and total phenolic acid concentrations. Principal component analysis (PCA) was applied to analyse spectral data to identify clusters in the PCA score plots and outliers. Partial least squares (PLS) regression was applied to spectral data after PCA to develop prediction models for each quality attribute. The spectra were subjected to a test set validation by dividing the data into calibration (n = 48) and test validation (n = 32) sets. An extra set of 40 fruit harvested from a different part of the orchard was used for external validation. PLS-discriminant analysis (PLS-DA) models were developed to sort fruit based on canopy position and RBD susceptibility. Fruit position within the canopy had a significant influence on rind biochemical properties. Outside fruit had higher rind carbohydrates, phenolic acids and dry matter content and lower RBD index than inside fruit. The data distribution in the PCA and PLS-DA models displayed four clusters that could easily be identified. These clusters allowed distinction between fruit from different preharvest treatments. NIR calibration and validation results demonstrated that colour index, dry matter, total carbohydrates and mass loss were predicted with significant accuracy, with residual predictive deviation (RPD) for prediction of 3.83, 3.58, 3.15 and 2.61, respectively. The good correlation between spectral information and carbohydrate content demonstrated the potential of Vis/NIR as a non-destructive tool to predict fruit susceptibility to RBD

    CytoNorm : a normalization algorithm for cytometry data

    Get PDF
    High-dimensional flow cytometry has matured to a level that enables deep phenotyping of cellular systems at a clinical scale. The resulting high-content data sets allow characterizing the human immune system at unprecedented single cell resolution. However, the results are highly dependent on sample preparation and measurements might drift over time. While various controls exist for assessment and improvement of data quality in a single sample, the challenges of cross-sample normalization attempts have been limited to aligning marker distributions across subjects. These approaches, inspired by bulk genomics and proteomics assays, ignore the single-cell nature of the data and risk the removal of biologically relevant signals. This work proposes CytoNorm, a normalization algorithm to ensure internal consistency between clinical samples based on shared controls across various study batches. Data from the shared controls is used to learn the appropriate transformations for each batch (e.g., each analysis day). Importantly, some sources of technical variation are strongly influenced by the amount of protein expressed on specific cell types, requiring several population-specific transformations to normalize cells from a heterogeneous sample. To address this, our approach first identifies the overall cellular distribution using a clustering step, and calculates subset-specific transformations on the control samples by computing their quantile distributions and aligning them with splines. These transformations are then applied to all other clinical samples in the batch to remove the batch-specific variations. We evaluated the algorithm on a customized data set with two shared controls across batches. One control sample was used for calculation of the normalization transformations and the second control was used as a blinded test set and evaluated with Earth Mover's distance. Additional results are provided using two real-world clinical data sets. Overall, our method compared favorably to standard normalization procedures

    Discriminative and informative features for biomolecular text mining with ensemble feature selection

    Get PDF
    Motivation: In the field of biomolecular text mining, black box behavior of machine learning systems currently limits understanding of the true nature of the predictions. However, feature selection (FS) is capable of identifying the most relevant features in any supervised learning setting, providing insight into the specific properties of the classification algorithm. This allows us to build more accurate classifiers while at the same time bridging the gap between the black box behavior and the end-user who has to interpret the results

    A full-body motion capture gait dataset of 138 able-bodied adults across the life span and 50 stroke survivors

    Get PDF
    \ua9 2023, The Author(s).This reference dataset contains biomechanical data of 138 able-bodied adults (21–86 years) and 50 stroke survivors walking bare-footed at their preferred speed. It is unique due to its size, and population, including adults across the life-span and over 70 years, as well as stroke survivors. Full-body kinematics (PiG-model), kinetics and muscle activity of 14 back and lower limbs muscles was collected with a Vicon motion capture system, ground-embedded force plates, and a synchronized surface EMG system. The data is reliable to compare within and between groups as the same methodology and infrastructure were used to gather all data. Both source files (C3D) and post-processed ready-to-use stride-normalized kinematics, kinetics and EMG data (MAT-file, Excel file) are available, allowing high flexibility and accessibility of analysis for both researchers and clinicians. These records are valuable to examine ageing, typical and hemiplegic gait, while also offering a wide range of reference data which can be utilized for age-matched controls during normal walking

    Machine learning for automatic prediction of the quality of electrophysiological recordings

    Get PDF
    The quality of electrophysiological recordings varies a lot due to technical and biological variability and neuroscientists inevitably have to select “good” recordings for further analyses. This procedure is time-consuming and prone to selection biases. Here, we investigate replacing human decisions by a machine learning approach. We define 16 features, such as spike height and width, select the most informative ones using a wrapper method and train a classifier to reproduce the judgement of one of our expert electrophysiologists. Generalisation performance is then assessed on unseen data, classified by the same or by another expert. We observe that the learning machine can be equally, if not more, consistent in its judgements as individual experts amongst each other. Best performance is achieved for a limited number of informative features; the optimal feature set being different from one data set to another. With 80–90% of correct judgements, the performance of the system is very promising within the data sets of each expert but judgments are less reliable when it is used across sets of recordings from different experts. We conclude that the proposed approach is relevant to the selection of electrophysiological recordings, provided parameters are adjusted to different types of experiments and to individual experimenters

    Digging into acceptor splice site prediction : an iterative feature selection approach

    Get PDF
    Feature selection techniques are often used to reduce data dimensionality, increase classification performance, and gain insight into the processes that generated the data. In this paper, we describe an iterative procedure of feature selection and feature construction steps, improving the classification of acceptor splice sites, an important subtask of gene prediction. We show that acceptor prediction can benefit from feature selection, and describe how feature selection techniques can be used to gain new insights in the classification of acceptor sites. This is illustrated by the identification of a new, biologically motivated feature: the AG-scanning feature. The results described in this paper contribute both to the domain of gene prediction, and to research in feature selection techniques, describing a new wrapper based feature weighting method that aids in knowledge discovery when dealing with complex datasets

    Differential expression of lncRNAs during the HIV replication cycle: an underestimated layer in the HIV-host interplay.

    Get PDF
    Studying the effects of HIV infection on the host transcriptome has typically focused on protein-coding genes. However, recent advances in the field of RNA sequencing revealed that long non-coding RNAs (lncRNAs) add an extensive additional layer to the cell's molecular network. Here, we performed transcriptome profiling throughout a primary HIV infection in vitro to investigate lncRNA expression at the different HIV replication cycle processes (reverse transcription, integration and particle production). Subsequently, guilt-by-association, transcription factor and co-expression analysis were performed to infer biological roles for the lncRNAs identified in the HIV-host interplay. Many lncRNAs were suggested to play a role in mechanisms relying on proteasomal and ubiquitination pathways, apoptosis, DNA damage responses and cell cycle regulation. Through transcription factor binding analysis, we found that lncRNAs display a distinct transcriptional regulation profile as compared to protein coding mRNAs, suggesting that mRNAs and lncRNAs are independently modulated. In addition, we identified five differentially expressed lncRNA-mRNA pairs with mRNA involvement in HIV pathogenesis with possible cis regulatory lncRNAs that control nearby mRNA expression and function. Altogether, the present study demonstrates that lncRNAs add a new dimension to the HIV-host interplay and should be further investigated as they may represent targets for controlling HIV replication

    Identification of disease-causing genes using microarray data mining and gene ontology

    Get PDF
    Background: One of the best and most accurate methods for identifying disease-causing genes is monitoring gene expression values in different samples using microarray technology. One of the shortcomings of microarray data is that they provide a small quantity of samples with respect to the number of genes. This problem reduces the classification accuracy of the methods, so gene selection is essential to improve the predictive accuracy and to identify potential marker genes for a disease. Among numerous existing methods for gene selection, support vector machine-based recursive feature elimination (SVMRFE) has become one of the leading methods, but its performance can be reduced because of the small sample size, noisy data and the fact that the method does not remove redundant genes. Methods: We propose a novel framework for gene selection which uses the advantageous features of conventional methods and addresses their weaknesses. In fact, we have combined the Fisher method and SVMRFE to utilize the advantages of a filtering method as well as an embedded method. Furthermore, we have added a redundancy reduction stage to address the weakness of the Fisher method and SVMRFE. In addition to gene expression values, the proposed method uses Gene Ontology which is a reliable source of information on genes. The use of Gene Ontology can compensate, in part, for the limitations of microarrays, such as having a small number of samples and erroneous measurement results. Results: The proposed method has been applied to colon, Diffuse Large B-Cell Lymphoma (DLBCL) and prostate cancer datasets. The empirical results show that our method has improved classification performance in terms of accuracy, sensitivity and specificity. In addition, the study of the molecular function of selected genes strengthened the hypothesis that these genes are involved in the process of cancer growth. Conclusions: The proposed method addresses the weakness of conventional methods by adding a redundancy reduction stage and utilizing Gene Ontology information. It predicts marker genes for colon, DLBCL and prostate cancer with a high accuracy. The predictions made in this study can serve as a list of candidates for subsequent wet-lab verification and might help in the search for a cure for cancers

    Identify error-sensitive patterns by decision tree

    Full text link
    © Springer International Publishing Switzerland 2015. When errors are inevitable during data classification, finding a particular part of the classification model which may be more susceptible to error than others, when compared to finding an Achilles’ heel of the model in a casual way, may help uncover specific error-sensitive value patterns and lead to additional error reduction measures. As an initial phase of the investigation, this study narrows the scope of problem by focusing on decision trees as a pilot model, develops a simple and effective tagging method to digitize individual nodes of a binary decision tree for node-level analysis, to link and track classification statistics for each node in a transparent way, to facilitate the identification and examination of the potentially “weakest” nodes and error-sensitive value patterns in decision trees, to assist cause analysis and enhancement development. This digitization method is not an attempt to re-develop or transform the existing decision tree model, but rather, a pragmatic node ID formulation that crafts numeric values to reflect the tree structure and decision making paths, to expand post-classification analysis to detailed node-level. Initial experiments have shown successful results in locating potentially high-risk attribute and value patterns; this is an encouraging sign to believe this study worth further exploration
    corecore