1,784 research outputs found

    Detection of polarization from the E^4\Pi-A^4\Pi system of FeH in sunspot spectra

    Full text link
    Here we report the first detection of polarization signals induced by the Zeeman effect in spectral lines of the E^4\Pi-A^4\Pi system of FeH located around 1.6 μ\mum. Motivated by the tentative detection of this band in the intensity spectrum of late-type dwarfs, we have investigated the full Stokes sunspot spectrum finding circular and linear polarization signatures that we associate with the FeH lines of the E^4\Pi-A^4\Pi band system. We investigate the Zeeman effect in these molecular transitions pointing out that in Hund's case (a) coupling the effective Land\'e factors are never negative. For this reason, the fact that our spectropolarimetric observations indicate that the Land\'e factors of pairs of FeH lines have opposite signs, prompt us to conclude that the E^4\Pi-A^4\Pi system must be in intermediate angular momentum coupling between Hund's cases (a) and (b). We emphasize that theoretical and/or laboratory investigations of this molecular system are urgently needed for exploiting its promising diagnostic capabilities.Comment: 11 pages, 4 figures, accepted for publication in Astrophysical Journal Letter

    Error propagation in polarimetric demodulation

    Full text link
    The polarization analysis of the light is typically carried out using modulation schemes. The light of unknown polarization state is passed through a set of known modulation optics and a detector is used to measure the total intensity passing the system. The modulation optics is modified several times and, with the aid of such several measurements, the unknown polarization state of the light can be inferred. How to find the optimal demodulation process has been investigated in the past. However, since the modulation matrix has to be measured for a given instrument and the optical elements can present problems of repeatability, some uncertainty is present in the elements of the modulation matrix and/or covariances between these elements. We analyze in detail this issue, presenting analytical formulae for calculating the covariance matrix produced by the propagation of such uncertainties on the demodulation matrix, on the inferred Stokes parameters and on the efficiency of the modulation process. We demonstrate that, even if the covariance matrix of the modulation matrix is diagonal, the covariance matrix of the demodulation matrix is, in general, non-diagonal because matrix inversion is a nonlinear operation. This propagates through the demodulation process and induces correlations on the inferred Stokes parameters.Comment: 18 pages, 3 figures, accepted for publication in Applied Optic

    On the Role of Text Preprocessing in Neural Network Architectures: An Evaluation Study on Text Categorization and Sentiment Analysis

    Full text link
    Text preprocessing is often the first step in the pipeline of a Natural Language Processing (NLP) system, with potential impact in its final performance. Despite its importance, text preprocessing has not received much attention in the deep learning literature. In this paper we investigate the impact of simple text preprocessing decisions (particularly tokenizing, lemmatizing, lowercasing and multiword grouping) on the performance of a standard neural text classifier. We perform an extensive evaluation on standard benchmarks from text categorization and sentiment analysis. While our experiments show that a simple tokenization of input text is generally adequate, they also highlight significant degrees of variability across preprocessing techniques. This reveals the importance of paying attention to this usually-overlooked step in the pipeline, particularly when comparing different models. Finally, our evaluation provides insights into the best preprocessing practices for training word embeddings.Comment: Blackbox EMNLP 2018. 7 page

    From Word to Sense Embeddings: A Survey on Vector Representations of Meaning

    Get PDF
    Over the past years, distributed semantic representations have proved to be effective and flexible keepers of prior knowledge to be integrated into downstream applications. This survey focuses on the representation of meaning. We start from the theoretical background behind word vector space models and highlight one of their major limitations: the meaning conflation deficiency, which arises from representing a word with all its possible meanings as a single vector. Then, we explain how this deficiency can be addressed through a transition from the word level to the more fine-grained level of word senses (in its broader acceptation) as a method for modelling unambiguous lexical meaning. We present a comprehensive overview of the wide range of techniques in the two main branches of sense representation, i.e., unsupervised and knowledge-based. Finally, this survey covers the main evaluation procedures and applications for this type of representation, and provides an analysis of four of its important aspects: interpretability, sense granularity, adaptability to different domains and compositionality.Comment: 46 pages, 8 figures. Published in Journal of Artificial Intelligence Researc

    NASARI: a novel approach to a Semantically-Aware Representation of items

    Get PDF
    The semantic representation of individual word senses and concepts is of fundamental importance to several applications in Natural Language Processing. To date, concept modeling techniques have in the main based their representation either on lexicographic resources, such as WordNet, or on encyclopedic resources, such as Wikipedia. We propose a vector representation technique that combines the complementary knowledge of both these types of resource. Thanks to its use of explicit semantics combined with a novel cluster-based dimensionality reduction and an effective weighting scheme, our representation attains state-of-the-art performance on multiple datasets in two standard benchmarks: word similarity and sense clustering. We are releasing our vector representations at http://lcl.uniroma1.it/nasari/
    corecore