71 research outputs found

    Profiling Cryptocurrency Influencers with Few-Shot Learning Using Data Augmentation and ELECTRA

    Get PDF
    With this work we propose an application of the ELECTRA Transformer, fine-tuned on two augmented version of the same training dataset. Our team developed the novel framework for taking part at the Profiling Cryptocurrency Influencers with Few-shot Learning task hosted at PAN@CLEF2023. Our proposed strategy consists of an early data augmentation stage followed by a fine-tuning of ELECTRA. At the first stage we augment the original training dataset provided by the organizers using backtranslation. Using this augmented version of the training dataset, we perform a fine tuning of ELECTRA. Finally, using the fine-tuned version of ELECTRA, we inference the labels of the samples provided in the test set. To develop and test our model we used a two-ways validation on the training set. Firstly, we evaluate all the metrics on the augmented training set, and then we evaluate on the original training set. The metrics we considered span from accuracy to Macro F1, to Micro F1, to Recall and Precision. According to the official evaluator, our best submission reached a Macro F1 value equal to 0.3762

    Is text preprocessing still worth the time? A comparative survey on the influence of popular preprocessing methods on Transformers and traditional classifiers

    Get PDF
    With the advent of the modern pre-trained Transformers, the text preprocessing has started to be neglected and not specifically addressed in recent NLP literature. However, both from a linguistic and from a computer science point of view, we believe that even when using modern Transformers, text preprocessing can significantly impact on the performance of a classification model. We want to investigate and compare, through this study, how preprocessing impacts on the Text Classification (TC) performance of modern and traditional classification models. We report and discuss the preprocessing techniques found in the literature and their most recent variants or applications to address TC tasks in different domains. In order to assess how much the preprocessing affects classification performance, we apply the three top referenced preprocessing techniques (alone or in combination) to four publicly available datasets from different domains. Then, nine machine learning models – including modern Transformers – get the preprocessed text as input. The results presented show that an educated choice on the text preprocessing strategy to employ should be based on the task as well as on the model considered. Outcomes in this survey show that choosing the best preprocessing technique – in place of the worst – can significantly improve accuracy on the classification (up to 25%, as in the case of an XLNet on the IMDB dataset). In some cases, by means of a suitable preprocessing strategy, even a simple Naïve Bayes classifier proved to outperform (i.e., by 2% in accuracy) the best performing Transformer. We found that Transformers and traditional models exhibit a higher impact of the preprocessing on the TC performance. Our main findings are: (1) also on modern pre-trained language models, preprocessing can affect performance, depending on the datasets and on the preprocessing technique or combination of techniques used, (2) in some cases, using a proper preprocessing strategy, simple models can outperform Transformers on TC tasks, (3) similar classes of models exhibit similar level of sensitivity to text preprocessing

    Spatio-temporal log-Gaussian Cox processes on earthquake events

    Get PDF
    This work presents an application of spatio-temporal log-Gaussian Cox processes for the description of earthquake events. To explain the overall spatial trend, spatial geological information in the study area such as faults and volcanoes are introduced in the model. Moreover, an anisotropic specification of the covariance matrix of the Gaussian process is used to improve the explanation of the phenomenon. We apply and compare different models to explain the seismic events occurred in Alaska over the last decades

    Compact Three Dimensional Black Hole: Topology Change and Closed Timelike Curve (minor changes)

    Full text link
    We present a compactified version of the 3-dimensional black hole recently found by considering extra identifications and determine the analytical continuation of the solution beyond its coordinate singularity by extending the identifications to the extended region of the spacetime. In the extended region of the spacetime, we find a topology change and non-trivial closed timelike curves both in the ordinary 3-dimensional black hole and in the compactified one. Especially, in the case of the compactified 3-dimensional black hole, we show an example of topology change from one double torus to eight spheres with three punctures.Comment: 20 pages revtex.sty 8 figures contained, TIT/HEP-245/COSMO-4

    Stable Topologies of Event Horizon

    Get PDF
    In our previous work, it was shown that the topology of an event horizon (EH) is determined by the past endpoints of the EH. A torus EH (the collision of two EH) is caused by the two-dimensional (one-dimensional) set of the endpoints. In the present article, we examine the stability of the topology of the EH. We see that a simple case of a single spherical EH is unstable. Furthermore, in general, an EH with handles (a torus, a double torus, ...) is structurally stable in the sense of catastrophe theory.Comment: 21 pages, revtex, five figures containe

    Quantum Stability of (2+1)-Spacetimes with Non-Trivial Topology

    Get PDF
    Quantum fields are investigated in the (2+1)-open-universes with non-trivial topologies by the method of images. The universes are locally de Sitter spacetime and anti-de Sitter spacetime. In the present article we study spacetimes whose spatial topologies are a torus with a cusp and a sphere with three cusps as a step toward the more general case. A quantum energy momentum tensor is obtained by the point stripping method. Though the cusps are no singularities, the latter cusps cause the divergence of the quantum field. This suggests that only the latter cusps are quantum mechanically unstable. Of course at the singularity of the background spacetime the quantum field diverges. Also the possibility of the divergence of topological effect by a negative spatial curvature is discussed. Since the volume of the negatively curved space is larger than that of the flat space, one see so many images of a single source by the non-trivial topology. It is confirmed that this divergence does not appear in our models of topologies. The results will be applicable to the case of three dimensional multi black hole\cite{BR}.Comment: 17 pages, revtex, 3 uuencoded figures containe

    Detection of Hate Speech Spreaders using convolutional neural networks

    Get PDF
    In this paper we describe a deep learning model based on a Convolutional Neural Network (CNN). The model was developed for the Profiling Hate Speech Spreaders (HSSs) task proposed by PAN 2021 organizers and hosted at the 2021 CLEF Conference. Our approach to the task of classifying an author as HSS or not (nHSS) takes advantage of a CNN based on a single convolutional layer. In this binary classification task, on the tests performed using a 5-fold cross validation, the proposed model reaches a maximum accuracy of 0.80 on the multilingual (i.e., English and Spanish) training set, and a minimum loss value of 0.51 on the same set. As announced by the task organizers, the trained model presented is able to reach an overall accuracy of 0.79 on the full test set. This overall accuracy is obtained averaging the accuracy achieved by the model on both languages. In particular, with regard to the Spanish test set, the organizers announced that our model achieves an accuracy of 0.85, while on the English test set the same model achieved - as announced by the organizers too - an accuracy of 0.73. Thanks to the model presented in this paper, our team won the 2021 PAN competition on profiling HSSs
    • …
    corecore