71 research outputs found
Profiling Cryptocurrency Influencers with Few-Shot Learning Using Data Augmentation and ELECTRA
With this work we propose an application of the ELECTRA Transformer, fine-tuned on two augmented version of the same training dataset. Our team developed the novel framework for taking part at the Profiling Cryptocurrency Influencers with Few-shot Learning task hosted at PAN@CLEF2023. Our proposed strategy consists of an early data augmentation stage followed by a fine-tuning of ELECTRA. At the first stage we augment the original training dataset provided by the organizers using backtranslation. Using this augmented version of the training dataset, we perform a fine tuning of ELECTRA. Finally, using the fine-tuned version of ELECTRA, we inference the labels of the samples provided in the test set. To develop and test our model we used a two-ways validation on the training set. Firstly, we evaluate all the metrics on the augmented training set, and then we evaluate on the original training set. The metrics we considered span from accuracy to Macro F1, to Micro F1, to Recall and Precision. According to the official evaluator, our best submission reached a Macro F1 value equal to 0.3762
Is text preprocessing still worth the time? A comparative survey on the influence of popular preprocessing methods on Transformers and traditional classifiers
With the advent of the modern pre-trained Transformers, the text preprocessing has started to be neglected and not specifically addressed in recent NLP literature. However, both from a linguistic and from a computer science point of view, we believe that even when using modern Transformers, text preprocessing can significantly impact on the performance of a classification model. We want to investigate and compare, through this study, how preprocessing impacts on the Text Classification (TC) performance of modern and traditional classification models. We report and discuss the preprocessing techniques found in the literature and their most recent variants or applications to address TC tasks in different domains. In order to assess how much the preprocessing affects classification performance, we apply the three top referenced preprocessing techniques (alone or in combination) to four publicly available datasets from different domains. Then, nine machine learning models – including modern Transformers – get the preprocessed text as input. The results presented show that an educated choice on the text preprocessing strategy to employ should be based on the task as well as on the model considered. Outcomes in this survey show that choosing the best preprocessing technique – in place of the worst – can significantly improve accuracy on the classification (up to 25%, as in the case of an XLNet on the IMDB dataset). In some cases, by means of a suitable preprocessing strategy, even a simple Naïve Bayes classifier proved to outperform (i.e., by 2% in accuracy) the best performing Transformer. We found that Transformers and traditional models exhibit a higher impact of the preprocessing on the TC performance. Our main findings are: (1) also on modern pre-trained language models, preprocessing can affect performance, depending on the datasets and on the preprocessing technique or combination of techniques used, (2) in some cases, using a proper preprocessing strategy, simple models can outperform Transformers on TC tasks, (3) similar classes of models exhibit similar level of sensitivity to text preprocessing
Spatio-temporal log-Gaussian Cox processes on earthquake events
This work presents an application of spatio-temporal log-Gaussian Cox processes for the
description of earthquake events. To explain the overall spatial trend, spatial geological information
in the study area such as faults and volcanoes are introduced in the model. Moreover, an anisotropic
specification of the covariance matrix of the Gaussian process is used to improve the explanation of the
phenomenon. We apply and compare different models to explain the seismic events occurred in Alaska
over the last decades
Compact Three Dimensional Black Hole: Topology Change and Closed Timelike Curve (minor changes)
We present a compactified version of the 3-dimensional black hole recently
found by considering extra identifications and determine the analytical
continuation of the solution beyond its coordinate singularity by extending the
identifications to the extended region of the spacetime. In the extended region
of the spacetime, we find a topology change and non-trivial closed timelike
curves both in the ordinary 3-dimensional black hole and in the compactified
one. Especially, in the case of the compactified 3-dimensional black hole, we
show an example of topology change from one double torus to eight spheres with
three punctures.Comment: 20 pages revtex.sty 8 figures contained, TIT/HEP-245/COSMO-4
Stable Topologies of Event Horizon
In our previous work, it was shown that the topology of an event horizon (EH)
is determined by the past endpoints of the EH. A torus EH (the collision of two
EH) is caused by the two-dimensional (one-dimensional) set of the endpoints. In
the present article, we examine the stability of the topology of the EH. We see
that a simple case of a single spherical EH is unstable. Furthermore, in
general, an EH with handles (a torus, a double torus, ...) is structurally
stable in the sense of catastrophe theory.Comment: 21 pages, revtex, five figures containe
Quantum Stability of (2+1)-Spacetimes with Non-Trivial Topology
Quantum fields are investigated in the (2+1)-open-universes with non-trivial
topologies by the method of images. The universes are locally de Sitter
spacetime and anti-de Sitter spacetime. In the present article we study
spacetimes whose spatial topologies are a torus with a cusp and a sphere with
three cusps as a step toward the more general case. A quantum energy momentum
tensor is obtained by the point stripping method. Though the cusps are no
singularities, the latter cusps cause the divergence of the quantum field. This
suggests that only the latter cusps are quantum mechanically unstable. Of
course at the singularity of the background spacetime the quantum field
diverges. Also the possibility of the divergence of topological effect by a
negative spatial curvature is discussed. Since the volume of the negatively
curved space is larger than that of the flat space, one see so many images of a
single source by the non-trivial topology. It is confirmed that this divergence
does not appear in our models of topologies. The results will be applicable to
the case of three dimensional multi black hole\cite{BR}.Comment: 17 pages, revtex, 3 uuencoded figures containe
Detection of Hate Speech Spreaders using convolutional neural networks
In this paper we describe a deep learning model based on a Convolutional Neural Network (CNN). The model was developed for the Profiling Hate Speech Spreaders (HSSs) task proposed by PAN 2021 organizers and hosted at the 2021 CLEF Conference. Our approach to the task of classifying an author as HSS or not (nHSS) takes advantage of a CNN based on a single convolutional layer. In this binary classification task, on the tests performed using a 5-fold cross validation, the proposed model reaches a maximum accuracy of 0.80 on the multilingual (i.e., English and Spanish) training set, and a minimum loss value of 0.51 on the same set. As announced by the task organizers, the trained model presented is able to reach an overall accuracy of 0.79 on the full test set. This overall accuracy is obtained averaging the accuracy achieved by the model on both languages. In particular, with regard to the Spanish test set, the organizers announced that our model achieves an accuracy of 0.85, while on the English test set the same model achieved - as announced by the organizers too - an accuracy of 0.73. Thanks to the model presented in this paper, our team won the 2021 PAN competition on profiling HSSs
- …