72 research outputs found
Comparing Transformer-based NER approaches for analysing textual medical diagnoses
The automated analysis of medical documents has grown in research interest in recent years as a consequence of the social relevance of the thematic and the difficulties often encountered with short and
very specific documents. In particular, this fervent area of research has stimulated the development of
several techniques of automatic document classification, question answering, and name entity recognition (NER). Nevertheless, many open issues must be addressed to obtain results that are satisfactory for
a field in which the effectiveness of predictions is a fundamental factor in order not to make mistakes
that could compromise people’s lives. To this end, we focused on the name entity recognition task from
medical documents and, in this work, we will discuss the results we obtained by our hybrid approach.
In order to take advantage of the most relevant findings in the field of natural language processing, we
decided to focus on deep neural network models. We compared several configurations of our model by
varying the transformer architecture, such as BERT, RoBERTa and ELECTRA, until we obtained a configuration that we considered the best for our goals. The most promising model was used to participate
in the SpRadIE task of the annual CLEF (Conference and Labs of the Evaluation Forum). The obtained
results are encouraging and can be of reference for future studies on the topic
Preface to the Sixth Workshop on Natural Language for Artificial Intelligence (NL4AI)
Natural Language Processing (NLP) is an important research topic in Artificial Intelligence (AI), as it is the target of different scientific and industrial interests. Natural Language is at the crossroad of Learning, Knowledge Representation, and Cognitive Modeling. Several recent AI achievements have repeatedly shown their beneficial impact on complex inference tasks, with huge application perspectives in linguistic modeling, processing, and inferences. However, Natural Language Understanding is still a rich research topic, whose cross-fertilization spans
a number of independent areas such as Cognitive Computing, Robotics as well as HumanComputer Interaction. For AI, Natural Languages are the research focus of paradigms and applications but, at the same time, they act as cornerstones of automation, autonomy, and learnability for most intelligent phenomena ranging from Vision to Planning and Social Behaviors. A reflection about such diverse and promising interactions is an important target for current AI studies, fully in the core mission of AI*IA. This workshop, supported by the Special Interest Group on NLP of AI*IA1 and by the Italian Association of Computational Linguistics
(AILC)2, aims at providing a broad overview of recent activities in the eld of Human Language Technologies (HLT) in Italy. In this context, the organization of NL4AI 2021 [1] provided researchers with the opportunity to share experiences and insights about AI applications focused on NLP in several domains. The 2022 edition of NL4AI is co-located with the 21th International Conference of the Italian Association for Artificial Intelligence (AIxIA 2022), taking place on November 30th in Udine, Italy. The program of the meeting is available on the official workshop
website3. We received 17 submissions, 14 of which were accepted after peer-review
GM-CTSC at SemEval-2020 Task 1: Gaussian Mixtures Cross Temporal Similarity Clustering
This paper describes the system proposed for the SemEval-2020 Task 1:
Unsupervised Lexical Semantic Change Detection. We focused our approach on the
detection problem. Given the semantics of words captured by temporal word
embeddings in different time periods, we investigate the use of unsupervised
methods to detect when the target word has gained or loosed senses. To this
end, we defined a new algorithm based on Gaussian Mixture Models to cluster the
target similarities computed over the two periods. We compared the proposed
approach with a number of similarity-based thresholds. We found that, although
the performance of the detection methods varies across the word embedding
algorithms, the combination of Gaussian Mixture with Temporal Referencing
resulted in our best system
Covid19/IT the digital side of Covid19: A picture from Italy with clustering and taxonomy
The Covid19 pandemic has significantly impacted on our lives, triggering a strong reaction resulting in vaccines, more effective diagnoses and therapies, policies to contain the pandemic outbreak, to name but a few. A significant contribution to their success comes from the computer science and information technology communities, both in support to other disciplines and as the primary driver of solutions for, e.g., diagnostics, social distancing, and contact tracing. In this work, we surveyed the Italian computer science and engineering community initiatives against the Covid19 pandemic. The 128 responses thus collected document the response of such a community during the first pandemic wave in Italy (February-May 2020), through several initiatives carried out by both single researchers and research groups able to promptly react to Covid19, even remotely. The data obtained by the survey are here reported, discussed and further investigated by Natural Language Processing techniques, to generate semantic clusters based on embedding representations of the surveyed activity descriptions. The resulting clusters have been then used to extend an existing Covid19 taxonomy with the classification of related research activities in computer science and information technology areas, summarizing this work contribution through a reproducible survey-to-taxonomy methodology
LLaMAntino: LLaMA 2 Models for Effective Text Generation in Italian Language
Large Language Models represent state-of-the-art linguistic models designed
to equip computers with the ability to comprehend natural language. With its
exceptional capacity to capture complex contextual relationships, the LLaMA
(Large Language Model Meta AI) family represents a novel advancement in the
field of natural language processing by releasing foundational models designed
to improve the natural language understanding abilities of the transformer
architecture thanks to their large amount of trainable parameters (7, 13, and
70 billion parameters). In many natural language understanding tasks, these
models obtain the same performances as private company models such as OpenAI
Chat-GPT with the advantage to make publicly available weights and code for
research and commercial uses. In this work, we investigate the possibility of
Language Adaptation for LLaMA models, explicitly focusing on addressing the
challenge of Italian Language coverage. Adopting an open science approach, we
explore various tuning approaches to ensure a high-quality text generated in
Italian suitable for common tasks in this underrepresented language in the
original models' datasets. We aim to release effective text generation models
with strong linguistic properties for many tasks that seem challenging using
multilingual or general-purpose LLMs. By leveraging an open science philosophy,
this study contributes to Language Adaptation strategies for the Italian
language by introducing the novel LLaMAntino family of Italian LLMs
Semantically-Aware Retrieval of Oceanographic Phenomena Annotated on Satellite Images
Scientists in the marine domain process satellite images in order to extract information
that can be used for monitoring, understanding, and forecasting of marine phenomena, such as
turbidity, algal blooms and oil spills. The growing need for effective retrieval of related information
has motivated the adoption of semantically aware strategies on satellite images with different spatiotemporal and spectral characteristics. A big issue of these approaches is the lack of coincidence
between the information that can be extracted from the visual data and the interpretation that the
same data have for a user in a given situation. In this work, we bridge this semantic gap by connecting
the quantitative elements of the Earth Observation satellite images with the qualitative information,
modelling this knowledge in a marine phenomena ontology and developing a question answering
mechanism based on natural language that enables the retrieval of the most appropriate data for each
user’s needs. The main objective of the presented methodology is to realize the content-based search
of Earth Observation images related to the marine application domain on an application-specific
basis that can answer queries such as “Find oil spills that occurred this year in the Adriatic Sea”
AlBERTo: Modeling Italian Social Media Language with BERT
Natural Language Processing tasks recently achieved considerable interest and progresses following the development of numerous innovative artificial intelligence models released in recent years. The increase in available computing power has made possible the application of machine learning approaches on a considerable amount of textual data, demonstrating how they can obtain very encouraging results in challenging NLP tasks by generalizing the properties of natural language directly from the data. Models such as ELMo, GPT/GPT-2, BERT, ERNIE, and RoBERTa have proved to be extremely useful in NLP tasks such as entailment, sentiment analysis, and question answering. The availability of these resources mainly in the English language motivated us towards the realization of AlBERTo, a natural language model based on BERT and trained on the Italian language. We decided to train AlBERTo from scratch on social network language, Twitter in particular, because many of the classic tasks of content analysis are oriented to data extracted from the digital sphere of users. The model was distributed to the community through a repository on GitHub and the Transformers library (Wolf et al. 2019) released by the development group huggingface.co. We have evaluated the validity of the model on the classification tasks of sentiment polarity, irony, subjectivity, and hate speech. The specifications of the model, the code developed for training and fine-tuning, and the instructions for using it in a research project are freely available
IT-Covid19-IT: la risposta della comunità informatica italiana alla pandemia
La pandemia Covid19 ha avuto un forte impatto sulle nostre vite, anche da accademici. Ne è scaturita una reazione veemente della comunità scientifica i cui risultati sono sotto gli occhi di tutti: vaccini, terapie più puntuali ed efficaci, politiche di contenimento mirate, etc. A tutto ciò, l’informatica ha contribuito in maniera determinante, spesso con funzioni di supporto e servizio alle altre discipline, talvolta in primo piano con applicazioni specifiche, per esempio, per il distanziamento sociale ed il tracciamento dei contatti. Questo articolo prova a fare una fotografia della reazione della comunità informatica italiana alla pandemia Covid19, elaborando i dati ottenuti da un censimento condotto nel maggio 2020, a seguito della prima ondata, dalla Task Force Covid19-IT istituita allo scopo dal CINI (Consorzio Interuniversitario Nazionale per l’Informatica). I dati ottenuti dalle 131 proposte censite raccontano di una risposta decisa ed articolata della comunità, nata spontaneamente da centinaia di iniziative autonome distribuite su tutto il territorio nazionale e che deve proseguire, magari evolvendo in forme più organizzate
- …