14 research outputs found
Multi-model Fusion Attention Network for News Text Classification
At present, the classification prediction task based on news content or news headline has the problems of inaccurate classification and attention deviation. In this paper, a multi-model fusion attention network for news text classification (MFAN) is proposed to train news content and news titles in parallel. Firstly, the multi-head attention mechanism is used to obtain the category information of news content through a dynamic word vector, focusing on the semantic information that significantly influences the downstream classification task. Secondly, the semantic information of news headlines is obtained by using the improved version of the long-short-term memory network, and the attention is focused on the words that have a great influence on the final results, which improves the effectiveness of model classification. Finally, the classification fusion module fuses the probability scores of news text and news headlines in proportion to improve the accuracy of text classification. The experimental test on the Tenth China Software cup dataset shows that the F1 - Score index of the MFAN model reaches 97.789 %. The experimental results show that the MFAN model can effectively and accurately predict the classification of news texts
Presenting an approach based on weighted CapsuleNet networks for Arabic and Persian multi-domain sentiment analysis
Sentiment classification is a fundamental task in natural language
processing, assigning one of the three classes, positive, negative, or neutral,
to free texts. However, sentiment classification models are highly domain
dependent; the classifier may perform classification with reasonable accuracy
in one domain but not in another due to the Semantic multiplicity of words
getting poor accuracy. This article presents a new Persian/Arabic multi-domain
sentiment analysis method using the cumulative weighted capsule networks
approach. Weighted capsule ensemble consists of training separate capsule
networks for each domain and a weighting measure called domain belonging degree
(DBD). This criterion consists of TF and IDF, which calculates the dependency
of each document for each domain separately; this value is multiplied by the
possible output that each capsule creates. In the end, the sum of these
multiplications is the title of the final output, and is used to determine the
polarity. And the most dependent domain is considered the final output for each
domain. The proposed method was evaluated using the Digikala dataset and
obtained acceptable accuracy compared to the existing approaches. It achieved
an accuracy of 0.89 on detecting the domain of belonging and 0.99 on detecting
the polarity. Also, for the problem of dealing with unbalanced classes, a
cost-sensitive function was used. This function was able to achieve 0.0162
improvements in accuracy for sentiment classification. This approach on Amazon
Arabic data can achieve 0.9695 accuracies in domain classification
Accelerated materials language processing enabled by GPT
Materials language processing (MLP) is one of the key facilitators of
materials science research, as it enables the extraction of structured
information from massive materials science literature. Prior works suggested
high-performance MLP models for text classification, named entity recognition
(NER), and extractive question answering (QA), which require complex model
architecture, exhaustive fine-tuning and a large number of human-labelled
datasets. In this study, we develop generative pretrained transformer
(GPT)-enabled pipelines where the complex architectures of prior MLP models are
replaced with strategic designs of prompt engineering. First, we develop a
GPT-enabled document classification method for screening relevant documents,
achieving comparable accuracy and reliability compared to prior models, with
only small dataset. Secondly, for NER task, we design an entity-centric
prompts, and learning few-shot of them improved the performance on most of
entities in three open datasets. Finally, we develop an GPT-enabled extractive
QA model, which provides improved performance and shows the possibility of
automatically correcting annotations. While our findings confirm the potential
of GPT-enabled MLP models as well as their value in terms of reliability and
practicability, our scientific methods and systematic approach are applicable
to any materials science domain to accelerate the information extraction of
scientific literature
Multimodal cyberbullying detection using capsule network with dynamic routing and deep convolutional neural network
Cyberbullying is the use of information technology networks by individuals’ to humiliate, tease, embarrass, taunt, defame and disparage a target without any face-to-face contact. Social media is the 'virtual playground' used by bullies with the upsurge of social networking sites such as Facebook, Instagram, YouTube and Twitter. It is critical to implement models and systems for automatic detection and resolution of bullying content available online as the ramifications can lead to a societal epidemic. This paper presents a deep neural model for cyberbullying detection in three different modalities of social data, namely textual, visual and info-graphic (text embedded along with an image). The all-in-one architecture, CapsNet–ConvNet, consists of a capsule network (CapsNet) deep neural network with dynamic routing for predicting the textual bullying content and a convolution neural network (ConvNet) for predicting the visual bullying content. The info-graphic content is discretized by separating text from the image using Google Lens of Google Photos app. The perceptron-based decision-level late fusion strategy for multimodal learning is used to dynamically combine the predictions of discrete modalities and output the final category as bullying or non-bullying type. Experimental evaluation is done on a mix-modal dataset which contains 10,000 comments and posts scrapped from YouTube, Instagram and Twitter. The proposed model achieves a superlative performance with the AUC–ROC of 0.98
A Review of Uncertain Decision-Making Methods in Energy Management Using Text Mining and Data Analytics
The managerial and environmental studies conducted in the energy research area reflect its substantial importance, particularly when optimizing and modifying consumption patterns, transitioning to renewable sources away from fossil ones, and designing plans and systems. The aim of this study is to provide a systematic review of the literature allowing us to identify which research subjects have been prioritized in the fields of energy and sustainability in recent years, determine the potential reasons explaining these trends, and categorize the techniques applied to analyze the uncertainty faced by decision-makers. We review articles published in highly ranked journals through the period 2003–2020 and apply text analytics to cluster their main characteristics; that is, we rely on pre-processing and text mining techniques. We analyze the title, abstract, keywords, and research methodology of the articles through clustering and topic modeling and illustrate what methods and fields constitute the main focus of researchers. We demonstrate the substantial importance of fuzzy-related methods and decision-making techniques such as the Analytical Hierarchy Process and Technique for Order Preferences by Similarity to Ideal Solutions (TOPSIS). We also show that subjects such as renewable energy, energy planning, sustainable energy, energy policy, and wind energy have gained relevance among researchers in recent years
Site-Specific Rules Extraction in Precision Agriculture
El incremento sostenible en la producción alimentaria para satisfacer las necesidades de una población mundial en aumento es un verdadero reto cuando tenemos en cuenta el impacto constante de plagas y enfermedades en los cultivos. Debido a las importantes pérdidas económicas que se producen, el uso de tratamientos químicos es demasiado alto; causando contaminación del medio ambiente y resistencia a distintos tratamientos. En este contexto, la comunidad agrícola divisa la aplicación de tratamientos más específicos para cada lugar, así como la validación automática con la conformidad legal. Sin embargo, la especificación de estos tratamientos se encuentra en regulaciones expresadas en lenguaje natural. Por este motivo, traducir regulaciones a una representación procesable por máquinas está tomando cada vez más importancia en la agricultura de precisión.Actualmente, los requisitos para traducir las regulaciones en reglas formales están lejos de ser cumplidos; y con el rápido desarrollo de la ciencia agrícola, la verificación manual de la conformidad legal se torna inabordable.En esta tesis, el objetivo es construir y evaluar un sistema de extracción de reglas para destilar de manera efectiva la información relevante de las regulaciones y transformar las reglas de lenguaje natural a un formato estructurado que pueda ser procesado por máquinas. Para ello, hemos separado la extracción de reglas en dos pasos. El primero es construir una ontología del dominio; un modelo para describir los desórdenes que producen las enfermedades en los cultivos y sus tratamientos. El segundo paso es extraer información para poblar la ontología. Puesto que usamos técnicas de aprendizaje automático, implementamos la metodología MATTER para realizar el proceso de anotación de regulaciones. Una vez creado el corpus, construimos un clasificador de categorías de reglas que discierne entre obligaciones y prohibiciones; y un sistema para la extracción de restricciones en reglas, que reconoce información relevante para retener el isomorfismo con la regulación original. Para estos componentes, empleamos, entre otra técnicas de aprendizaje profundo, redes neuronales convolucionales y “Long Short- Term Memory”. Además, utilizamos como baselines algoritmos más tradicionales como “support-vector machines” y “random forests”.Como resultado, presentamos la ontología PCT-O, que ha sido alineada con otras ontologías como NCBI, PubChem, ChEBI y Wikipedia. El modelo puede ser utilizado para la identificación de desórdenes, el análisis de conflictos entre tratamientos y la comparación entre legislaciones de distintos países. Con respecto a los sistemas de extracción, evaluamos empíricamente el comportamiento con distintas métricas, pero la métrica F1 es utilizada para seleccionar los mejores sistemas. En el caso del clasificador de categorías de reglas, el mejor sistema obtiene un macro F1 de 92,77% y un F1 binario de 85,71%. Este sistema usa una red “bidirectional long short-term memory” con “word embeddings” como entrada. En relación al extractor de restricciones de reglas, el mejor sistema obtiene un micro F1 de 88,3%. Este extractor utiliza como entrada una combinación de “character embeddings” junto a “word embeddings” y una red neuronal “bidirectional long short-term memory”.<br /