136 research outputs found
Multi-model Fusion Attention Network for News Text Classification
At present, the classification prediction task based on news content or news headline has the problems of inaccurate classification and attention deviation. In this paper, a multi-model fusion attention network for news text classification (MFAN) is proposed to train news content and news titles in parallel. Firstly, the multi-head attention mechanism is used to obtain the category information of news content through a dynamic word vector, focusing on the semantic information that significantly influences the downstream classification task. Secondly, the semantic information of news headlines is obtained by using the improved version of the long-short-term memory network, and the attention is focused on the words that have a great influence on the final results, which improves the effectiveness of model classification. Finally, the classification fusion module fuses the probability scores of news text and news headlines in proportion to improve the accuracy of text classification. The experimental test on the Tenth China Software cup dataset shows that the F1 - Score index of the MFAN model reaches 97.789 %. The experimental results show that the MFAN model can effectively and accurately predict the classification of news texts
DXVNet-ViT-Huge (JFT) Multimode Classification Network Based on Vision Transformer
Aiming at the problem that traditional CNN network is not good at extracting global features of images, Based on DXVNet network, Conditional Random Fields (CRF) component and pre-trained ViT-Huge (Vision Transformer) are adopted in this paper Transformer model expands and builds a brand new DXVNet-ViT-Huge (JFT) network. CRF component can help the network learn the constraint conditions of each word corresponding prediction label, improve the D-GRU method based word label prediction errors, and improve the accuracy of sequence annotation. The Transformer architecture of the ViT (Huge) model can extract the global feature information of the image, while CNN is better at extracting the local features of the image. Therefore, the ViT (Huge) Huge pre-training model and CNN pre-training model adopt the multi-modal feature fusion technology. Two complementary image feature information is fused by Bi-GRU to improve the performance of network classification. The experimental results show that the newly constructed Dxvnet-Vit-Huge (JFT) model achieves good performance, and the F1 values in the two real public data sets are 6.03% and 7.11% higher than the original DXVNet model, respectively
Emotion Analysis of Ideological and Political Education Using a GRU Deep Neural Network
Theoretical research into the emotional attributes of ideological and political education can improve our ability to understand human emotion and solve socio-emotional problems. To that end, this study undertook an analysis of emotion in ideological and political education by integrating a gate recurrent unit (GRU) with an attention mechanism. Based on the good results achieved by BERT in the downstream network, we use the long focusing attention mechanism assisted by two-way GRU to extract relevant information and global information of ideological and political education and emotion analysis, respectively. The two kinds of information complement each other, and the accuracy of emotion information can be further improved by combining neural network model. Finally, the validity and domain adaptability of the model were verified using several publicly available, fine-grained emotion datasets
ODTC: An online darknet traffic classification model based on multimodal self-attention chaotic mapping features
Darknet traffic classification is significantly important to network management and security. To achieve fast and accurate classification performance, this paper proposes an online classification model based on multimodal self-attention chaotic mapping features. On the one hand, the payload content of the packet is input into the network integrating CNN and BiGRU to extract local space-time features. On the other hand, the flow level abstract features processed by the MLP are introduced. To make up for the lack of the indistinct feature learning, a feature amplification module that uses logistic chaotic mapping to amplify fuzzy features is introduced. In addition, a multi-head attention mechanism is used to excavate the hidden relationships between different features. Besides, to better support new traffic classes, a class incremental learning model is developed with the weighted loss function to achieve continuous learning with reduced network parameters. The experimental results on the public CICDarketSec2020 dataset show that the accuracy of the proposed model is improved in multiple categories; however, the time and memory consumption is reduced by about 50. Compared with the existing state-of-the-art traffic classification models, the proposed model has better classification performance
Optimized Ensemble Approach for Multi-model Event Detection in Big data
Event detection acts an important role among modern society and it is a popular computer process that permits to detect the events automatically. Big data is more useful for the event detection due to large size of data. Multimodal event detection is utilized for the detection of events using heterogeneous types of data. This work aims to perform for classification of diverse events using Optimized Ensemble learning approach. The Multi-modal event data including text, image and audio are sent to the user devices from cloud or server where three models are generated for processing audio, text and image. At first, the text, image and audio data is processed separately. The process of creating a text model includes pre-processing using Imputation of missing values and data normalization. Then the textual feature extraction using integrated N-gram approach. The Generation of text model using Convolutional two directional LSTM (2DCon_LSTM). The steps involved in image model generation are pre-processing using Min-Max Gaussian filtering (MMGF). Image feature extraction using VGG-16 network model and generation of image model using Tweaked auto encoder (TAE) model. The steps involved in audio model generation are pre-processing using Discrete wavelet transform (DWT). Then the audio feature extraction using Hilbert Huang transform (HHT) and Generation of audio model using Attention based convolutional capsule network (Attn_CCNet). The features obtained by the generated models of text, image and audio are fused together by feature ensemble approach. From the fused feature vector, the optimal features are trained through improved battle royal optimization (IBRO) algorithm. A deep learning model called Convolutional duo Gated recurrent unit with auto encoder (C-Duo GRU_AE) is used as a classifier. Finally, different types of events are classified where the global model are then sent to the user devices with high security and offers better decision making process. The proposed methodology achieves better performances are Accuracy (99.93%), F1-score (99.91%), precision (99.93%), Recall (99.93%), processing time (17seconds) and training time (0.05seconds). Performance analysis exceeds several comparable methodologies in precision, recall, accuracy, F1 score, training time, and processing time. This designates that the proposed methodology achieves improved performance than the compared schemes. In addition, the proposed scheme detects the multi-modal events accurately
Social media bot detection with deep learning methods: a systematic review
Social bots are automated social media accounts governed by software and controlled by humans at the backend. Some bots have good purposes, such as automatically posting information about news and even to provide help during emergencies. Nevertheless, bots have also been used for malicious purposes, such as for posting fake news or rumour spreading or manipulating political campaigns. There are existing mechanisms that allow for detection and removal of malicious bots automatically. However, the bot landscape changes as the bot creators use more sophisticated methods to avoid being detected. Therefore, new mechanisms for discerning between legitimate and bot accounts are much needed. Over the past few years, a few review studies contributed to the social media bot detection research by presenting a comprehensive survey on various detection methods including cutting-edge solutions like machine learning (ML)/deep learning (DL) techniques. This paper, to the best of our knowledge, is the first one to only highlight the DL techniques and compare the motivation/effectiveness of these techniques among themselves and over other methods, especially the traditional ML ones. We present here a refined taxonomy of the features used in DL studies and details about the associated pre-processing strategies required to make suitable training data for a DL model. We summarize the gaps addressed by the review papers that mentioned about DL/ML studies to provide future directions in this field. Overall, DL techniques turn out to be computation and time efficient techniques for social bot detection with better or compatible performance as traditional ML techniques
HyMo: Vulnerability Detection in Smart Contracts using a Novel Multi-Modal Hybrid Model
With blockchain technology rapidly progress, the smart contracts have become
a common tool in a number of industries including finance, healthcare,
insurance and gaming. The number of smart contracts has multiplied, and at the
same time, the security of smart contracts has drawn considerable attention due
to the monetary losses brought on by smart contract vulnerabilities. Existing
analysis techniques are capable of identifying a large number of smart contract
security flaws, but they rely too much on rigid criteria established by
specialists, where the detection process takes much longer as the complexity of
the smart contract rises. In this paper, we propose HyMo as a multi-modal
hybrid deep learning model, which intelligently considers various input
representations to consider multimodality and FastText word embedding
technique, which represents each word as an n-gram of characters with BiGRU
deep learning technique, as a sequence processing model that consists of two
GRUs to achieve higher accuracy in smart contract vulnerability detection. The
model gathers features using various deep learning models to identify the smart
contract vulnerabilities. Through a series of studies on the currently publicly
accessible dataset such as ScrawlD, we show that our hybrid HyMo model has
excellent smart contract vulnerability detection performance. Therefore, HyMo
performs better detection of smart contract vulnerabilities against other
approaches
- …