685 research outputs found
FAKE NEWS DETECTION ON THE WEB: A DEEP LEARNING BASED APPROACH
The acceptance and popularity of social media platforms for the dispersion and proliferation of news articles have led to the spread of questionable and untrusted information (in part) due to the ease by which misleading content can be created and shared among the communities. While prior research has attempted to automatically classify news articles and tweets as credible and non-credible. This work complements such research by proposing an approach that utilizes the amalgamation of Natural Language Processing (NLP), and Deep Learning techniques such as Long Short-Term Memory (LSTM).
Moreover, in Information System’s paradigm, design science research methodology (DSRM) has become the major stream that focuses on building and evaluating an artifact to solve emerging problems. Hence, DSRM can accommodate deep learning-based models with the availability of adequate datasets. Two publicly available datasets that contain labeled news articles and tweets have been used to validate the proposed model’s effectiveness. This work presents two distinct experiments, and the results demonstrate that the proposed model works well for both long sequence news articles and short-sequence texts such as tweets. Finally, the findings suggest that the sentiments, tagging, linguistics, syntactic, and text embeddings are the features that have the potential to foster fake news detection through training the proposed model on various dimensionality to learn the contextual meaning of the news content
GISTIC2.0 facilitates sensitive and confident localization of the targets of focal somatic copy-number alteration in human cancers
We describe methods with enhanced power and specificity to identify genes targeted by somatic copy-number alterations (SCNAs) that drive cancer growth. By separating SCNA profiles into underlying arm-level and focal alterations, we improve the estimation of background rates for each category. We additionally describe a probabilistic method for defining the boundaries of selected-for SCNA regions with user-defined confidence. Here we detail this revised computational approach, GISTIC2.0, and validate its performance in real and simulated datasets
U-Net and its variants for medical image segmentation: theory and applications
U-net is an image segmentation technique developed primarily for medical
image analysis that can precisely segment images using a scarce amount of
training data. These traits provide U-net with a very high utility within the
medical imaging community and have resulted in extensive adoption of U-net as
the primary tool for segmentation tasks in medical imaging. The success of
U-net is evident in its widespread use in all major image modalities from CT
scans and MRI to X-rays and microscopy. Furthermore, while U-net is largely a
segmentation tool, there have been instances of the use of U-net in other
applications. As the potential of U-net is still increasing, in this review we
look at the various developments that have been made in the U-net architecture
and provide observations on recent trends. We examine the various innovations
that have been made in deep learning and discuss how these tools facilitate
U-net. Furthermore, we look at image modalities and application areas where
U-net has been applied.Comment: 42 pages, in IEEE Acces
Gene expression programming for Efficient Time-series Financial Forecasting
Stock market prediction is of immense interest to trading companies and buyers due to
high profit margins. The majority of successful buying or selling activities occur close
to stock price turning trends. This makes the prediction of stock indices and analysis a
crucial factor in the determination that whether the stocks will increase or decrease the
next day. Additionally, precise prediction of the measure of increase or decrease of
stock prices also plays an important role in buying/selling activities. This research
presents two core aspects of stock-market prediction. Firstly, it presents a Networkbased
Fuzzy Inference System (ANFIS) methodology to integrate the capabilities of
neural networks with that of fuzzy logic. A specialised extension to this technique is
known as the genetic programming (GP) and gene expression programming (GEP) to
explore and investigate the outcome of the GEP criteria on the stock market price
prediction.
The research presented in this thesis aims at the modelling and prediction of short-tomedium
term stock value fluctuations in the market via genetically tuned stock market
parameters. The technique uses hierarchically defined GP and gene-expressionprogramming
(GEP) techniques to tune algebraic functions representing the fittest
equation for stock market activities. The technology achieves novelty by proposing a
fractional adaptive mutation rate Elitism (GEP-FAMR) technique to initiate a balance
between varied mutation rates between varied-fitness chromosomes thereby improving
prediction accuracy and fitness improvement rate. The methodology is evaluated
against five stock market companies with each having its own trading circumstances
during the past 20+ years. The proposed GEP/GP methodologies were evaluated based
on variable window/population sizes, selection methods, and Elitism, Rank and Roulette
selection methods. The Elitism-based approach showed promising results with a low
error-rate in the resultant pattern matching with an overall accuracy of 95.96% for
short-term 5-day and 95.35% for medium-term 56-day trading periods. The
contribution of this research to theory is that it presented a novel evolutionary
methodology with modified selection operators for the prediction of stock exchange
data via Gene expression programming. The methodology dynamically adapts the
mutation rate of different fitness groups in each generation to ensure a diversification
II
balance between high and low fitness solutions. The GEP-FAMR approach was
preferred to Neural and Fuzzy approaches because it can address well-reported
problems of over-fitting, algorithmic black-boxing, and data-snooping issues via GP
and GEP algorithmsSaudi Cultural Burea
Mathematics in Software Reliability and Quality Assurance
This monograph concerns the mathematical aspects of software reliability and quality assurance and consists of 11 technical papers in this emerging area. Included are the latest research results related to formal methods and design, automatic software testing, software verification and validation, coalgebra theory, automata theory, hybrid system and software reliability modeling and assessment
Advances in Evolutionary Algorithms
With the recent trends towards massive data sets and significant computational power, combined with evolutionary algorithmic advances evolutionary computation is becoming much more relevant to practice. Aim of the book is to present recent improvements, innovative ideas and concepts in a part of a huge EA field
Introducing deep learning -based methods into the variant calling analysis pipeline
Biological interpretation of the genetic variation enhances our understanding of normal and pathological phenotypes, and may lead to the development of new therapeutics.
However, it is heavily dependent on the genomic data analysis, which might be inaccurate due to the various sequencing errors and inconsistencies caused by these errors. Modern analysis pipelines already utilize heuristic and statistical techniques, but the rate of falsely identified mutations remains high and variable, particular sequencing technology, settings and variant type.
Recently, several tools based on deep neural networks have been published. The neural networks are supposed to find motifs in the data that were not previously seen.
The performance of these novel tools is assessed in terms of precision and recall, as well as computational efficiency. Following the established best practices in both variant detection and benchmarking, the discussed tools demonstrate accuracy metrics and computational efficiency that spur further discussion
- …