11,417 research outputs found

    A hybrid medical text classification framework: integrating attentive rule construction and neural network

    Get PDF
    The main objective of this work is to improve the quality and transparency of the medical text classification solutions. Conventional text classification methods provide users with only a restricted mechanism (based on frequency) for selecting features. In this paper, a three-stage hybrid method combining the threshold-gated attentive bi-directional Long Short-Term Memory (ABLSTM) and the regular expression based classifier is proposed for medical text classification tasks. The bi-directional Long Short-Term Memory (LSTM) architecture with an attention layer allows the network to weigh words according to their perceived importance and focus on crucial parts of a sentence. Feature words (or keywords) extracted by ABLSTM model are utilized to guide the regular expression rule construction. Our proposed approach leverages the advantages of both the interpretability of rule-based algorithms and the computational power of deep learning approaches for a production-ready scenario. Experimental results on real-world medical online query data clearly validate the superiority of our system in selecting domain-specific and topic-related features. Results show that the proposed approach achieves an accuracy of 0.89 and an F1-score of 0.92 respectively. Furthermore, our experimentation also illustrates the versatility of regular expressions as a user-level tool for focusing on desired patterns and providing interpretable solutions for human modification

    On Cognitive Preferences and the Plausibility of Rule-based Models

    Get PDF
    It is conventional wisdom in machine learning and data mining that logical models such as rule sets are more interpretable than other models, and that among such rule-based models, simpler models are more interpretable than more complex ones. In this position paper, we question this latter assumption by focusing on one particular aspect of interpretability, namely the plausibility of models. Roughly speaking, we equate the plausibility of a model with the likeliness that a user accepts it as an explanation for a prediction. In particular, we argue that, all other things being equal, longer explanations may be more convincing than shorter ones, and that the predominant bias for shorter models, which is typically necessary for learning powerful discriminative models, may not be suitable when it comes to user acceptance of the learned models. To that end, we first recapitulate evidence for and against this postulate, and then report the results of an evaluation in a crowd-sourcing study based on about 3.000 judgments. The results do not reveal a strong preference for simple rules, whereas we can observe a weak preference for longer rules in some domains. We then relate these results to well-known cognitive biases such as the conjunction fallacy, the representative heuristic, or the recogition heuristic, and investigate their relation to rule length and plausibility.Comment: V4: Another rewrite of section on interpretability to clarify focus on plausibility and relation to interpretability, comprehensibility, and justifiabilit

    Scientific reference style using rule-based machine learning

    Get PDF
    Regular Expressions (RegEx) can be employed as a technique for supervised learning to define and search for specific patterns inside text. This work devised a method that utilizes regular expressions to convert the reference style of academic papers into several styles, dependent on the specific needs of the target publication or conference. Our research aimed to detect distinctive patterns of reference styles using RegEx and compare them with a dataset including various reference styles. We gathered a diverse range of reference format categories, encompassing seven distinct classes, from various sources such as academic papers, journals, conference proceedings, and books. Our approach involves employing RegEx to convert one referencing format to another based on the user's specific preferences. The proposed model demonstrated an accuracy of 57.26% for book references and 57.56% for journal references. We used the similarity ratio and Levenshtein distance to evaluate the dataset's performance. The model achieved a 97.8% similarity ratio with a Levenshtein distance of 2. Notably, the APA style for journal references yielded the best results. However, the effectiveness of the extraction function varies depending on the reference style. For APA style, the model showed a 99.97% similarity ratio with a Levenshtein distance of 1. Overall, our proposed model outperforms baseline machine learning models in this task. This study introduces an automated program that utilizes regular expressions to modify academic reference formats. This will enhance the efficiency, precision, and adaptability of academic publishing

    Towards Intelligent Databases

    Get PDF
    This article is a presentation of the objectives and techniques of deductive databases. The deductive approach to databases aims at extending with intensional definitions other database paradigms that describe applications extensionaUy. We first show how constructive specifications can be expressed with deduction rules, and how normative conditions can be defined using integrity constraints. We outline the principles of bottom-up and top-down query answering procedures and present the techniques used for integrity checking. We then argue that it is often desirable to manage with a database system not only database applications, but also specifications of system components. We present such meta-level specifications and discuss their advantages over conventional approaches

    Metaheuristic design of feedforward neural networks: a review of two decades of research

    Get PDF
    Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era

    Information Extraction Berbasis Rule untuk Laporan Keuangan Perusahaan

    Get PDF
    Laporan keuangan merupakan salah satu sumber informasi penting bagi investor dalam membuat keputusan investasi. Namun, data yang terkandung di dalam laporan keuangan sangat besar sehingga proses untuk mencari informasi penting didalamnya perlu waktu yang cukup lama. Hal ini memunculkan kebutuhan perangkat yang bisa mengidentifikasi dan mengekstrak informasi yang relevan dari laporan keuangan secara otomatis. Oleh karena itu, sebuah model information extraction berbasis rule diusulkan dalam tugas akhir ini. Model ekstraksi informasi yang diusulkan, dirancang supaya dapat mengautomasi proses ekstraksi informasi yang relevan bagi investor seperti posisi keuangan, laba rugi, dan arus kas perusahaan. Model ini menggunakan rule yang sudah ditentukan sebelumnya oleh peneliti untuk mengekstrak informasi tersebut. Salah satu rule yang digunakan adalah regular expression, yang biasa digunakan untuk mengambil informasi yang terkandung dalam teks. Proses evaluasi model akan dilakukan dengan mengukur tingkat accuracy dan f1-score informasi yang didapatkan dari berbagai laporan keuangan. Penelitian ini menghasilkan model information extraction yang mampu mendapatkan nilai accuracy keseluruhan sebesar 86,07% dalam mendapatkan 16 label informasi finansial yang berbeda. Model ini, diharapkan dapat membantu investor dalam pengambilan keputusan investasi dengan memberikan akses yang mudah dan cepat ke informasi yang terdapat dalam laporan keuangan tanpa mengurangi tingkat akurasi informasi yang didapatkan secara signifikan

    Ant Colony Optimization

    Get PDF
    Ant Colony Optimization (ACO) is the best example of how studies aimed at understanding and modeling the behavior of ants and other social insects can provide inspiration for the development of computational algorithms for the solution of difficult mathematical problems. Introduced by Marco Dorigo in his PhD thesis (1992) and initially applied to the travelling salesman problem, the ACO field has experienced a tremendous growth, standing today as an important nature-inspired stochastic metaheuristic for hard optimization problems. This book presents state-of-the-art ACO methods and is divided into two parts: (I) Techniques, which includes parallel implementations, and (II) Applications, where recent contributions of ACO to diverse fields, such as traffic congestion and control, structural optimization, manufacturing, and genomics are presented

    A Classification Approach for Automated Reasoning Systems--A Case Study in Graph Theory

    Get PDF
    Reasoning systems which create classifications of structured objects face the problem of how object descriptions can be used to reflect their components as well as relations among these components. Current reasoning systems on graph theory do not adequately provide models to discover complex relations among mathematical concepts (eg: relations involving subgraphs) mainly due to the inability to solve this problem. This thesis presents an approach to construct a knowledge-based system, GC (Graph Classification), which overcomes this difficulty in performing automated reasoning in graph theory. We describe graph concepts based on an attribute called Linear Recursive Constructivity (LRC). LRC defines classes by an algebraic formula supported by background knowledge of graph types. We use subsumption checking on decomposed algebraic expressions of graph classes as a major proof method. The search is guided by case-split-based inferencing. Using the approach GC has generated proofs for many theorems such as any two distinct cycles (closed paths) having a common edge e contain a cycle not traversing e , if cycle C1 contains edges e1, e2, and cycle C2 contains edges e2, e3, then there exists a cycle that contains e1 and e3 and the union of a tree and a path is a tree if they have only a single common vertex. The main contributions of this thesis are: (1) Development of a classification-based knowledge representation and a reasoning approach for graph concepts, thus providing a simple model for structured mathematical objects. (2) Development of an algebraic theory for simplifying and decomposing graph concepts. (3) Development of a proof search and a case-splitting technique with the guidance of graph type knowledge. (4) Development of a proving mechanism that can be generate constructive proofs by manipulating only simple linear formalization of theorems

    Improving the translation environment for professional translators

    Get PDF
    When using computer-aided translation systems in a typical, professional translation workflow, there are several stages at which there is room for improvement. The SCATE (Smart Computer-Aided Translation Environment) project investigated several of these aspects, both from a human-computer interaction point of view, as well as from a purely technological side. This paper describes the SCATE research with respect to improved fuzzy matching, parallel treebanks, the integration of translation memories with machine translation, quality estimation, terminology extraction from comparable texts, the use of speech recognition in the translation process, and human computer interaction and interface design for the professional translation environment. For each of these topics, we describe the experiments we performed and the conclusions drawn, providing an overview of the highlights of the entire SCATE project

    Creating Business Intelligence through Machine Learning: An Effective Business Decision Making Tool

    Get PDF
    Growing technological progressions have given rise to many issues concerning the contemporary decision making in business, which is a difficult phenomenon without Business Intelligence/ Machine Learning. The linking of machine learning with business intelligence is not only pivotal for business decision making but also for the business intelligence in totality, owing to the reason that in absence of machine learning, decision making couldn’t take place efficaciously. Machines need to learn, re-learn, and then only they can help your learning process. The below paper seeks to make this concept simple/ easy by removing the ambiguities using a general framework. In order to prove the impact of machine learning on business intelligence, we need to forecast the trends, what is going around the world – business has to stay updated, then only it can be a successful endeavour.  The paper posits the basic theories and definitions of business intelligence and machine learning. To learn from the past and forecast the future trends, many companies are adopting business intelligence tools and systems. Companies have understood the brilliance of enforcing achievements of the goals defined by their business strategies through business intelligence concepts and with the help of machine learning. It describes the insights on the role and requirement of real time BI by examining the business needs. Keywords: Business Intelligence (BI); Machine Learning (ML); Artificial Neural Networks (ANN); Self-Organizing Maps (SOM); Data Mining (DM); Data Warehousing (DW)
    corecore