7 research outputs found

    Klasifikasi Tulisan Tangan Pada Resep Obat Menggunakan Convolutional Neural Network

    Get PDF
    Obat merupakan bahan kimia yang dapat merepresentasikan tubuh secara fisiologi dan psikologi ketika dikonsumsi. Obat sebagai alat bantu untuk menyembuhkan dari berbagai macam penyakit. Dengan berkembangnya zaman dan bertambahnya wawasan, menyebabkan bertambah juga jenis obat-obatan yang memiliki banyak manfaat dan kegunaanya. Penelitian ini bertujuan untuk mendeteksi nama obat dalam resep dokter menggunakan Convolutional Neural Network (CNN) dengan transfer learning. Metode transfer learning merupakan metode yang popular dalam mengklasifikasi gambar digital yang berguna untuk mempercepat proses klasifikasi. Penelitian ini membandingkan lima artistektur transfer learning yaitu VGG16, Resnet, Xception, LeNet, dan GoogleNet. Penelitian ini juga menggunakan grayscaling, resizing, dan median filter pada tahap preprocessing. Preprocessing digunakan untuk meningkatkan kualitas citra pada citra resep obat dan menghilangkan noise pada citra. ResNet-50 merupakan arsitektur terbaik untuk mengklasifikasi nama obat. Pada percobaan menggunakan ResNet-50, mendapatkan F1 score tertinggi yaitu sebesar 97,56% dan waktu training rata-rata 0,25 detik setiap epoch. Dapat disimpulkan Resnet merupakan arsitektur terbaik untuk mengklasifikasikan nama obat dalam citra resep dokter serta dapat mendeteksi nama obat secara akurat

    Application and Theory of Multimedia Signal Processing Using Machine Learning or Advanced Methods

    Get PDF
    This Special Issue is a book composed by collecting documents published through peer review on the research of various advanced technologies related to applications and theories of signal processing for multimedia systems using ML or advanced methods. Multimedia signals include image, video, audio, character recognition and optimization of communication channels for networks. The specific contents included in this book are data hiding, encryption, object detection, image classification, and character recognition. Academics and colleagues who are interested in these topics will find it interesting to read

    Advanced document data extraction techniques to improve supply chain performance

    Get PDF
    In this thesis, a novel machine learning technique to extract text-based information from scanned images has been developed. This information extraction is performed in the context of scanned invoices and bills used in financial transactions. These financial transactions contain a considerable amount of data that must be extracted, refined, and stored digitally before it can be used for analysis. Converting this data into a digital format is often a time-consuming process. Automation and data optimisation show promise as methods for reducing the time required and the cost of Supply Chain Management (SCM) processes, especially Supplier Invoice Management (SIM), Financial Supply Chain Management (FSCM) and Supply Chain procurement processes. This thesis uses a cross-disciplinary approach involving Computer Science and Operational Management to explore the benefit of automated invoice data extraction in business and its impact on SCM. The study adopts a multimethod approach based on empirical research, surveys, and interviews performed on selected companies.The expert system developed in this thesis focuses on two distinct areas of research: Text/Object Detection and Text Extraction. For Text/Object Detection, the Faster R-CNN model was analysed. While this model yields outstanding results in terms of object detection, it is limited by poor performance when image quality is low. The Generative Adversarial Network (GAN) model is proposed in response to this limitation. The GAN model is a generator network that is implemented with the help of the Faster R-CNN model and a discriminator that relies on PatchGAN. The output of the GAN model is text data with bonding boxes. For text extraction from the bounding box, a novel data extraction framework consisting of various processes including XML processing in case of existing OCR engine, bounding box pre-processing, text clean up, OCR error correction, spell check, type check, pattern-based matching, and finally, a learning mechanism for automatizing future data extraction was designed. Whichever fields the system can extract successfully are provided in key-value format.The efficiency of the proposed system was validated using existing datasets such as SROIE and VATI. Real-time data was validated using invoices that were collected by two companies that provide invoice automation services in various countries. Currently, these scanned invoices are sent to an OCR system such as OmniPage, Tesseract, or ABBYY FRE to extract text blocks and later, a rule-based engine is used to extract relevant data. While the system’s methodology is robust, the companies surveyed were not satisfied with its accuracy. Thus, they sought out new, optimized solutions. To confirm the results, the engines were used to return XML-based files with text and metadata identified. The output XML data was then fed into this new system for information extraction. This system uses the existing OCR engine and a novel, self-adaptive, learning-based OCR engine. This new engine is based on the GAN model for better text identification. Experiments were conducted on various invoice formats to further test and refine its extraction capabilities. For cost optimisation and the analysis of spend classification, additional data were provided by another company in London that holds expertise in reducing their clients' procurement costs. This data was fed into our system to get a deeper level of spend classification and categorisation. This helped the company to reduce its reliance on human effort and allowed for greater efficiency in comparison with the process of performing similar tasks manually using excel sheets and Business Intelligence (BI) tools.The intention behind the development of this novel methodology was twofold. First, to test and develop a novel solution that does not depend on any specific OCR technology. Second, to increase the information extraction accuracy factor over that of existing methodologies. Finally, it evaluates the real-world need for the system and the impact it would have on SCM. This newly developed method is generic and can extract text from any given invoice, making it a valuable tool for optimizing SCM. In addition, the system uses a template-matching approach to ensure the quality of the extracted information

    Mortality case note review use for hospital care quality improvement: A methodological, psychological and qualitative investigation

    Get PDF
    Evaluating hospital care quality using case-note reviews is mandated in the United Kingdom and is endorsed by many high-income countries. This thesis separately addresses both the validity of case-note reviews and the use of case-note reviews for care quality improvement. On case-note reviewing validity, there are moderate-to-high levels of disagreement (variability) between multiple clinician case-note reviewers when evaluating the overall care quality of the same case-note. The sources of this disagreement (variability) are unknown. On case-note review use, the potential factors which affect case-note reviewing in hospitals has not been well-studied in relation to their contribution to hospital care quality improvement. This thesis presents the findings of three original studies and seeks to both identify the sources for this reviewer variability and the organizational factors which influence case-note review’s likely contribution to hospital quality improvement. The introduction discusses the policy context and offers a critique of hospital mortality statistics with the prospective use of case-note reviews as an alternative approach for detecting care quality issues. Chapter 1 involved a systematic review of preventable mortality rates and a characterization of their measurement properties for evaluating care quality and subsequent hospital ranking. Findings concluded that a limitation of studies not accounting for variation between different hospitals, assuming equal variance, in the ranking process. Case-note reviews are presented as a workable alternative, to which this thesis is devoted to investigating. Chapter 2 presents the findings of an original systematic review which identified cognitive biases and heuristics related to case-note review care quality judgements. Cognitive biases and heuristics, sourced from two systematic reviews, are investigated with their plausible influence upon case-note reviewer care quality judgments using clinical scenarios derived using a systematic literature search and informed by a panel consensus. Findings indicate the plausible influence of cognitive biases and heuristics. Chapter 3 investigates the influence of reviewer attitudes; their demographics and patient case-note review characteristics upon case-note reviewer care quality judgements. Selected attitudes did not significantly influence care quality judgements and a significant proportion of care quality judgement variability is unexplained by the included independent variables. Chapter 4 describes case study fieldwork in an acute NHS Trust which explored the organizational processes around case-note review including its embedding, information flow and its perceived quality improvement contribution. We found that case-note reviews were well-embedded, with there being limited information flow from ward-to-board. Chapter 5 is a critical reflection of the research process and the assumptions made in this thesis. Chapter 6 summarizes the thesis, discusses practical implications, and identifies opportunities for future research for quality improvement from case-note reviews

    High-Performance Modelling and Simulation for Big Data Applications

    Get PDF
    This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications

    High-Performance Modelling and Simulation for Big Data Applications

    Get PDF
    This open access book was prepared as a Final Publication of the COST Action IC1406 “High-Performance Modelling and Simulation for Big Data Applications (cHiPSet)“ project. Long considered important pillars of the scientific method, Modelling and Simulation have evolved from traditional discrete numerical methods to complex data-intensive continuous analytical optimisations. Resolution, scale, and accuracy have become essential to predict and analyse natural and complex systems in science and engineering. When their level of abstraction raises to have a better discernment of the domain at hand, their representation gets increasingly demanding for computational and data resources. On the other hand, High Performance Computing typically entails the effective use of parallel and distributed processing units coupled with efficient storage, communication and visualisation systems to underpin complex data-intensive applications in distinct scientific and technical domains. It is then arguably required to have a seamless interaction of High Performance Computing with Modelling and Simulation in order to store, compute, analyse, and visualise large data sets in science and engineering. Funded by the European Commission, cHiPSet has provided a dynamic trans-European forum for their members and distinguished guests to openly discuss novel perspectives and topics of interests for these two communities. This cHiPSet compendium presents a set of selected case studies related to healthcare, biological data, computational advertising, multimedia, finance, bioinformatics, and telecommunications

    Making Government Work: Electronic Delivery of Federal Services

    Get PDF
    This report focuses on key topics and issues that are central to the successful use of electronic deli very by government. This report provides Congress with alternative strategies for improving the performance of government by using modern information technologies
    corecore