59 research outputs found

    A Survey on Particle Swarm Optimization for Association Rule Mining

    Get PDF
    Association rule mining (ARM) is one of the core techniques of data mining to discover potentially valuable association relationships from mixed datasets. In the current research, various heuristic algorithms have been introduced into ARM to address the high computation time of traditional ARM. Although a more detailed review of the heuristic algorithms based on ARM is available, this paper differs from the existing reviews in that we expected it to provide a more comprehensive and multi-faceted survey of emerging research, which could provide a reference for researchers in the field to help them understand the state-of-the-art PSO-based ARM algorithms. In this paper, we review the existing research results. Heuristic algorithms for ARM were divided into three main groups, including biologically inspired, physically inspired, and other algorithms. Additionally, different types of ARM and their evaluation metrics are described in this paper, and the current status of the improvement in PSO algorithms is discussed in stages, including swarm initialization, algorithm parameter optimization, optimal particle update, and velocity and position updates. Furthermore, we discuss the applications of PSO-based ARM algorithms and propose further research directions by exploring the existing problems.publishedVersio

    Swarm intelligence algorithms adaptation for various search spaces

    Get PDF
    U današnje vrijeme postoji mnogo algoritama inteligencije rojeva koji se uspiješno koriste za rešavanje raznih teških problema optimizacije. Zajednicki elementi svih ovih algoritama su operator za lokalnu pretragu (eksploataciju) oko prona enih obecavajucih rješenja i operator globalne pretrage (eksploracije) koji pomaže u bijegu iz lokalnih optimuma. Algoritmi inteligencije rojeva obicno se inicijalno testiraju na neogranicenim, ogranicenim ili visoko-dimenzionalnim skupovima standardnih test funkcija. Nadalje, mogu se poboljšati, prilagoditi, izmijeniti, hibridizirati, kombinirati s lokalnom pretragom. Konacna svrha je korištenje takve metaheuristike za optimizaciju problema iz stvarnog svijeta. Domeni rješenja odnosno prostori pretrage prakticnih teških problema optimizacije mogu biti razliciti. Rješenja mogu biti vektori iz skupa realnih brojeva, cijelih brojeva ali mogu biti i kompleksnije strukture. Algoritmi inteligencije rojeva moraju se prilagoditi za razlicite prostore pretrage što može biti jednostavno podešavanje parametera algoritma ili prilagodba za cjelobrojna rješenja jednostavnim zaokruživanjem dobivenih realnih rješenja ali za pojedine prostore pretrage potrebnao je skoro kompletno prepravljanja algoritma ukljucujuci i operatore ekploatacije i ekploracije zadržavajuci samo proces vo enja odnosno inteligenciju roja. U disertaciji je predstavljeno nekoliko algoritama inteligencije rojeva i njihova prilagodba za razlicite prostore pretrage i primjena na prakticne probleme. Ova disertacija ima za cilj analizirati i prilagoditi, u zavisnosti od funkcije cilja i prostora rješenja, algoritme inteligencije rojeva. Predmet disertacije ukljucuje sveobuhvatan pregled postojecih implementacija algoritama inteligencije rojeva. Disertacija tako er obuhvaca komparativnu analizu, prikaz slabosti i snaga jednih algoritama u odnosu na druge zajedno s istraživanjem prilagodbi algoritama inteligencije rojeva za razlicite prostore pretrage i njihova primjena na prakticne problem. Razmatrani su problemi sa realnim rješenjima kao što su optimizacija stroja potpornih vektora, grupiranje podataka, sa cijelobrojnim rješenjima kao što je slucaj problema segmentacije digitalnih slika i za probleme gdje su rješenja posebne strukture kao što su problemi planiranja putanje robota i triangulacije minimalne težine. Modificirani i prilago eni algoritmi inteligencije rojeva za razlicite prostore pretrage i primjenih na prakticne probleme testirani su na standardnim skupovima test podataka i uspore eni s drugim suvremenim metodama za rješavanje promatranih problema iz literature. Pokazane su uspješne prilagodbe algoritama inteligencije rojeva za razne prostore pretrage. Ovako prilago eni algoritmi su u svim slucajevima postigli bolje rezultate u usporedbi sa metodama iz literature, što dovodi do zakljucka da je moguce prilagoditi algoritme inteligencije rojeva za razne prostore pretrage ukljucujuci i kompleksne strukture i postici bolje rezultate u usporedbi sa metodama iz literature

    UM ESTUDO DE MAPEAMENTO SISTEMÁTICO DA MINERAÇÃO DE DADOS PARA CENÁRIOS DE BIG DATA

    Get PDF
    O volume de dados produzidos tem crescido em larga escala nos últimos anos. Esses dados são de diferentes fontes e diversificados formatos, caracterizando as principais dimensões do Big Data: grande volume, alta velocidade de crescimento e grande variedade de dados. O maior desafio é como gerar informação de qualidade para inferir insights significativos de tais dados variados e grandes. A Mineração de Dados é o processo de identificar, em dados, padrões válidos, novos e potencialmente úteis. No entanto, a infraestrutura de tecnologia da informação tradicional não é capaz de atender as demandas deste novo cenário. O termo atualmente conhecido como Big Data Mining refere-se à extração de informação a partir de grandes bases de dados. Uma questão a ser respondida é como a comunidade científica está abordando o processo de Big Data Mining? Seria válido identificar quais tarefas, métodos e algoritmos vêm sendo aplicados para extrair conhecimento neste contexto. Este artigo tem como objetivo identificar na literatura os trabalhos de pesquisa já realizados no contexto do Big Data Mining. Buscou-se identificar as áreas mais abordadas, os tipos de problemas tratados, as tarefas aplicadas na extração de conhecimento, os métodos aplicados para a realização das tarefas, os algoritmos para a implementação dos métodos, os tipos de dados que vêm sendo minerados, fonte e estrutura dos mesmos. Um estudo de mapeamento sistemático foi conduzido, foram examinados 78 estudos primários. Os resultados obtidos apresentam uma compreensão panorâmica da área investigada, revelando as principais tarefas, métodos e algoritmos aplicados no Big Data Mining

    Evolutionary Computation 2020

    Get PDF
    Intelligent optimization is based on the mechanism of computational intelligence to refine a suitable feature model, design an effective optimization algorithm, and then to obtain an optimal or satisfactory solution to a complex problem. Intelligent algorithms are key tools to ensure global optimization quality, fast optimization efficiency and robust optimization performance. Intelligent optimization algorithms have been studied by many researchers, leading to improvements in the performance of algorithms such as the evolutionary algorithm, whale optimization algorithm, differential evolution algorithm, and particle swarm optimization. Studies in this arena have also resulted in breakthroughs in solving complex problems including the green shop scheduling problem, the severe nonlinear problem in one-dimensional geodesic electromagnetic inversion, error and bug finding problem in software, the 0-1 backpack problem, traveler problem, and logistics distribution center siting problem. The editors are confident that this book can open a new avenue for further improvement and discoveries in the area of intelligent algorithms. The book is a valuable resource for researchers interested in understanding the principles and design of intelligent algorithms

    SVMAUD: Using textual information to predict the audience level of written works using support vector machines

    Get PDF
    Information retrieval systems should seek to match resources with the reading ability of the individual user; similarly, an author must choose vocabulary and sentence structures appropriate for his or her audience. Traditional readability formulas, including the popular Flesch-Kincaid Reading Age and the Dale-Chall Reading Ease Score, rely on numerical representations of text characteristics, including syllable counts and sentence lengths, to suggest audience level of resources. However, the author’s chosen vocabulary, sentence structure, and even the page formatting can alter the predicted audience level by several levels, especially in the case of digital library resources. For these reasons, the performance of readability formulas when predicting the audience level of digital library resources is very low. Rather than relying on these inputs, machine learning methods, including cosine, Naïve Bayes, and Support Vector Machines (SVM), can suggest the grade level of an essay based on the vocabulary chosen by the author. The audience level prediction and essay grading problems share the same inputs, expert-labeled documents, and outputs, a numerical score representing quality or audience level. After a human expert labels a representative sample of resources with audience level, the proposed SVM-based audience level prediction program, SVMAUD, constructs a vocabulary for each audience level; then, the text in an unlabeled resource is compared with this predefined vocabulary to suggest the most appropriate audience level. Two readability formulas and four machine learning programs are evaluated with respect to predicting human-expert entered audience levels based on the text contained in an unlabeled resource. In a collection containing 10,238 expert-labeled HTML-based digital library resources, the Flesch-Kincaid Reading Age and the Dale-Chall Reading Ease Score predict the specific audience level with F-measures of 0.10 and 0.05, respectively. Conversely, cosine, Naïve Bayes, the Collins-Thompson and Callan model, and SVMAUD improve these F-measures to 0.57, 0.61, 0.68, and 0.78, respectively. When a term’s weight is adjusted based on the HTML tag in which it occurs, the specific audience level prediction performance of cosine, Naïve Bayes, the Collins-Thompson and Callan method, and SVMAUD improves to 0.68, 0.70, 0.75, and 0.84, respectively. When title, keyword, and abstract metadata is used for training, cosine, Naïve Bayes, the Collins-Thompson and Callan model, and SVMAUD specific audience level prediction F-measures are found to be 0.61, 0.68, 0.75, and 0.86, respectively. When cosine, Naïve Bayes, the Collins-Thompson and Callan method, and SVMAUD are trained and tested using resources from a single subject category, the specific audience level prediction F- measure performance improves to 0.63, 0.70, 0.77, and 0.87, respectively. SVMAUD experiences the highest audience level prediction performance among all methods under evaluation in this study. After SVMAUD is properly trained, it can be used to predict the audience level of any written work

    Speciation analysis of <sup>129</sup>I in the Environment

    Get PDF

    AMS and ICP-MS for determination of long-lived environmental radionuclides

    Get PDF

    Understanding Quantum Technologies 2022

    Full text link
    Understanding Quantum Technologies 2022 is a creative-commons ebook that provides a unique 360 degrees overview of quantum technologies from science and technology to geopolitical and societal issues. It covers quantum physics history, quantum physics 101, gate-based quantum computing, quantum computing engineering (including quantum error corrections and quantum computing energetics), quantum computing hardware (all qubit types, including quantum annealing and quantum simulation paradigms, history, science, research, implementation and vendors), quantum enabling technologies (cryogenics, control electronics, photonics, components fabs, raw materials), quantum computing algorithms, software development tools and use cases, unconventional computing (potential alternatives to quantum and classical computing), quantum telecommunications and cryptography, quantum sensing, quantum technologies around the world, quantum technologies societal impact and even quantum fake sciences. The main audience are computer science engineers, developers and IT specialists as well as quantum scientists and students who want to acquire a global view of how quantum technologies work, and particularly quantum computing. This version is an extensive update to the 2021 edition published in October 2021.Comment: 1132 pages, 920 figures, Letter forma
    corecore