85 research outputs found
A model for error detection and correction
The aim of our work is to investigate how a relatively small
set of clauses can be transformed into a running program
capable of solving a number of problems. The problems are
chosen from the domain of simple arithmetic, algebra and
letter series completion. We describe how the problems are
solved, how errors are detected and corrected by modification
of the existing clauses.Various techniques useful in the process of error detection
and correction are described in detail. Two types of errors
are dealt with: selection errors arising due to incorrect
selection of clauses, and instantiation errors arising when
the partial results (subgoals) are not specific enough.The system described was implemented in Prolog
Comparison of SVM and some older classification algorithms in text classification tasks
Document classification has already been widely studied. In fact, some studies compared feature selection techniques or feature space transformation whereas some others compared the performance of different algorithms. Recently, following the rising interest towards the Support Vector Machine, various studies showed that SVM outperforms other classification algorithms. So should we just not bother about other classification algorithms and opt always for SVM We have decided to investigate this issue and compared SVM to kNN and naive Bayes on binary classification tasks. An important issue is to compare optimized versions of these algorithms, which is what we have done. Our results show all the classifiers achieved comparable performance on most problems. One surprising result is that SVM was not a clear winner, despite quite good overall performance. If a suitable preprocessing is used with kNN, this algorithm continues to achieve very good results and scales up well with the number of documents, which is not the case for SVM. As for naive Bayes, it also achieved good performance.IFIP International Conference on Artificial Intelligence in Theory and Practice - Knowledge Acquisition and Data MiningRed de Universidades con Carreras en Informática (RedUNCI
Association and Temporality between News and Tweets
With the advent of social media, the boundaries of mainstream journalism and social networks are becoming blurred. User-generated content is increasing, and hence, journalists dedicate considerable time searching platforms such as Facebook and Twitter to announce, spread, and monitor news and crowd check information. Many studies have looked at social networks as news sources, but the relationship and interconnections between this type of platform and news media have not been thoroughly investigated. In this work, we have studied a series of news articles and examined a set of related comments on a social network during a period of six months. Specifically, a sample of articles from generalist Portuguese news sources published in the first semester of 2016 was clustered, and the resulting clusters were then associated with tweets of Portuguese users with the recourse to a similarity measure. Focusing on a subset of clusters, we have performed a temporal analysis by examining the evolution of the two types of documents (articles and tweets) and the timing of when they appeared. It appears that for some stories, namely Brexit and the European Football Cup, the publishing of news articles intensifies on key dates (event-oriented), while the discussion on social media is more balanced throughout the months leading up to those events.info:eu-repo/semantics/publishedVersio
Comparison of SVM and some older classification algorithms in text classification tasks
Document classification has already been widely studied. In fact, some studies compared feature selection techniques or feature space transformation whereas some others compared the performance of different algorithms. Recently, following the rising interest towards the Support Vector Machine, various studies showed that SVM outperforms other classification algorithms. So should we just not bother about other classification algorithms and opt always for SVM We have decided to investigate this issue and compared SVM to kNN and naive Bayes on binary classification tasks. An important issue is to compare optimized versions of these algorithms, which is what we have done. Our results show all the classifiers achieved comparable performance on most problems. One surprising result is that SVM was not a clear winner, despite quite good overall performance. If a suitable preprocessing is used with kNN, this algorithm continues to achieve very good results and scales up well with the number of documents, which is not the case for SVM. As for naive Bayes, it also achieved good performance.IFIP International Conference on Artificial Intelligence in Theory and Practice - Knowledge Acquisition and Data MiningRed de Universidades con Carreras en Informática (RedUNCI
Visualização da relevância relativa de investigadores a partir da sua produção textual
A construção de uma rede de afinidade de investigadores através do processamento
automático das suas publicações permite obter uma perspetiva que vai para além das redes
estabelecidas através da coautoria. A definição da importância de cada investigador parte do seu
volume de produção bibliográfica, i.e., número de publicações, e também da sua centralidade
na rede geral de investigadores. De facto, a centralidade de um investigador numa rede revela
a sua importância nos fluxos de comunicação com os outros investigadores, pressupondo deste
modo que a comunicação entre investigadores é, em si própria, um fator relevante para a vida
organizacional e na sua produção.
Tanto os conceitos de rede como de centralidade são melhor interpretados de forma gráfica.
Neste estudo, exploramos o fluxo de trabalho que proporcionará estas visualizações e focamos
na seleção empírica da medida de centralidade mais adequada. Propomos também um método de visualização da centralidade que facilite a interpretação das medidas selecionadas.Building a researchers affinity network through the automatic processing of their
publications allows us to gain a perspective that goes beyond the networks established through
co-authorship. The definition of the importance of each researcher is defined upon their
bibliographic production volume, i.e., number of publications, and also upon their centrality in
the general network of researchers. In fact, the centrality of a researcher in a network reveals its
importance in communication flows with other researchers, thus assuming that communication
between researchers is itself a relevant factor for organizational life and in its production.
Both network and centrality concepts are better interpreted in a graphical way. In this
study, we explore the workflow that will provide these visualizations and focus in the empirical
selection of the most appropriate centrality measure. We also propose a centrality visualization
method that facilitates the interpretation of the selected measures
Evolving Networks and Social Network Analysis Methods and Techniques
Evolving networks by definition are networks that change as a function of time. They are a natural extension of network science since almost all real-world networks evolve over time, either by adding or by removing nodes or links over time: elementary actor-level network measures like network centrality change as a function of time, popularity and influence of individuals grow or fade depending on processes, and events occur in networks during time intervals. Other problems such as network-level statistics computation, link prediction, community detection, and visualization gain additional research importance when applied to dynamic online social networks (OSNs). Due to their temporal dimension, rapid growth of users, velocity of changes in networks, and amount of data that these OSNs generate, effective and efficient methods and techniques for small static networks are now required to scale and deal with the temporal dimension in case of streaming settings. This chapter reviews the state of the art in selected aspects of evolving social networks presenting open research challenges related to OSNs. The challenges suggest that significant further research is required in evolving social networks, i.e., existent methods, techniques, and algorithms must be rethought and designed toward incremental and dynamic versions that allow the efficient analysis of evolving networks
Metalearning
This open access book as one of the fastest-growing areas of research in machine learning, metalearning studies principled methods to obtain efficient models and solutions by adapting machine learning and data mining processes. This adaptation usually exploits information from past experience on other tasks and the adaptive processes can involve machine learning approaches. As a related area to metalearning and a hot topic currently, automated machine learning (AutoML) is concerned with automating the machine learning processes. Metalearning and AutoML can help AI learn to control the application of different learning methods and acquire new solutions faster without unnecessary interventions from the user. This book offers a comprehensive and thorough introduction to almost all aspects of metalearning and AutoML, covering the basic concepts and architecture, evaluation, datasets, hyperparameter optimization, ensembles and workflows, and also how this knowledge can be used to select, combine, compose, adapt and configure both algorithms and models to yield faster and better solutions to data mining and data science problems. It can thus help developers to develop systems that can improve themselves through experience. This book is a substantial update of the first edition published in 2009. It includes 18 chapters, more than twice as much as the previous version. This enabled the authors to cover the most relevant topics in more depth and incorporate the overview of recent research in the respective area. The book will be of interest to researchers and graduate students in the areas of machine learning, data mining, data science and artificial intelligence. ; Metalearning is the study of principled methods that exploit metaknowledge to obtain efficient models and solutions by adapting machine learning and data mining processes. While the variety of machine learning and data mining techniques now available can, in principle, provide good model solutions, a methodology is still needed to guide the search for the most appropriate model in an efficient way. Metalearning provides one such methodology that allows systems to become more effective through experience. This book discusses several approaches to obtaining knowledge concerning the performance of machine learning and data mining algorithms. It shows how this knowledge can be reused to select, combine, compose and adapt both algorithms and models to yield faster, more effective solutions to data mining problems. It can thus help developers improve their algorithms and also develop learning systems that can improve themselves. The book will be of interest to researchers and graduate students in the areas of machine learning, data mining and artificial intelligence
- …