4 research outputs found

    Clustering Arabic Tweets for Sentiment Analysis

    Get PDF
    The focus of this study is to evaluate the impact of linguistic preprocessing and similarity functions for clustering Arabic Twitter tweets. The experiments apply an optimized version of the standard K-Means algorithm to assign tweets into positive and negative categories. The results show that root-based stemming has a significant advantage over light stemming in all settings. The Averaged Kullback-Leibler Divergence similarity function clearly outperforms the Cosine, Pearson Correlation, Jaccard Coefficient and Euclidean functions. The combination of the Averaged Kullback-Leibler Divergence and root-based stemming achieved the highest purity of 0.764 while the second-best purity was 0.719. These results are of importance as it is contrary to normal-sized documents where, in many information retrieval applications, light stemming performs better than root-based stemming and the Cosine function is commonly used

    Clustering Arabic Tweets for Sentiment Analysis

    Get PDF
    The focus of this study is to evaluate the impact of linguistic preprocessing and similarity functions for clustering Arabic Twitter tweets. The experiments apply an optimized version of the standard K-Means algorithm to assign tweets into positive and negative categories. The results show that root-based stemming has a significant advantage over light stemming in all settings. The Averaged Kullback-Leibler Divergence similarity function clearly outperforms the Cosine, Pearson Correlation, Jaccard Coefficient and Euclidean functions. The combination of the Averaged Kullback-Leibler Divergence and root-based stemming achieved the highest purity of 0.764 while the second-best purity was 0.719. These results are of importance as it is contrary to normal-sized documents where, in many information retrieval applications, light stemming performs better than root-based stemming and the Cosine function is commonly used

    ENHANCING PRIVACY IN MULTI-AGENT SYSTEMS

    Full text link
    La pérdida de privacidad se está convirtiendo en uno de los mayores problemas en el mundo de la informática. De hecho, la mayoría de los usuarios de Internet (que hoy en día alcanzan la cantidad de 2 billones de usuarios en todo el mundo) están preocupados por su privacidad. Estas preocupaciones también se trasladan a las nuevas ramas de la informática que están emergiendo en los ultimos años. En concreto, en esta tesis nos centramos en la privacidad en los Sistemas Multiagente. En estos sistemas, varios agentes (que pueden ser inteligentes y/o autónomos) interactúan para resolver problemas. Estos agentes suelen encapsular información personal de los usuarios a los que representan (nombres, preferencias, tarjetas de crédito, roles, etc.). Además, estos agentes suelen intercambiar dicha información cuando interactúan entre ellos. Todo esto puede resultar en pérdida de privacidad para los usuarios, y por tanto, provocar que los usuarios se muestren adversos a utilizar estas tecnologías. En esta tesis nos centramos en evitar la colección y el procesado de información personal en Sistemas Multiagente. Para evitar la colección de información, proponemos un modelo para que un agente sea capaz de decidir qué atributos (de la información personal que tiene sobre el usuario al que representa) revelar a otros agentes. Además, proporcionamos una infraestructura de agentes segura, para que una vez que un agente decide revelar un atributo a otro, sólo este último sea capaz de tener acceso a ese atributo, evitando que terceras partes puedan acceder a dicho atributo. Para evitar el procesado de información personal proponemos un modelo de gestión de las identidades de los agentes. Este modelo permite a los agentes la utilización de diferentes identidades para reducir el riesgo del procesado de información. Además, también describimos en esta tesis la implementación de dicho modelo en una plataforma de agentes.Such Aparicio, JM. (2011). ENHANCING PRIVACY IN MULTI-AGENT SYSTEMS [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/13023Palanci

    Combi-BP: automating the data-oriented optimization in business processes. From declarative to executable models.

    Get PDF
    One of the main objectives of a business expert is to model the business goals of an enterprise process. Several languages have been created to describe the necessary activities to achieve the objective, especially in the business process context. These languages can be divided into imperative and declarative ones. Declarative languages tend to be used when the speci c model is unknown, being possible to describe what has to be done instead of how. Otherwise, imperative languages permit to describe how the things have to be done and then, the imperative models can be executed in any Business Process Management System (BPMS). The declarative descriptions are more exible, since they permit to describe the model in a more relaxed way, which means that various process executions can follow the same declarative description. However, both paradigms are focused on the activities order description, but unfortunately, the data perspective is missed. Furthermore, the optimization of a business goal which depends on the exchanged data during the execution of the business process has not been included in previous proposals. There are no solutions that allow the business experts to describe nor execute a declarative description where the executed model depends on the exchanged data between the involved activities in each instance. In this thesis dissertation, an approach to support this data-oriented optimization in business process is presented. A data-oriented optimization problem is a process whose main purpose is to obtain the best business product. In order to obtain this business product, the process must combine several activities by taking into account the existing data-structure and data-value dependencies. Both kind of dependencies are established by a set of constraints that relate the data (consumed and provided by the activities) and the data given by the customer. Therefore, the BPs under the scope of our research are those which are centred on developing sound data in business processes, analysing how data-structure and data-value dependencies can aect the correct business process execution. However, if the data provided at runtime for the activities that conform the model have not got enough level of quality, then business process will not be successfully executed. The base of the proposal is focused on the combination of the advantages of both paradigms: the exibility of the declaratives, and the automatic execution in a BPMS of the imperatives. On the one hand, we want to describe a exible model using a declarative description where the exchanged data and an optimization objective are included. In the other hand, this declarative model must be executed in a generic business process management system with the aim of support any instance of the process. Therefore, how the declarative description can be transformed into an imperative business process is developed. The transformation methodology that we propose is based on Model-Driven Architecture. Firstly, the declarative is transformed into an imperative which takes into account the data-structure dependencies. The exibility of the declarative speci cation is kept thanks to the use of Constraint Programming. On the other hand, the resulting imperative model is enriched with new intelligent techniques, also based on Constraint Programming, in order to solve the data-value dependencies. Finally, a methodology and an implementation are developed in order to make the business process aware of the data-quality aspects
    corecore