12 research outputs found

    Engaging stakeholders through Facebook. The case of Global Compact LEAD participants

    Get PDF
    Facebook has deeply modified the way people communicate and interact. From a business perspective, Facebook has enormous potential as a means of communication and stakeholder engagement. It enables companies to share contents rapidly and efficiently with a large number of stakeholders worldwide. People can personalise their Facebook profile to receive updates from selected companies. Moreover, people can reply to such posts or simply manifest their approval by liking or sharing the posts. In this way, people also propagate corporate information among their own friends. The dramatic diffusion of Facebook should encourage companies to virtually interact with a network of stakeholders 2.0, using Facebook as a stakeholder engagement tool. The evolution to Web 2.0 goes with a general change in the social and business environment. In today’s world, both policy makers and the public expect that companies work in a sustainable way and consult their stakeholders about corporate strategies, operations and performance. The discussion should concern social and ecological cares as well as economic issues. In this sense, the engagement of the Facebook community could considerable enlarge and improve the dialogue. This paper offers a theoretical and empirical analysis to answer the following research question: do sustainability-oriented companies use Facebook as an effective means of stakeholder engagement? The paper contains an investigation based on UN Global Compact LEAD members, characterised by strong commitment and cooperation with governments, civil society, labour and the UN in order to promote sustainable practices. To evaluate the contribution of Facebook to the dialogue on sustainability, the investigation considered the types of contents published by the LEAD companies on their Facebook pages in 30 days. According to the subject, seven categories of posts emerged from the analysis: human rights and social citizenship; labour; environment; anti-corruption; strategy, business activity and economic performance; news on products and services; other. To evaluate the use of Facebook for stakeholder engagement 2.0, the investigation verified how many “likes”, comments and “shares” each post received and how often the company replied. The analysis showed that some LEAD members did not have a Facebook profile, which is unacceptable nowadays. Moreover, the companies with an official page rarely covered all three perspectives of sustainability (social, environmental, and economic issues). Furthermore, companies rarely replied to stakeholders’ comments. Based on the empirical evidence, most LEAD participants should modify the way they used Facebook. Therefore, the results of this research may help them improve stakeholder engagement 2.0

    Geotag Propagation with User Trust Modeling

    Get PDF
    The amount of information that people share on social networks is constantly increasing. People also comment, annotate, and tag their own content (videos, photos, notes, etc.), as well as the content of others. In many cases, the content is tagged manually. One way to make this time-consuming manual tagging process more efficient is to propagate tags from a small set of tagged images to the larger set of untagged images automatically. In such a scenario, however, a wrong or a spam tag can damage the integrity and reliability of the automated propagation system. Users may make mistakes in tagging, or irrelevant tags and content may be added maliciously for advertisement or self-promotion. Therefore, a certain mechanism insuring the trustworthiness of users or published content is needed. In this chapter, we discuss several image retrieval methods based on tags, various approaches to trust modeling and spam protection in social networks, and trust modeling in geotagging systems. We then consider a specific example of automated geotag propagation system that adopts a user trust model. The tag propagation in images relies on the similarity between image content (famous landmarks) and its context (associated geotags). For each tagged image, similar untagged images are found by the robust graph-based object duplicate detection and the known tags are propagated accordingly. The user trust value is estimated based on a social feedback from the users of the photo-sharing system and only tags from trusted users are propagated. This approach demonstrates that a practical tagging system significantly benefits from the intelligent combination of efficient propagation algorithm and a user-centered trust model

    Information quality in online social media and big data collection: an example of Twitter spam detection

    Get PDF
    La popularité des médias sociaux en ligne (Online Social Media - OSM) est fortement liée à la qualité du contenu généré par l'utilisateur (User Generated Content - UGC) et la protection de la vie privée des utilisateurs. En se basant sur la définition de la qualité de l'information, comme son aptitude à être exploitée, la facilité d'utilisation des OSM soulève de nombreux problèmes en termes de la qualité de l'information ce qui impacte les performances des applications exploitant ces OSM. Ces problèmes sont causés par des individus mal intentionnés (nommés spammeurs) qui utilisent les OSM pour disséminer des fausses informations et/ou des informations indésirables telles que les contenus commerciaux illégaux. La propagation et la diffusion de telle information, dit spam, entraînent d'énormes problèmes affectant la qualité de services proposés par les OSM. La majorité des OSM (comme Facebook, Twitter, etc.) sont quotidiennement attaquées par un énorme nombre d'utilisateurs mal intentionnés. Cependant, les techniques de filtrage adoptées par les OSM se sont avérées inefficaces dans le traitement de ce type d'information bruitée, nécessitant plusieurs semaines ou voir plusieurs mois pour filtrer l'information spam. En effet, plusieurs défis doivent être surmontées pour réaliser une méthode de filtrage de l'information bruitée . Les défis majeurs sous-jacents à cette problématique peuvent être résumés par : (i) données de masse ; (ii) vie privée et sécurité ; (iii) hétérogénéité des structures dans les réseaux sociaux ; (iv) diversité des formats du UGC ; (v) subjectivité et objectivité. Notre travail s'inscrit dans le cadre de l'amélioration de la qualité des contenus en termes de messages partagés (contenu spam) et de profils des utilisateurs (spammeurs) sur les OSM en abordant en détail les défis susmentionnés. Comme le spam social est le problème le plus récurant qui apparaît sur les OSM, nous proposons deux approches génériques pour détecter et filtrer le contenu spam : i) La première approche consiste à détecter le contenu spam (par exemple, les tweets spam) dans un flux en temps réel. ii) La seconde approche est dédiée au traitement d'un grand volume des données relatives aux profils utilisateurs des spammeurs (par exemple, les comptes Twitter). Pour filtrer le contenu spam en temps réel, nous introduisons une approche d'apprentissage non supervisée qui permet le filtrage en temps réel des tweets spams dans laquelle la fonction de classification est adaptée automatiquement. La fonction de classification est entraîné de manière itérative et ne requière pas une collection de données annotées manuellement. Dans la deuxième approche, nous traitons le problème de classification des profils utilisateurs dans le contexte d'une collection de données à grande échelle. Nous proposons de faire une recherche dans un espace réduit de profils utilisateurs (une communauté d'utilisateurs) au lieu de traiter chaque profil d'utilisateur à part. Ensuite, chaque profil qui appartient à cet espace réduit est analysé pour prédire sa classe à l'aide d'un modèle de classification binaire. Les expériences menées sur Twitter ont montré que le modèle de classification collective non supervisé proposé est capable de générer une fonction efficace de classification binaire en temps réel des tweets qui s'adapte avec l'évolution des stratégies des spammeurs sociaux sur Twitter. L'approche proposée surpasse les performances de deux méthodes de l'état de l'art de détection de spam en temps réel. Les résultats de la deuxième approche ont démontré que l'extraction des métadonnées des spams et leur exploitation dans le processus de recherche de profils de spammeurs est réalisable dans le contexte de grandes collections de profils Twitter. L'approche proposée est une alternative au traitement de tous les profils existants dans le OSM.The popularity of OSM is mainly conditioned by the integrity and the quality of UGC as well as the protection of users' privacy. Based on the definition of information quality as fitness for use, the high usability and accessibility of OSM have exposed many information quality (IQ) problems which consequently decrease the performance of OSM dependent applications. Such problems are caused by ill-intentioned individuals who misuse OSM services to spread different kinds of noisy information, including fake information, illegal commercial content, drug sales, mal- ware downloads, and phishing links. The propagation and spreading of noisy information cause enormous drawbacks related to resources consumptions, decreasing quality of service of OSM-based applications, and spending human efforts. The majority of popular social networks (e.g., Facebook, Twitter, etc) over the Web 2.0 is daily attacked by an enormous number of ill-intentioned users. However, those popular social networks are ineffective in handling the noisy information, requiring several weeks or months to detect them. Moreover, different challenges stand in front of building a complete OSM-based noisy information filtering methods that can overcome the shortcomings of OSM information filters. These challenges are summarized in: (i) big data; (ii) privacy and security; (iii) structure heterogeneity; (iv) UGC format diversity; (v) subjectivity and objectivity; (vi) and service limitations In this thesis, we focus on increasing the quality of social UGC that are published and publicly accessible in forms of posts and profiles over OSNs through addressing in-depth the stated serious challenges. As the social spam is the most common IQ problem appearing over the OSM, we introduce a design of two generic approaches for detecting and filtering out the spam content. The first approach is for detecting the spam posts (e.g., spam tweets) in a real-time stream, while the other approach is dedicated for handling a big data collection of social profiles (e.g., Twitter accounts). For filtering the spam content in real-time, we introduce an unsupervised collective-based framework that automatically adapts a supervised spam tweet classification function in order to have an updated real-time classifier without requiring manual annotated data-sets. In the second approach, we treat the big data collections through minimizing the search space of profiles that needs advanced analysis, instead of processing every user's profile existing in the collections. Then, each profile falling in the reduced search space is further analyzed in an advanced way to produce an accurate decision using a binary classification model. The experiments conducted on Twitter online social network have shown that the unsupervised collective-based framework is able to produce updated and effective real- time binary tweet-based classification function that adapts the high evolution of social spammer's strategies on Twitter, outperforming the performance of two existing real- time spam detection methods. On the other hand, the results of the second approach have demonstrated that performing a preprocessing step for extracting spammy meta-data values and leveraging them in the retrieval process is a feasible solution for handling a large collections of Twitter profiles, as an alternative solution for processing all profiles existing in the input data collection. The introduced approaches open different opportunities for information science researcher to leverage our solutions in other information filtering problems and applications. Our long term perspective consists of (i) developing a generic platform covering most common OSM for instantly checking the quality of a given piece of information where the forms of the input information could be profiles, website links, posts, and plain texts; (ii) and transforming and adapting our methods to handle additional IQ problems such as rumors and information overloading

    Analyse des réseaux sociaux : vers une adaptation de la navigation sociale

    Get PDF
    L'avènement du web 2.0, centré utilisateur, a fait émerger une quantité importante d'informations (personnelles, collectives, partagées, "aimées", etc.). Ces informations peuvent constituer une aide pour les utilisateurs en les guidant vers l'information recherchée. Cependant, cette quantité rend l'accès à l'information partagée de plus en plus difficile, vu la diversité des contenus qui peuvent intéresser l'utilisateur. La désorientation de l'utilisateur est donc l'un des principaux problèmes liés aux médias sociaux. Pour surmonter ce problème, l'adaptation constitue une solution classique qui peut être appliquée dans un contexte social. Avec l'évolution des réseaux sociaux, de nouvelles notions apparaissent comme la navigation sociale, qui est une manière de naviguer en étant influencé par les autres utilisateurs du réseau. Une autre notion importante est celle de "tag". Ce terme définit les annotations sociales créées par les utilisateurs et associées à des ressources. La navigation peut être dès lors effectuée aussi bien par les liens qu'à travers les tags. Adapter la navigation sociale, signifie la rendre plus ciblée pour chaque utilisateur selon ses intérêts. Concrètement, cela peut se faire en recommandant à chaque utilisateur des tags, qu'il pourra suivre ou non. Pour cela, il faut garantir une détection adéquate des intérêts de l'utilisateur ainsi que la prise en compte de leur évolution. Cependant, nous sommes confrontés à des limites liées à : i) la détection des intérêts, puisque ces derniers peuvent être déduits de plusieurs ressources sociales (des amis, des ressources, des tags, etc.). Leur pertinence est primordiale afin de garantir un résultat d'adaptation adéquat. ii) la mise à jour du profil utilisateur. En effet, l'utilisateur social, est caractérisé par sa grande activité sociale, et par conséquent ses intérêts doivent refléter ses "vrais" intérêts à chaque période de temps afin d'aboutir à une adaptation fiable. Afin de résoudre les problèmes affectant la qualité d'adaptation de la navigation sociale cités ci-dessus, nous avons proposé en premier lieu, une approche de détection des intérêts de l'utilisateur. Cette approche analyse les tags des utilisateurs selon le contenu de leurs ressources respectives. La plupart des recherches ne considèrent pas l'exactitude des tags vis-à-vis du contenu des ressources : cette exactitude reflète si l'utilisateur peut vraiment être intéressé par le contenu ou pas. Les tags précis sont ceux qui reflètent fidèlement le contenu des ressources. Ceci est effectué grâce à l'interrogation du réseau de l'utilisateur et de l'analyse de son comportement d'annotation. Notre approche repose sur l'hypothèse qu'un utilisateur qui annote la ressource par des tags reflétant le contenu de ladite ressource, reflète mieux ses "vrais" intérêts. Nous avons proposé en deuxième lieu, une approche de mise à jour des intérêts des utilisateurs. Nous nous sommes intéressés aux techniques d'enrichissement du profil utilisateur est effectué par l'ajout des intérêts jugés pertinents à un moment donné. L'enrichissement dans un contexte social est effectué selon l'information sociale comme les personnes proches qui partagent avec l'utilisateur des comportements en communs, selon le comportement d'annotation des utilisateurs, et selon les métadonnées des ressources annotées. Le choix de ces informations est effectué selon l'étude de leur influence sur l'évolution des intérêts de l'utilisateur. L'approche d'enrichissement nous a servi à proposer des recommandations (de tags) selon les nouveaux tags ajoutés au profil utilisateur.Ces deux contributions ont été testées sur la base sociale Delicious. Elles ont montré un taux de précision assez important. Elles ont aussi prouvé leur efficacité par rapport à des méthodes classiques. De plus, le taux d'ambigüité associé aux tags a été fortement réduit, grâce au filtrage implicite des tags non pertinents par rapport au contenu des ressources.The advent of Web 2.0, user-centered, has given rise to a significant amount of information (personal, collective, shared, "loved", etc.). This information is a way to help users and guide them to the information sought. However, this quantity makes access to shared information more and more difficult, given the diversity of content that may interest the user. Disorientation of the user is one of the main problems related to social media. To overcome such problem, adaptation is a standard solution that can be applied in a social context. With the evolution of these social networks, new concepts appear such as social navigation, which is a way to navigate while being influenced by other users in the network: Another important concept is that of "tag". This term is defined as social annotations created by users and associated to resources. Navigation can be therefore carried out by both links and tags. Adapting social navigation means making it more targeted for each user according to their interests. In practice, this can be done by recommending tags to each user, so he can follow or not. To adapt the social navigation, we must ensure proper detection of the user's interests and taking into account their evolution. However, we are faced with some problems: i) the detection of interest, since they can be derived from several social resources (friends, resources, tags, etc.). Their relevance is primordial to ensure adequate adaptation result. ii) updating the user profile. Indeed, the social user, is characterized by its great social activity, and therefore its interests should reflect its "real" interest each time period in order to achieve a reliable adaptation. To solve the problems affecting the quality of adaptation of social navigation quoted above, we first proposed a method for detecting the user's interests. This proposal aims to overcome the detection of irrelevant interests issues. This approach analyzes the user tags depending on the content of their respective resources. Unlike most research, who do not consider the accuracy of tags with the contents of resource, the accuracy reflects whether the user is really interested with the content or not. This is done by querying the user's network and analysis of the user annotation behavior. The approach is based on the assumption that a user annotates the resource by tags reflecting the content of this resource better reflects its "true" interests. Following the proposal of the interests of detection approach, we conducted second, the treatment of the problem of updating these interests. We were interested to the user profile enrichment techniques, performed by adding interests deemed relevant at a given time. The enrichment in a social context is performed according to social information such as neighbours who share the user behaviors in common, according to the user annotation behavior, and according to the metadata annotated resources. The choice of such information shall follow the study of their influence on the changing interests of the user. The approach we used enrichment propose recommendations (tags) according to the new tags added to the user profile. Both contributions were tested on the social database Delicious. They showed a sizeable accuracy rate. They have also proven their efficiency compared to conventional methods. In addition, the rate of ambiguity associated with the tags has been greatly reduced, thanks to the implicit filtering of irrelevant tags relative to resource content

    Learning plan networks in conversational video games

    Get PDF
    Thesis (S.M.)--Massachusetts Institute of Technology, School of Architecture and Planning, Program in Media Arts and Sciences, 2007.Includes bibliographical references (p. 121-123).We look forward to a future where robots collaborate with humans in the home and workplace, and virtual agents collaborate with humans in games and training simulations. A representation of common ground for everyday scenarios is essential for these agents if they are to be effective collaborators and communicators. Effective collaborators can infer a partner's goals and predict future actions. Effective communicators can infer the meaning of utterances based on semantic context. This thesis introduces a computational cognitive model of common ground called a Plan Network. A Plan Network is a statistical model that provides representations of social roles, object affordances, and expected patterns of behavior and language. I describe a methodology for unsupervised learning of a Plan Network using a multiplayer video game, visualization of this network, and evaluation of the learned model with respect to human judgment of typical behavior. Specifically, I describe learning the Restaurant Plan Network from data collected from over 5,000 players of an online game called The Restaurant Game.by Jeffrey David Orkin.S.M

    Using eye tracking for evaluation of information visualisation in web search interfaces

    Get PDF
    Search result organization and presentation is an important component of a web search system, and can have a substantial impact on the ability of users to find useful information. Most web search result interfaces include textual information, including for example the document title, URL, and a short query-biased summary of the content. Recent studies have developed various novel visual summaries, aiming to improve the effectiveness of search results. In this thesis, the impact and efficacy of presenting additional visual summaries are investigated through a series of four studies. User interaction with the search results was captured using eye tracking data. In the first study we compare the effectiveness of three publicly available search interfaces for supporting navigational search tasks. The three interfaces varied primarily in the proportion of visual versus textual cues that were used to display a search result. Our analysis shows that users' search completion time varies greatly among interfaces, and an appropriate combination of textual and visual information leads to the shortest search completion time and the least number of wrong answers. Another outcome of this experiment is the identification of factors that should be accounted for in subsequent, more controlled, experiments with visual summaries, including the size of the visual summaries and interface design. An understanding of the features and limitations of the eye tracker, particularly for IR studies, was also obtained. To obtain a richer understanding of a user's information seeking strategies and the impact of presenting additional visual summaries, five interfaces were designed: text-only, thumbnail, image, tag and visual snippet. In the second study, fifty participants carried out searches on five informational topics, using the five different interfaces. Findings show that visual summaries significantly impact on the behaviour of users, but not on their performance when predicting the relevance of answer resources. In the third study, fifty participants carried out five navigational topics using the five different interfaces. The results show that apart from the salient image interface, users perform statistically significantly better in terms of time required and effort required to answer given navigational search topics when additional visual summaries are presented. The fourth study was conducted with both navigational and informational topics, for a more detailed comparison between the best-performing interfaces identified in the previous studies: salient images for informational searches, and thumbnails for navigational searches. The findings confirm our previous results. Overall, the salient image interface can significantly increase user performance with informational topics, while thumbnails can help users to predict relevant answers, in a significantly shorter time, with navigational search topics

    Construction de corpus généraux et spécialisés à partir du Web

    Get PDF
    At the beginning of the first chapter the interdisciplinary setting between linguistics, corpus linguistics, and computational linguistics is introduced. Then, the notion of corpus is put into focus. Existing corpus and text definitions are discussed. Several milestones of corpus design are presented, from pre-digital corpora at the end of the 1950s to web corpora in the 2000s and 2010s. The continuities and changes between the linguistic tradition and web native corpora are exposed.In the second chapter, methodological insights on automated text scrutiny in computer science, computational linguistics and natural language processing are presented. The state of the art on text quality assessment and web text filtering exemplifies current interdisciplinary research trends on web texts. Readability studies and automated text classification are used as a paragon of methods to find salient features in order to grasp text characteristics. Text visualization exemplifies corpus processing in the digital humanities framework. As a conclusion, guiding principles for research practice are listed, and reasons are given to find a balance between quantitative analysis and corpus linguistics, in an environment which is spanned by technological innovation and artificial intelligence techniques.Third, current research on web corpora is summarized. I distinguish two main approaches to web document retrieval: restricted retrieval and web crawling. The notion of web corpus preprocessing is introduced and salient steps are discussed. The impact of the preprocessing phase on research results is assessed. I explain why the importance of preprocessing should not be underestimated and why it is an important task for linguists to learn new skills in order to confront the whole data gathering and preprocessing phase.I present my work on web corpus construction in the fourth chapter. My analyses concern two main aspects, first the question of corpus sources (or prequalification), and secondly the problem of including valid, desirable documents in a corpus (or document qualification). Last, I present work on corpus visualization consisting of extracting certain corpus characteristics in order to give indications on corpus contents and quality.Le premier chapitre s'ouvre par un description du contexte interdisciplinaire. Ensuite, le concept de corpus est présenté en tenant compte de l'état de l'art. Le besoin de disposer de preuves certes de nature linguistique mais embrassant différentes disciplines est illustré par plusieurs scénarios de recherche. Plusieurs étapes clés de la construction de corpus sont retracées, des corpus précédant l'ère digitale à la fin des années 1950 aux corpus web des années 2000 et 2010. Les continuités et changements entre la tradition en linguistique et les corpus tirés du web sont exposés.Le second chapitre rassemble des considérations méthodologiques. L'état de l'art concernant l'estimation de la qualité de textes est décrit. Ensuite, les méthodes utilisées par les études de lisibilité ainsi que par la classification automatique de textes sont résumées. Des dénominateurs communs sont isolés. Enfin, la visualisation de textes démontre l'intérêt de l'analyse de corpus pour les humanités numériques. Les raisons de trouver un équilibre entre analyse quantitative et linguistique de corpus sont abordées.Le troisième chapitre résume l'apport de la thèse en ce qui concerne la recherche sur les corpus tirés d'internet. La question de la collection des données est examinée avec une attention particulière, tout spécialement le cas des URLs sources. La notion de prétraitement des corpus web est introduite, ses étapes majeures sont brossées. L'impact des prétraitements sur le résultat est évalué. La question de la simplicité et de la reproducibilité de la construction de corpus est mise en avant.La quatrième partie décrit l'apport de la thèse du point de vue de la construction de corpus proprement dite, à travers la question des sources et le problèmes des documents invalides ou indésirables. Une approche utilisant un éclaireur léger pour préparer le parcours du web est présentée. Ensuite, les travaux concernant la sélection de documents juste avant l'inclusion dans un corpus sont résumés : il est possible d'utiliser les apports des études de lisibilité ainsi que des techniques d'apprentissage artificiel au cours de la construction du corpus. Un ensemble de caractéristiques textuelles testées sur des échantillons annotés évalue l'efficacité du procédé. Enfin, les travaux sur la visualisation de corpus sont abordés : extraction de caractéristiques à l'échelle d'un corpus afin de donner des indications sur sa composition et sa qualité

    Information Reliability on the Social Web - Models and Applications in Intelligent User Interfaces

    Get PDF
    The Social Web is undergoing continued evolution, changing the paradigm of information production, processing and sharing. Information sources have shifted from institutions to individual users, vastly increasing the amount of information available online. To overcome the information overload problem, modern filtering algorithms have enabled people to find relevant information in efficient ways. However, noisy, false and otherwise useless information remains a problem. We believe that the concept of information reliability needs to be considered along with information relevance to adapt filtering algorithms to today's Social Web. This approach helps to improve information search and discovery and can also improve user experience by communicating aspects of information reliability.This thesis first shows the results of a cross-disciplinary study into perceived reliability by reporting on a novel user experiment. This is followed by a discussion of modeling, validating, and communicating information reliability, including its various definitions across disciplines. A selection of important reliability attributes such as source credibility, competence, influence and timeliness are examined through different case studies. Results show that perceived reliability of information can vary greatly across contexts. Finally, recent studies on visual analytics, including algorithm explanations and interactive interfaces are discussed with respect to their impact on the perception of information reliability in a range of application domains
    corecore