6 research outputs found

    Assessing Public Opinions Through Web 2.0: A Case Study on Wal-Mart

    Get PDF
    The recent advancement of Web 2.0 enables people to exchange their opinions on a variety of topics. Among these discussions, the opinions of employees, customers, and investors are of great interest to companies. Insight into such perspectives can help managers make better decisions on business policies and strategy. However, assessing online opinions is a nontrivial task. The high volume of messages, casual writing style, and the significant amount of noise require the application of sophisticated text mining techniques to digest the data. Previous research has successfully applied sentiment analysis to assess online opinions on specific items and topics. In this research, we propose the integration of topic analysis with sentiment analysis methods to assess the public opinions expressed in forums with diverse topics of discussion. Using a Wal- Mart-related Web forum as an example, we found that combining the two types of analysis can provide us with improved ability to assess public opinions on a company. Through further analysis on one cluster of discussions, several abnormal traffic and sentiment patterns were identified related to Wal-Mart events. The case study validates the propose framework as an IT artifact to assess online public opinion on companies of interest. Our research promotes further efforts to combine topic and sentiment analysis techniques in online research supporting business decision making

    Social Network - An autonomous system designed for radio recommendation.

    No full text
    International audienceThis paper describes the functions of a system proposed for the music tube recommendation from social network data base. Such a system enables the automatic collection, evaluation and rating of music critics, the possibility to rate music tube by auditors and the recommendation of tubes depended from auditor's pro les in form of regional internet radio. First, the system searches and retrieves probable music reviews from the Internet. Subsequently, the system carries out an evaluation and rating of those reviews. From this list of music tubes the system directly allows notation from our application. Finally the system automatically create the record list di used each day depended form the region, the year season, day hours and age of listeners. Our system uses linguistics and statistic methods for classifying music opinions and data mining techniques for recommendation part needed for recorded list creation. The principal task is the creation of popular intelligent radio adaptive on auditor's age and region - IA-Regional-Radio

    Brand Positioning Map and Analysis Using Web Scraping and Advertisement Analysis

    Get PDF
    There’s a significant increase in online consumer forums. When customers set out to buy a product they use these forums to form an opinion. Our research focuses on comparing Brand positioning maps based on consumer reviews. We also analyse the impact of advertisements and expert reviews. Our goal is to show that combining consumer reviews with ads and electronic media will help us analyze the effectiveness of advertising on brand positioning maps. This approach shall also help us in making association graphs for a brand using words of perception/opinion associated with that brand/product. Which may in turn assist companies in improving the focus of their advertisements to persuade the required set of crowd and influence the public perception

    Approximated Prompt Tuning for Vision-Language Pre-trained Models

    Full text link
    Prompt tuning is a parameter-efficient way to deploy large-scale pre-trained models to downstream tasks by adding task-specific tokens. In terms of vision-language pre-trained (VLP) models, prompt tuning often requires a large number of learnable tokens to bridge the gap between the pre-training and downstream tasks, which greatly exacerbates the already high computational overhead. In this paper, we revisit the principle of prompt tuning for Transformer-based VLP models and reveal that the impact of soft prompt tokens can be actually approximated via independent information diffusion steps, thereby avoiding the expensive global attention modeling and reducing the computational complexity to a large extent. Based on this finding, we propose a novel Approximated Prompt Tuning (APT) approach towards efficient VL transfer learning. To validate APT, we apply it to two representative VLP models, namely ViLT and METER, and conduct extensive experiments on a bunch of downstream tasks. Meanwhile, the generalization of APT is also validated on CLIP for image classification. The experimental results not only show the superior performance gains and computation efficiency of APT against the conventional prompt tuning methods, e.g., +6.6% accuracy and -64.62% additional computation overhead on METER, but also confirm its merits over other parameter-efficient transfer learning approaches

    Temporal patterns of happiness and information in a global social network: Hedonometrics and Twitter

    Get PDF
    Individual happiness is a fundamental societal metric. Normally measured through self-report, happiness has often been indirectly characterized and overshadowed by more readily quantifiable economic indicators such as gross domestic product. Here, we examine expressions made on the online, global microblog and social networking service Twitter, uncovering and explaining temporal variations in happiness and information levels over timescales ranging from hours to years. Our data set comprises over 46 billion words contained in nearly 4.6 billion expressions posted over a 33 month span by over 63 million unique users. In measuring happiness, we use a real-time, remote-sensing, non-invasive, text-based approach---a kind of hedonometer. In building our metric, made available with this paper, we conducted a survey to obtain happiness evaluations of over 10,000 individual words, representing a tenfold size improvement over similar existing word sets. Rather than being ad hoc, our word list is chosen solely by frequency of usage and we show how a highly robust metric can be constructed and defended.Comment: 27 pages, 17 figures, 3 tables. Supplementary Information: 1 table, 52 figure

    Domain-specific language models for multi-label classification of medical text

    Get PDF
    Recent advancements in machine learning-based medical text multi-label classifications can be used to enhance the understanding of the human body and aid the need for patient care. This research considers predicting medical codes from electronic health records (EHRs) as multi-label problems, where the number of labels ranged from 15 to 923. It is motivated by the advancements in domain-specific language models to better understand and represent electronic health records and improve the predictive accuracy of medical codes. The thesis presents an extensive empirical study of language models for binary and multi-label medical text classifications. Domain-specific multi-sourced fastText pre-trained embeddings are introduced. Experimental results show considerable improvements to predictive accuracy when such embeddings are used to represent medical text. It is shown that using domain-specific transformer models outperforms results for multi-label problems with fixed sequence length. If processing time is not an issue for a long medical text, then TransformerXL will be the best model to use. Experimental results show significant improvements over other models, including state-of-the-art results, when TransformerXL is used for down-streaming tasks such as predicting medical codes. The thesis considers concatenated language models to handle long medical documents and text data from multiple sources of EHRs. Experimental results show improvements in overall micro and macro F1 scores, and such improvements are achieved with fewer resources. In addition, it is shown that concatenated domain-specific transformers improve F1 scores of infrequent labels across several multi-label problems, especially with long-tail labels
    corecore