6 research outputs found
Assessing Public Opinions Through Web 2.0: A Case Study on Wal-Mart
The recent advancement of Web 2.0 enables people to exchange their opinions on a variety of topics. Among these discussions, the opinions of employees, customers, and investors are of great interest to companies. Insight into such perspectives can help managers make better decisions on business policies and strategy. However, assessing online opinions is a nontrivial task. The high volume of messages, casual writing style, and the significant amount of noise require the application of sophisticated text mining techniques to digest the data. Previous research has successfully applied sentiment analysis to assess online opinions on specific items and topics. In this research, we propose the integration of topic analysis with sentiment analysis methods to assess the public opinions expressed in forums with diverse topics of discussion. Using a Wal- Mart-related Web forum as an example, we found that combining the two types of analysis can provide us with improved ability to assess public opinions on a company. Through further analysis on one cluster of discussions, several abnormal traffic and sentiment patterns were identified related to Wal-Mart events. The case study validates the propose framework as an IT artifact to assess online public opinion on companies of interest. Our research promotes further efforts to combine topic and sentiment analysis techniques in online research supporting business decision making
Social Network - An autonomous system designed for radio recommendation.
International audienceThis paper describes the functions of a system proposed for the music tube recommendation from social network data base. Such a system enables the automatic collection, evaluation and rating of music critics, the possibility to rate music tube by auditors and the recommendation of tubes depended from auditor's pro les in form of regional internet radio. First, the system searches and retrieves probable music reviews from the Internet. Subsequently, the system carries out an evaluation and rating of those reviews. From this list of music tubes the system directly allows notation from our application. Finally the system automatically create the record list di used each day depended form the region, the year season, day hours and age of listeners. Our system uses linguistics and statistic methods for classifying music opinions and data mining techniques for recommendation part needed for recorded list creation. The principal task is the creation of popular intelligent radio adaptive on auditor's age and region - IA-Regional-Radio
Brand Positioning Map and Analysis Using Web Scraping and Advertisement Analysis
There’s a significant increase in online consumer forums. When customers set out to buy a product they use these forums to form an opinion. Our research focuses on comparing Brand positioning maps based on consumer reviews. We also analyse the impact of advertisements and expert reviews. Our goal is to show that combining consumer reviews with ads and electronic media will help us analyze the effectiveness of advertising on brand positioning maps. This approach shall also help us in making association graphs for a brand using words of perception/opinion associated with that brand/product. Which may in turn assist companies in improving the focus of their advertisements to persuade the required set of crowd and influence the public perception
Approximated Prompt Tuning for Vision-Language Pre-trained Models
Prompt tuning is a parameter-efficient way to deploy large-scale pre-trained
models to downstream tasks by adding task-specific tokens. In terms of
vision-language pre-trained (VLP) models, prompt tuning often requires a large
number of learnable tokens to bridge the gap between the pre-training and
downstream tasks, which greatly exacerbates the already high computational
overhead. In this paper, we revisit the principle of prompt tuning for
Transformer-based VLP models and reveal that the impact of soft prompt tokens
can be actually approximated via independent information diffusion steps,
thereby avoiding the expensive global attention modeling and reducing the
computational complexity to a large extent. Based on this finding, we propose a
novel Approximated Prompt Tuning (APT) approach towards efficient VL transfer
learning. To validate APT, we apply it to two representative VLP models, namely
ViLT and METER, and conduct extensive experiments on a bunch of downstream
tasks. Meanwhile, the generalization of APT is also validated on CLIP for image
classification. The experimental results not only show the superior performance
gains and computation efficiency of APT against the conventional prompt tuning
methods, e.g., +6.6% accuracy and -64.62% additional computation overhead on
METER, but also confirm its merits over other parameter-efficient transfer
learning approaches
Temporal patterns of happiness and information in a global social network: Hedonometrics and Twitter
Individual happiness is a fundamental societal metric. Normally measured
through self-report, happiness has often been indirectly characterized and
overshadowed by more readily quantifiable economic indicators such as gross
domestic product. Here, we examine expressions made on the online, global
microblog and social networking service Twitter, uncovering and explaining
temporal variations in happiness and information levels over timescales ranging
from hours to years. Our data set comprises over 46 billion words contained in
nearly 4.6 billion expressions posted over a 33 month span by over 63 million
unique users. In measuring happiness, we use a real-time, remote-sensing,
non-invasive, text-based approach---a kind of hedonometer. In building our
metric, made available with this paper, we conducted a survey to obtain
happiness evaluations of over 10,000 individual words, representing a tenfold
size improvement over similar existing word sets. Rather than being ad hoc, our
word list is chosen solely by frequency of usage and we show how a highly
robust metric can be constructed and defended.Comment: 27 pages, 17 figures, 3 tables. Supplementary Information: 1 table,
52 figure
Domain-specific language models for multi-label classification of medical text
Recent advancements in machine learning-based medical text multi-label classifications can be used to enhance the understanding of the human body and aid the need for patient care. This research considers predicting medical codes from electronic health records (EHRs) as multi-label problems, where the number of labels ranged from 15 to 923. It is motivated by the advancements in domain-specific language models to better understand and represent electronic health records and improve the predictive accuracy of medical codes.
The thesis presents an extensive empirical study of language models for binary and multi-label medical text classifications. Domain-specific multi-sourced fastText pre-trained embeddings are introduced. Experimental results show considerable improvements to predictive accuracy when such embeddings are used to represent medical text. It is shown that using domain-specific transformer models outperforms results for multi-label problems with fixed sequence length. If processing time is not an issue for a long medical text, then TransformerXL will be the best model to use. Experimental results show significant improvements over other models, including state-of-the-art results, when TransformerXL is used for down-streaming tasks such as predicting medical codes.
The thesis considers concatenated language models to handle long medical documents and text data from multiple sources of EHRs. Experimental results show improvements in overall micro and macro F1 scores, and such improvements are achieved with fewer resources. In addition, it is shown that concatenated domain-specific transformers improve F1 scores of infrequent labels across several multi-label problems, especially with long-tail labels