196,120 research outputs found
Sentiment without Sentiment Analysis: Using the Recommendation Outcome of Steam Game Reviews as Sentiment Predictor
This paper presents and explores a novel way to determine the sentiment of a Steam game review based on the predicted recommendation of the review, testing different regression models on a combination of Term Frequency-Inverse Document Frequency (TF-IDF) and Latent Dirichlet Allocation (LDA) features. A dataset of Steam game reviews extracted from the Programming games genre consisting of 21 games along with other significant features such as the number of helpful likes on the recommendation, number of hours played, and others. Based on the features, they are grouped into three datasets: 1) either having keyword features only, 2) keyword features with the numerical features, and 3) numerical features only. The three datasets were trained using five different regression models: Multilinear Regression, Lasso Regression, Ridge Regression, Support Vector Regression, and Multi-layer Perceptron Regression, which were then evaluated using RMSE, MAE, and MAPE. The review recommendation was predicted from each model, and the accuracy of the predictions were measured using the different error rates. The results of this research may prove helpful in the convergence of Machine Learning and Educational Games
Statistical and machine learning methods evaluated for incorporating soil and weather into corn nitrogen recommendations
Nitrogen (N) fertilizer recommendation tools could be improved for estimating corn (Zea mays L.) N needs by incorporating site-specific soil and weather information. However, an evaluation of analytical methods is needed to determine the success of incorporating this information. The objectives of this research were to evaluate statistical and machine learning (ML) algorithms for utilizing soil and weather information for improving corn N recommendation tools. Eight algorithms [stepwise, ridge regression, least absolute shrinkage and selection operator (Lasso), elastic net regression, principal component regression (PCR), partial least squares regression (PLSR), decision tree, and random forest] were evaluated using a dataset containing measured soil and weather variables from a regional database. The performance was evaluated based on how well these algorithms predicted corn economically optimal N rates (EONR) from 49 sites in the U.S. Midwest. Multiple algorithm modeling scenarios were examined with and without adjustment for multicollinearity and inclusion of two-way interaction terms to identify the soil and weather variables that could improve three dissimilar N recommendation tools. Results showed the out-of-sample root-mean-square error (RMSE) for the decision tree and some random forest modeling scenarios were better than the stepwise or ridge regression, but not significantly different than any other algorithm. The best ML algorithm for adjusting N recommendation tools was the random forest approach (r2 increased between 0.72 and 0.84 and the RMSE decreased between 41 and 94 kg N ha−1). However, the ML algorithm that best adjusted tools while using a minimal amount of variables was the decision tree. This method was simple, needing only one or two variables (regardless of modeling scenario) and provided moderate improvement as r2 values increased between 0.15 and 0.51 and RMSE decreased between 16 and 66 kg N ha−1. Using ML algorithms to adjust N recommendation tools with soil and weather information shows promising results for better N management in the U.S. Midwest
Statistical and machine learning methods evaluated for incorporating soil and weather into corn nitrogen recommendations
Nitrogen (N) fertilizer recommendation tools could be improved for estimating corn (Zea mays L.) N needs by incorporating site-specific soil and weather information. However, an evaluation of analytical methods is needed to determine the success of incorporating this information. The objectives of this research were to evaluate statistical and machine learning (ML) algorithms for utilizing soil and weather information for improving corn N recommendation tools. Eight algorithms [stepwise, ridge regression, least absolute shrinkage and selection operator (Lasso), elastic net regression, principal component regression (PCR), partial least squares regression (PLSR), decision tree, and random forest] were evaluated using a dataset containing measured soil and weather variables from a regional database. The performance was evaluated based on how well these algorithms predicted corn economically optimal N rates (EONR) from 49 sites in the U.S. Midwest. Multiple algorithm modeling scenarios were examined with and without adjustment for multicollinearity and inclusion of two-way interaction terms to identify the soil and weather variables that could improve three dissimilar N recommendation tools. Results showed the out-of-sample root-mean-square error (RMSE) for the decision tree and some random forest modeling scenarios were better than the stepwise or ridge regression, but not significantly different than any other algorithm. The best ML algorithm for adjusting N recommendation tools was the random forest approach (r2 increased between 0.72 and 0.84 and the RMSE decreased between 41 and 94 kg N ha−1). However, the ML algorithm that best adjusted tools while using a minimal amount of variables was the decision tree. This method was simple, needing only one or two variables (regardless of modeling scenario) and provided moderate improvement as r2 values increased between 0.15 and 0.51 and RMSE decreased between 16 and 66 kg N ha−1. Using ML algorithms to adjust N recommendation tools with soil and weather information shows promising results for better N management in the U.S. Midwest
F1000 recommendations as a new data source for research evaluation: A comparison with citations
F1000 is a post-publication peer review service for biological and medical
research. F1000 aims to recommend important publications in the biomedical
literature, and from this perspective F1000 could be an interesting tool for
research evaluation. By linking the complete database of F1000 recommendations
to the Web of Science bibliographic database, we are able to make a
comprehensive comparison between F1000 recommendations and citations. We find
that about 2% of the publications in the biomedical literature receive at least
one F1000 recommendation. Recommended publications on average receive 1.30
recommendations, and over 90% of the recommendations are given within half a
year after a publication has appeared. There turns out to be a clear
correlation between F1000 recommendations and citations. However, the
correlation is relatively weak, at least weaker than the correlation between
journal impact and citations. More research is needed to identify the main
reasons for differences between recommendations and citations in assessing the
impact of publications
Data Portraits and Intermediary Topics: Encouraging Exploration of Politically Diverse Profiles
In micro-blogging platforms, people connect and interact with others.
However, due to cognitive biases, they tend to interact with like-minded people
and read agreeable information only. Many efforts to make people connect with
those who think differently have not worked well. In this paper, we
hypothesize, first, that previous approaches have not worked because they have
been direct -- they have tried to explicitly connect people with those having
opposing views on sensitive issues. Second, that neither recommendation or
presentation of information by themselves are enough to encourage behavioral
change. We propose a platform that mixes a recommender algorithm and a
visualization-based user interface to explore recommendations. It recommends
politically diverse profiles in terms of distance of latent topics, and
displays those recommendations in a visual representation of each user's
personal content. We performed an "in the wild" evaluation of this platform,
and found that people explored more recommendations when using a biased
algorithm instead of ours. In line with our hypothesis, we also found that the
mixture of our recommender algorithm and our user interface, allowed
politically interested users to exhibit an unbiased exploration of the
recommended profiles. Finally, our results contribute insights in two aspects:
first, which individual differences are important when designing platforms
aimed at behavioral change; and second, which algorithms and user interfaces
should be mixed to help users avoid cognitive mechanisms that lead to biased
behavior.Comment: 12 pages, 7 figures. To be presented at ACM Intelligent User
Interfaces 201
- …