196,106 research outputs found

    Sentiment without Sentiment Analysis: Using the Recommendation Outcome of Steam Game Reviews as Sentiment Predictor

    Get PDF
    This paper presents and explores a novel way to determine the sentiment of a Steam game review based on the predicted recommendation of the review, testing different regression models on a combination of Term Frequency-Inverse Document Frequency (TF-IDF) and Latent Dirichlet Allocation (LDA) features. A dataset of Steam game reviews extracted from the Programming games genre consisting of 21 games along with other significant features such as the number of helpful likes on the recommendation, number of hours played, and others. Based on the features, they are grouped into three datasets: 1) either having keyword features only, 2) keyword features with the numerical features, and 3) numerical features only. The three datasets were trained using five different regression models: Multilinear Regression, Lasso Regression, Ridge Regression, Support Vector Regression, and Multi-layer Perceptron Regression, which were then evaluated using RMSE, MAE, and MAPE. The review recommendation was predicted from each model, and the accuracy of the predictions were measured using the different error rates. The results of this research may prove helpful in the convergence of Machine Learning and Educational Games

    Statistical and machine learning methods evaluated for incorporating soil and weather into corn nitrogen recommendations

    Get PDF
    Nitrogen (N) fertilizer recommendation tools could be improved for estimating corn (Zea mays L.) N needs by incorporating site-specific soil and weather information. However, an evaluation of analytical methods is needed to determine the success of incorporating this information. The objectives of this research were to evaluate statistical and machine learning (ML) algorithms for utilizing soil and weather information for improving corn N recommendation tools. Eight algorithms [stepwise, ridge regression, least absolute shrinkage and selection operator (Lasso), elastic net regression, principal component regression (PCR), partial least squares regression (PLSR), decision tree, and random forest] were evaluated using a dataset containing measured soil and weather variables from a regional database. The performance was evaluated based on how well these algorithms predicted corn economically optimal N rates (EONR) from 49 sites in the U.S. Midwest. Multiple algorithm modeling scenarios were examined with and without adjustment for multicollinearity and inclusion of two-way interaction terms to identify the soil and weather variables that could improve three dissimilar N recommendation tools. Results showed the out-of-sample root-mean-square error (RMSE) for the decision tree and some random forest modeling scenarios were better than the stepwise or ridge regression, but not significantly different than any other algorithm. The best ML algorithm for adjusting N recommendation tools was the random forest approach (r2 increased between 0.72 and 0.84 and the RMSE decreased between 41 and 94 kg N ha−1). However, the ML algorithm that best adjusted tools while using a minimal amount of variables was the decision tree. This method was simple, needing only one or two variables (regardless of modeling scenario) and provided moderate improvement as r2 values increased between 0.15 and 0.51 and RMSE decreased between 16 and 66 kg N ha−1. Using ML algorithms to adjust N recommendation tools with soil and weather information shows promising results for better N management in the U.S. Midwest

    Statistical and machine learning methods evaluated for incorporating soil and weather into corn nitrogen recommendations

    Get PDF
    Nitrogen (N) fertilizer recommendation tools could be improved for estimating corn (Zea mays L.) N needs by incorporating site-specific soil and weather information. However, an evaluation of analytical methods is needed to determine the success of incorporating this information. The objectives of this research were to evaluate statistical and machine learning (ML) algorithms for utilizing soil and weather information for improving corn N recommendation tools. Eight algorithms [stepwise, ridge regression, least absolute shrinkage and selection operator (Lasso), elastic net regression, principal component regression (PCR), partial least squares regression (PLSR), decision tree, and random forest] were evaluated using a dataset containing measured soil and weather variables from a regional database. The performance was evaluated based on how well these algorithms predicted corn economically optimal N rates (EONR) from 49 sites in the U.S. Midwest. Multiple algorithm modeling scenarios were examined with and without adjustment for multicollinearity and inclusion of two-way interaction terms to identify the soil and weather variables that could improve three dissimilar N recommendation tools. Results showed the out-of-sample root-mean-square error (RMSE) for the decision tree and some random forest modeling scenarios were better than the stepwise or ridge regression, but not significantly different than any other algorithm. The best ML algorithm for adjusting N recommendation tools was the random forest approach (r2 increased between 0.72 and 0.84 and the RMSE decreased between 41 and 94 kg N ha−1). However, the ML algorithm that best adjusted tools while using a minimal amount of variables was the decision tree. This method was simple, needing only one or two variables (regardless of modeling scenario) and provided moderate improvement as r2 values increased between 0.15 and 0.51 and RMSE decreased between 16 and 66 kg N ha−1. Using ML algorithms to adjust N recommendation tools with soil and weather information shows promising results for better N management in the U.S. Midwest

    F1000 recommendations as a new data source for research evaluation: A comparison with citations

    Get PDF
    F1000 is a post-publication peer review service for biological and medical research. F1000 aims to recommend important publications in the biomedical literature, and from this perspective F1000 could be an interesting tool for research evaluation. By linking the complete database of F1000 recommendations to the Web of Science bibliographic database, we are able to make a comprehensive comparison between F1000 recommendations and citations. We find that about 2% of the publications in the biomedical literature receive at least one F1000 recommendation. Recommended publications on average receive 1.30 recommendations, and over 90% of the recommendations are given within half a year after a publication has appeared. There turns out to be a clear correlation between F1000 recommendations and citations. However, the correlation is relatively weak, at least weaker than the correlation between journal impact and citations. More research is needed to identify the main reasons for differences between recommendations and citations in assessing the impact of publications

    Data Portraits and Intermediary Topics: Encouraging Exploration of Politically Diverse Profiles

    Full text link
    In micro-blogging platforms, people connect and interact with others. However, due to cognitive biases, they tend to interact with like-minded people and read agreeable information only. Many efforts to make people connect with those who think differently have not worked well. In this paper, we hypothesize, first, that previous approaches have not worked because they have been direct -- they have tried to explicitly connect people with those having opposing views on sensitive issues. Second, that neither recommendation or presentation of information by themselves are enough to encourage behavioral change. We propose a platform that mixes a recommender algorithm and a visualization-based user interface to explore recommendations. It recommends politically diverse profiles in terms of distance of latent topics, and displays those recommendations in a visual representation of each user's personal content. We performed an "in the wild" evaluation of this platform, and found that people explored more recommendations when using a biased algorithm instead of ours. In line with our hypothesis, we also found that the mixture of our recommender algorithm and our user interface, allowed politically interested users to exhibit an unbiased exploration of the recommended profiles. Finally, our results contribute insights in two aspects: first, which individual differences are important when designing platforms aimed at behavioral change; and second, which algorithms and user interfaces should be mixed to help users avoid cognitive mechanisms that lead to biased behavior.Comment: 12 pages, 7 figures. To be presented at ACM Intelligent User Interfaces 201
    • …
    corecore