3,353 research outputs found

    A Graph-Neural-Network-Based Social Network Recommendation Algorithm Using High-Order Neighbor Information

    Get PDF
    Social-network-based recommendation algorithms leverage rich social network information to alleviate the problem of data sparsity and boost the recommendation performance. However, traditional social-network-based recommendation algorithms ignore high-order collaborative signals or only consider the first-order collaborative signal when learning users’ and items’ latent representations, resulting in suboptimal recommendation performance. In this paper, we propose a graph neural network (GNN)-based social recommendation model that utilizes the GNN framework to capture high-order collaborative signals in the process of learning the latent representations of users and items. Specifically, we formulate the representations of entities, i.e., users and items, by stacking multiple embedding propagation layers to recursively aggregate multi-hop neighborhood information on both the user–item interaction graph and the social network graph. Hence, the collaborative signals hidden in both the user–item interaction graph and the social network graph are explicitly injected into the final representations of entities. Moreover, we ease the training process of the proposed GNN-based social recommendation model and alleviate overfitting by adopting a lightweight GNN framework that only retains the neighborhood aggregation component and abandons the feature transformation and nonlinear activation components. The experimental results on two real-world datasets show that our proposed GNN-based social recommendation method outperforms the state-of-the-art recommendation algorithms

    Personalized Video Recommendation Using Rich Contents from Videos

    Full text link
    Video recommendation has become an essential way of helping people explore the massive videos and discover the ones that may be of interest to them. In the existing video recommender systems, the models make the recommendations based on the user-video interactions and single specific content features. When the specific content features are unavailable, the performance of the existing models will seriously deteriorate. Inspired by the fact that rich contents (e.g., text, audio, motion, and so on) exist in videos, in this paper, we explore how to use these rich contents to overcome the limitations caused by the unavailability of the specific ones. Specifically, we propose a novel general framework that incorporates arbitrary single content feature with user-video interactions, named as collaborative embedding regression (CER) model, to make effective video recommendation in both in-matrix and out-of-matrix scenarios. Our extensive experiments on two real-world large-scale datasets show that CER beats the existing recommender models with any single content feature and is more time efficient. In addition, we propose a priority-based late fusion (PRI) method to gain the benefit brought by the integrating the multiple content features. The corresponding experiment shows that PRI brings real performance improvement to the baseline and outperforms the existing fusion methods

    Scale And Translation Invariant Collaborative Filtering Systems

    Get PDF
    Collaborative filtering systems are prediction algorithms over sparse data sets of user preferences. We modify a wide range of state-of-the-art collaborative filtering systems to make them scale and translation invariant and generally improve their accuracy without increasing their computational cost. Using the EachMovie and the Jester data sets, we show that learning-free constant time scale and translation invariant schemes outperforms other learning-free constant time schemes by at least 3% and perform as well as expensive memory-based schemes (within 4%). Over the Jester data set, we show that a scale and translation invariant Eigentaste algorithm outperforms Eigentaste 2.0 by 20%. These results suggest that scale and translation invariance is a desirable property

    CP-SLAM: Collaborative Neural Point-based SLAM System

    Full text link
    This paper presents a collaborative implicit neural simultaneous localization and mapping (SLAM) system with RGB-D image sequences, which consists of complete front-end and back-end modules including odometry, loop detection, sub-map fusion, and global refinement. In order to enable all these modules in a unified framework, we propose a novel neural point based 3D scene representation in which each point maintains a learnable neural feature for scene encoding and is associated with a certain keyframe. Moreover, a distributed-to-centralized learning strategy is proposed for the collaborative implicit SLAM to improve consistency and cooperation. A novel global optimization framework is also proposed to improve the system accuracy like traditional bundle adjustment. Experiments on various datasets demonstrate the superiority of the proposed method in both camera tracking and mapping.Comment: Accepted at NeurIPS 202

    Dynamic Sparse Learning: A Novel Paradigm for Efficient Recommendation

    Full text link
    In the realm of deep learning-based recommendation systems, the increasing computational demands, driven by the growing number of users and items, pose a significant challenge to practical deployment. This challenge is primarily twofold: reducing the model size while effectively learning user and item representations for efficient recommendations. Despite considerable advancements in model compression and architecture search, prevalent approaches face notable constraints. These include substantial additional computational costs from pre-training/re-training in model compression and an extensive search space in architecture design. Additionally, managing complexity and adhering to memory constraints is problematic, especially in scenarios with strict time or space limitations. Addressing these issues, this paper introduces a novel learning paradigm, Dynamic Sparse Learning (DSL), tailored for recommendation models. DSL innovatively trains a lightweight sparse model from scratch, periodically evaluating and dynamically adjusting each weight's significance and the model's sparsity distribution during the training. This approach ensures a consistent and minimal parameter budget throughout the full learning lifecycle, paving the way for "end-to-end" efficiency from training to inference. Our extensive experimental results underline DSL's effectiveness, significantly reducing training and inference costs while delivering comparable recommendation performance.Comment: 10 pages, 5 figures, 4 tables. Accecpted by WSDM 202

    Social Media Data Analysis Framework for Disaster Response

    Get PDF
    This paper presents a social media data analysis framework applied to multiple datasets. The method developed uses machine learning classifiers, where filtering binary classifiers based on deep bidirectional neural networks are trained on benchmark datasets of disaster responses for earthquakes and floods and extreme flood events. The classifiers consist of learning from discrete handcrafted features and fine-tuning approaches using deep bidirectional Transformer neural networks on these disaster response datasets. With the development of the multiclass classification approach, we compare the state-of-the-art results in one of the benchmark datasets containing the largest number of disaster-related categories. The multiclass classification approaches developed in this research with support vector machines provide a precision of 0.83 and 0.79 compared to Bernoulli naïve Bayes, which are 0.59 and 0.76, and multinomial naïve Bayes, which are 0.79 and 0.91, respectively. The binary classification methods based on the MDRM dataset show a higher precision with deep learning methods (DistilBERT) than BoW and TF-IDF, while in the case of UnifiedCEHMET dataset show a high performance for accuracy with the deep learning method in terms of severity, with a precision of 0.92 compared to BoW and TF-IDF method which has a precision of 0.68 and 0.70, respectively
    corecore