4,013 research outputs found
The Use of Clustering Methods in Memory-Based Collaborative Filtering for Ranking-Based Recommendation Systems
This research explores the application of clustering techniques and frequency normalization in collaborative filtering to enhance the performance of ranking-based recommendation systems. Collaborative filtering is a popular approach in recommendation systems that relies on user-item interaction data. In ranking-based recommendation systems, the goal is to provide users with a personalized list of items, sorted by their predicted relevance. In this study, we propose a novel approach that combines clustering and frequency normalization techniques. Clustering, in the context of data analysis, is a technique used to organize and group together users or items that share similar characteristics or features. This method proves beneficial in enhancing recommendation accuracy by uncovering hidden patterns within the data. Additionally, frequency normalization is utilized to mitigate potential biases in user-item interaction data, ensuring fair and unbiased recommendations. The research methodology involves data preprocessing, clustering algorithm selection, frequency normalization techniques, and evaluation metrics. Experimental results demonstrate that the proposed method outperforms traditional collaborative filtering approaches in terms of ranking accuracy and recommendation quality. This approach has the potential to enhance recommendation systems across various domains, including e-commerce, content recommendation, and personalized advertising
A reinforcement learning recommender system using bi-clustering and Markov Decision Process
Collaborative filtering (CF) recommender systems are static in nature and does not adapt well with changing user preferences. User preferences may change after interaction with a system or after buying a product. Conventional CF clustering algorithms only identifies the distribution of patterns and hidden correlations globally. However, the impossibility of discovering local patterns by these algorithms, headed to the popularization of bi-clustering algorithms. Bi-clustering algorithms can analyze all dataset dimensions simultaneously and consequently, discover local patterns that deliver a better understanding of the underlying hidden correlations. In this paper, we modelled the recommendation problem as a sequential decision-making problem using Markov Decision Processes (MDP). To perform state representation for MDP, we first converted user-item votings matrix to a binary matrix. Then we performed bi-clustering on this binary matrix to determine a subset of similar rows and columns. A bi-cluster merging algorithm is designed to merge similar and overlapping bi-clusters. These bi-clusters are then mapped to a squared grid (SG). RL is applied on this SG to determine best policy to give recommendation to users. Start state is determined using Improved Triangle Similarity (ITR similarity measure. Reward function is computed as grid state overlapping in terms of users and items in current and prospective next state. A thorough comparative analysis was conducted, encompassing a diverse array of methodologies, including RL-based, pure Collaborative Filtering (CF), and clustering methods. The results demonstrate that our proposed method outperforms its competitors in terms of precision, recall, and optimal policy learning
Unleashing the power of artificial intelligence for climate action in industrial markets
Artificial Intelligence (AI) is a game-changing capability in industrial markets that can accelerate humanity's race against climate change. Positioned in a resource-hungry and pollution-intensive industry, this study explores AI-powered climate service innovation capabilities and their overall effects. The study develops and validates an AI model, identifying three primary dimensions and nine subdimensions. Based on a dataset in the fast fashion industry, the findings show that the AI-powered climate service innovation capabilities significantly influence both environmental and market performance, in which environmental performance acts as a partial mediator. Specifically, the results identify the key elements of an AI-informed framework for climate action and show how this can be used to develop a range of mitigation, adaptation and resilience initiatives in response to climate change
Explainable recommender with geometric information bottleneck
Explainable recommender systems can explain their recommendation decisions, enhancing user trust in the systems. Most explainable recommender systems either rely on human-annotated rationales to train models for explanation generation or leverage the attention mechanism to extract important text spans from reviews as explanations. The extracted rationales are often confined to an individual review and may fail to identify the implicit features beyond the review text. To avoid the expensive human annotation process and to generate explanations beyond individual reviews, we propose to incorporate a geometric prior learnt from user-item interactions into a variational network which infers latent factors from user-item reviews. The latent factors from an individual user-item pair can be used for both recommendation and explanation generation, which naturally inherit the global characteristics encoded in the prior knowledge. Experimental results on three e-commerce datasets show that our model significantly improves the interpretability of a variational recommender using the Wasserstein distance while achieving performance comparable to existing content-based recommender systems in terms of recommendation behaviours
Multidisciplinary perspectives on Artificial Intelligence and the law
This open access book presents an interdisciplinary, multi-authored, edited collection of chapters on Artificial Intelligence (âAIâ) and the Law. AI technology has come to play a central role in the modern data economy. Through a combination of increased computing power, the growing availability of data and the advancement of algorithms, AI has now become an umbrella term for some of the most transformational technological breakthroughs of this age. The importance of AI stems from both the opportunities that it offers and the challenges that it entails. While AI applications hold the promise of economic growth and efficiency gains, they also create significant risks and uncertainty. The potential and perils of AI have thus come to dominate modern discussions of technology and ethics â and although AI was initially allowed to largely develop without guidelines or rules, few would deny that the law is set to play a fundamental role in shaping the future of AI. As the debate over AI is far from over, the need for rigorous analysis has never been greater. This book thus brings together contributors from different fields and backgrounds to explore how the law might provide answers to some of the most pressing questions raised by AI. An outcome of the CatĂłlica Research Centre for the Future of Law and its interdisciplinary working group on Law and Artificial Intelligence, it includes contributions by leading scholars in the fields of technology, ethics and the law.info:eu-repo/semantics/publishedVersio
Fuzzy Norm-Explicit Product Quantization for Recommender Systems
As the data resources grow, providing recommendations that best meet the demands has become a vital requirement in business and life to overcome the information overload problem. However, building a system suggesting relevant recommendations has always been a point of debate. One of the most cost-efficient techniques in terms of producing relevant recommendations at a low complexity is Product Quantization (PQ). PQ approaches have continued developing in recent years. This systemâs crucial challenge is improving product quantization performance in terms of recall measures without compromising its complexity. This makes the algorithm suitable for problems that require a greater number of potentially relevant items without disregarding others, at high-speed and low-cost to keep up with traffic. This is the case of online shops where the recommendations for the purpose are important, although customers can be susceptible to scoping other products. A recent approach has been exploiting the notion of norm sub-vectors encoded in product quantizers. This research proposes a fuzzy approach to perform norm-based product quantization. Type-2 Fuzzy sets (T2FSs) define the codebook allowing sub-vectors (T2FSs) to be associated with more than one element of the codebook, and next, its norm calculus is resolved by means of integration. Our method finesses the recall measure up, making the algorithm suitable for problems that require querying at most possible potential relevant items without disregarding others. The proposed approach is tested with three public recommender benchmark datasets and compared against seven PQ approaches for Maximum Inner-Product Search (MIPS). The proposed method outperforms all PQ approaches such as NEQ, PQ, and RQ up to +6%, +5%, and +8% by achieving a recall of 94%, 69%, 59% in Netflix, Audio, Cifar60k datasets, respectively. More and over, computing time and complexity nearly equals the most computationally efficient existing PQ method in the state-of-the-art
Advances and Challenges of Multi-task Learning Method in Recommender System: A Survey
Multi-task learning has been widely applied in computational vision, natural
language processing and other fields, which has achieved well performance. In
recent years, a lot of work about multi-task learning recommender system has
been yielded, but there is no previous literature to summarize these works. To
bridge this gap, we provide a systematic literature survey about multi-task
recommender systems, aiming to help researchers and practitioners quickly
understand the current progress in this direction. In this survey, we first
introduce the background and the motivation of the multi-task learning-based
recommender systems. Then we provide a taxonomy of multi-task learning-based
recommendation methods according to the different stages of multi-task learning
techniques, which including task relationship discovery, model architecture and
optimization strategy. Finally, we raise discussions on the application and
promising future directions in this area
A Comparative Analysis of Collaborative Filtering Similarity Measurements for Recommendation Systems
Collaborative Filtering (CF) is a widely used technique in recommendation systems to suggest items to users based on their previous interactions with the system. CF involves finding correlations between the preferences of different users and using those correlations to provide recommendations. This technique can be divided into user-based and item-based CF, both of which utilize similarity metrics to generate recommendations. Content-based filtering is another commonly used recommendation technique that analyzes the attributes of items to suggest similar items. To enhance the accuracy of recommendation systems, hybrid algorithms that combine CF and content-based filtering techniques have been developed. These hybrid systems leverage the strengths of both approaches to provide more accurate and personalized recommendations. In conclusion, collaborative filtering is an essential technique in recommendation systems, and the use of various similarity metrics and hybrid techniques can enhance the quality of recommendations
La traduzione specializzata allâopera per una piccola impresa in espansione: la mia esperienza di internazionalizzazione in cinese di Bioretics© S.r.l.
Global markets are currently immersed in two all-encompassing and unstoppable processes: internationalization and globalization. While the former pushes companies to look beyond the borders of their country of origin to forge relationships with foreign trading partners, the latter fosters the standardization in all countries, by reducing spatiotemporal distances and breaking down geographical, political, economic and socio-cultural barriers. In recent decades, another domain has appeared to propel these unifying drives: Artificial Intelligence, together with its high technologies aiming to implement human cognitive abilities in machinery. The âLanguage Toolkit â Le lingue straniere al servizio dellâinternazionalizzazione dellâimpresaâ project, promoted by the Department of Interpreting and Translation (ForlĂŹ Campus) in collaboration with the Romagna Chamber of Commerce (ForlĂŹ-Cesena and Rimini), seeks to help Italian SMEs make their way into the global market. It is precisely within this project that this dissertation has been conceived. Indeed, its purpose is to present the translation and localization project from English into Chinese of a series of texts produced by Bioretics© S.r.l.: an investor deck, the company website and part of the installation and use manual of the Aliquis© framework software, its flagship product. This dissertation is structured as follows: Chapter 1 presents the project and the company in detail; Chapter 2 outlines the internationalization and globalization processes and the Artificial Intelligence market both in Italy and in China; Chapter 3 provides the theoretical foundations for every aspect related to Specialized Translation, including website localization; Chapter 4 describes the resources and tools used to perform the translations; Chapter 5 proposes an analysis of the source texts; Chapter 6 is a commentary on translation strategies and choices
Mining Butterflies in Streaming Graphs
This thesis introduces two main-memory systems sGrapp and sGradd for performing the fundamental analytic tasks of biclique counting and concept drift detection over a streaming graph. A data-driven heuristic is used to architect the systems. To this end, initially, the growth patterns of bipartite streaming graphs are mined and the emergence principles of streaming motifs are discovered. Next, the discovered principles are (a) explained by a graph generator called sGrow; and (b) utilized to establish the requirements for efficient, effective, explainable, and interpretable management and processing of streams. sGrow is used to benchmark stream analytics, particularly in the case of concept drift detection.
sGrow displays robust realization of streaming growth patterns independent of initial conditions, scale and temporal characteristics, and model configurations. Extensive evaluations confirm the simultaneous effectiveness and efficiency of sGrapp and sGradd. sGrapp achieves mean absolute percentage error up to 0.05/0.14 for the cumulative butterfly count in streaming graphs with uniform/non-uniform temporal distribution and a processing throughput of 1.5 million data records per second. The throughput and estimation error of sGrapp are 160x higher and 0.02x lower than baselines. sGradd demonstrates an improving performance over time, achieves zero false detection rates when there is not any drift and when drift is already detected, and detects sequential drifts in zero to a few seconds after their occurrence regardless of drift intervals
- âŠ