133 research outputs found
Parsec: a state channel for the Internet of Value
We propose Parsec, a web-scale State channel for the Internet of Value to exterminate the consensus bottleneck in Blockchain by leveraging a network of state channels which enable to robustly transfer value off-chain. It acts as an infrastructure layer developed on top of Ethereum Blockchain, as a network protocol which allows coherent routing and interlocking channel transfers for trade-off between parties. A web-scale solution for state channels is implemented to enable a layer of value transfer to the internet. Existing network protocol on State Channels include Raiden for Ethereum and Lightning Network for Bitcoin. However, we intend to leverage existing web-scale technologies used by large Internet companies such as Uber, LinkedIn or Netflix. We use Apache Kafka to scale the global payment operation to trillions of operations per day enabling near-instant, low-fee, scalable, and privacy-sustainable payments. Our architecture follows Event Sourcing pattern which solves current issues of payment solutions such as scaling, transfer, interoperability, low-fees, micropayments and to name a few. To the best of knowledge, our proposed model achieve better performance than state-of-the-art lightning network on the Ethereum based (fork) cryptocoins
A Model-Agnostic Framework for Recommendation via Interest-aware Item Embeddings
Item representation holds significant importance in recommendation systems,
which encompasses domains such as news, retail, and videos. Retrieval and
ranking models utilise item representation to capture the user-item
relationship based on user behaviours. While existing representation learning
methods primarily focus on optimising item-based mechanisms, such as attention
and sequential modelling. However, these methods lack a modelling mechanism to
directly reflect user interests within the learned item representations.
Consequently, these methods may be less effective in capturing user interests
indirectly. To address this challenge, we propose a novel Interest-aware
Capsule network (IaCN) recommendation model, a model-agnostic framework that
directly learns interest-oriented item representations. IaCN serves as an
auxiliary task, enabling the joint learning of both item-based and
interest-based representations. This framework adopts existing recommendation
models without requiring substantial redesign. We evaluate the proposed
approach on benchmark datasets, exploring various scenarios involving different
deep neural networks, behaviour sequence lengths, and joint learning ratios of
interest-oriented item representations. Experimental results demonstrate
significant performance enhancements across diverse recommendation models,
validating the effectiveness of our approach.Comment: Accepted Paper under LBR track in the Seventeenth ACM Conference on
Recommender Systems (RecSys) 202
Faster Algorithms for the Constrained k-Means Problem
The classical center based clustering problems such as k-means/median/center assume that the optimal clusters satisfy the locality property that the points in the same cluster are close to each other. A number of clustering problems arise in machine learning where the optimal clusters do not follow such a locality property. For instance, consider the r-gather clustering problem where there is an additional constraint that each of the clusters should have at least r points or the capacitated clustering problem where there is an upper bound on the cluster sizes. Consider a variant of the k-means problem that may be regarded as a general version of such problems. Here, the optimal clusters O_1, ..., O_k are an arbitrary partition of the dataset and the goal is to output k-centers c_1, ..., c_k such that the objective function sum_{i=1}^{k} sum_{x in O_{i}} ||x - c_{i}||^2 is minimized. It is not difficult to argue that any algorithm (without knowing the optimal clusters) that outputs a single set of k centers, will not behave well as far as optimizing the above objective function is concerned. However, this does not rule out the existence of algorithms that output a list of such k centers such that at least one of these k centers behaves well. Given an error parameter epsilon > 0, let l denote the size of the smallest list of k-centers such that at least one of the k-centers gives a (1+epsilon) approximation w.r.t. the objective function above. In this paper, we show an upper bound on l by giving a randomized algorithm that outputs a list of 2^{~O(k/epsilon)} k-centers. We also give a closely matching lower bound of 2^{~Omega(k/sqrt{epsilon})}. Moreover, our algorithm runs in time O(n * d * 2^{~O(k/epsilon)}). This is a significant improvement over the previous result of Ding and Xu who gave an algorithm with running time O(n * d * (log{n})^{k} * 2^{poly(k/epsilon)}) and output a list of size O((log{n})^k * 2^{poly(k/epsilon)}). Our techniques generalize for the k-median problem and for many other settings where non-Euclidean distance measures are involved
A simple D^2-sampling based PTAS for k-means and other Clustering Problems
Given a set of points , the -means clustering
problem is to find a set of {\em centers} such that the objective function ,
where denotes the distance between and the closest center in ,
is minimized. This is one of the most prominent objective functions that have
been studied with respect to clustering.
-sampling \cite{ArthurV07} is a simple non-uniform sampling technique
for choosing points from a set of points. It works as follows: given a set of
points , the first point is chosen uniformly at
random from . Subsequently, a point from is chosen as the next sample
with probability proportional to the square of the distance of this point to
the nearest previously sampled points.
-sampling has been shown to have nice properties with respect to the
-means clustering problem. Arthur and Vassilvitskii \cite{ArthurV07} show
that points chosen as centers from using -sampling gives an
approximation in expectation. Ailon et. al. \cite{AJMonteleoni09}
and Aggarwal et. al. \cite{AggarwalDK09} extended results of \cite{ArthurV07}
to show that points chosen as centers using -sampling give
approximation to the -means objective function with high probability. In
this paper, we further demonstrate the power of -sampling by giving a
simple randomized -approximation algorithm that uses the
-sampling in its core
Effects of foraging in personalized content-based image recommendation
A major challenge of recommender systems is to help users locating interesting items. Personalized recommender systems have become very popular as they attempt to predetermine the needs of users and provide them with recommendations to personalize their navigation. However, few studies have addressed the question of what drives the users' attention to specific content within the collection and what influences the selection of interesting items. To this end, we employ the lens of Information Foraging Theory (IFT) to image recommendation to demonstrate how the user could utilize visual bookmarks to locate interesting images. We investigate a personalized content-based image recommendation system to understand what affects user attention by reinforcing visual attention cues based on IFT. We further find that visual bookmarks (cues) lead to a stronger scent of the recommended image collection. Our evaluation is based on the Pinterest image collection
Approximate Clustering with Same-Cluster Queries
Ashtiani et al. proposed a Semi-Supervised Active Clustering framework (SSAC), where the learner is allowed to make adaptive queries to a domain expert. The queries are of the kind "do two given points belong to the same optimal cluster?", where the answers to these queries are assumed to be consistent with a unique optimal solution. There are many clustering contexts where such same cluster queries are feasible. Ashtiani et al. exhibited the power of such queries by showing that any instance of the k-means clustering problem, with additional margin assumption, can be solved efficiently if one is allowed to make O(k^2 log{k} + k log{n}) same-cluster queries. This is interesting since the k-means problem, even with the margin assumption, is NP-hard.
In this paper, we extend the work of Ashtiani et al. to the approximation setting by showing that a few of such same-cluster queries enables one to get a polynomial-time (1+eps)-approximation algorithm for the k-means problem without any margin assumption on the input dataset. Again, this is interesting since the k-means problem is NP-hard to approximate within a factor (1+c) for a fixed constant 0 < c < 1. The number of same-cluster queries used by the algorithm is poly(k/eps) which is independent of the size n of the dataset. Our algorithm is based on the D^2-sampling technique, also known as the k-means++ seeding algorithm. We also give a conditional lower bound on the number of same-cluster queries showing that if the Exponential Time Hypothesis (ETH) holds, then any such efficient query algorithm needs to make Omega (k/poly log k) same-cluster queries. Our algorithm can be extended for the case where the query answers are wrong with some bounded probability. Another result we show for the k-means++ seeding is that a small modification of the k-means++ seeding within the SSAC framework converts it to a constant factor approximation algorithm instead of the well known O(log k)-approximation algorithm
- …