368 research outputs found
Replace Scoring with Arrangement: A Contextual Set-to-Arrangement Framework for Learning-to-Rank
Learning-to-rank is a core technique in the top-N recommendation task, where
an ideal ranker would be a mapping from an item set to an arrangement (a.k.a.
permutation). Most existing solutions fall in the paradigm of probabilistic
ranking principle (PRP), i.e., first score each item in the candidate set and
then perform a sort operation to generate the top ranking list. However, these
approaches neglect the contextual dependence among candidate items during
individual scoring, and the sort operation is non-differentiable. To bypass the
above issues, we propose Set-To-Arrangement Ranking (STARank), a new framework
directly generates the permutations of the candidate items without the need for
individually scoring and sort operations; and is end-to-end differentiable. As
a result, STARank can operate when only the ground-truth permutations are
accessible without requiring access to the ground-truth relevance scores for
items. For this purpose, STARank first reads the candidate items in the context
of the user browsing history, whose representations are fed into a
Plackett-Luce module to arrange the given items into a list. To effectively
utilize the given ground-truth permutations for supervising STARank, we
leverage the internal consistency property of Plackett-Luce models to derive a
computationally efficient list-wise loss. Experimental comparisons against 9
the state-of-the-art methods on 2 learning-to-rank benchmark datasets and 3
top-N real-world recommendation datasets demonstrate the superiority of STARank
in terms of conventional ranking metrics. Notice that these ranking metrics do
not consider the effects of the contextual dependence among the items in the
list, we design a new family of simulation-based ranking metrics, where
existing metrics can be regarded as special cases. STARank can consistently
achieve better performance in terms of PBM and UBM simulation-based metrics.Comment: CIKM 202
ReLLa: Retrieval-enhanced Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation
With large language models (LLMs) achieving remarkable breakthroughs in
natural language processing (NLP) domains, LLM-enhanced recommender systems
have received much attention and have been actively explored currently. In this
paper, we focus on adapting and empowering a pure large language model for
zero-shot and few-shot recommendation tasks. First and foremost, we identify
and formulate the lifelong sequential behavior incomprehension problem for LLMs
in recommendation domains, i.e., LLMs fail to extract useful information from a
textual context of long user behavior sequence, even if the length of context
is far from reaching the context limitation of LLMs. To address such an issue
and improve the recommendation performance of LLMs, we propose a novel
framework, namely Retrieval-enhanced Large Language models (ReLLa) for
recommendation tasks in both zero-shot and few-shot settings. For zero-shot
recommendation, we perform semantic user behavior retrieval (SUBR) to improve
the data quality of testing samples, which greatly reduces the difficulty for
LLMs to extract the essential knowledge from user behavior sequences. As for
few-shot recommendation, we further design retrieval-enhanced instruction
tuning (ReiT) by adopting SUBR as a data augmentation technique for training
samples. Specifically, we develop a mixed training dataset consisting of both
the original data samples and their retrieval-enhanced counterparts. We conduct
extensive experiments on a real-world public dataset (i.e., MovieLens-1M) to
demonstrate the superiority of ReLLa compared with existing baseline models, as
well as its capability for lifelong sequential behavior comprehension.Comment: Under Revie
Transgenically mediated shRNAs targeting conserved regions of foot-and-mouth disease virus provide heritable resistance in porcine cell lines and suckling mice
CodeApex: A Bilingual Programming Evaluation Benchmark for Large Language Models
With the emergence of Large Language Models (LLMs), there has been a
significant improvement in the programming capabilities of models, attracting
growing attention from researchers. We propose CodeApex, a bilingual benchmark
dataset focusing on the programming comprehension and code generation abilities
of LLMs. CodeApex comprises three types of multiple-choice questions:
conceptual understanding, commonsense reasoning, and multi-hop reasoning,
designed to evaluate LLMs on programming comprehension tasks. Additionally,
CodeApex utilizes algorithmic questions and corresponding test cases to assess
the code quality generated by LLMs. We evaluate 14 state-of-the-art LLMs,
including both general-purpose and specialized models. GPT exhibits the best
programming capabilities, achieving approximate accuracies of 50% and 56% on
the two tasks, respectively. There is still significant room for improvement in
programming tasks. We hope that CodeApex can serve as a reference for
evaluating the coding capabilities of LLMs, further promoting their development
and growth. Datasets are released at https://github.com/APEXLAB/CodeApex.git.
CodeApex submission website is https://apex.sjtu.edu.cn/codeapex/.Comment: 21 page
Optimal Reaction Coordinates
The dynamic behavior of complex systems with many degrees of freedom is often analyzed by projection onto one or a few reaction coordinates. The dynamics is then described in a simple and intuitive way as diffusion on the associated free energy pro le. In order to use such a picture for a quantitative description of the dynamics one needs to select the coordinate in an optimal way so as to minimize non-Markovian effects due to the projection. For equilibrium dynamics between two boundary states (e.g., a reaction) the optimal coordinate is known as the committor or the pfold coordinate in protein folding studies. While the dynamics projected on the committor is not Markovian, many important quantities of the original multidimensional dynamics on an arbitrarily complex landscape can be computed exactly. Here we summarize the derivation of this result, discuss different approaches to determine and validate the committor coordinate and present three illustrative applications: protein folding, the game of chess, and patient recovery dynamics after kidney transplant
HIF drives lipid deposition and cancer in ccRCC via repression of fatty acid metabolism
Clear cell renal cell carcinoma (ccRCC) is histologically defined by its lipid and glycogen-rich cytoplasmic deposits. Alterations in the VHL tumor suppressor stabilizing the hypoxiainducible factors (HIFs) are the most prevalent molecular features of clear cell tumors. The significance of lipid deposition remains undefined. We describe the mechanism of lipid deposition in ccRCC by identifying the rate-limiting component of mitochondrial fatty acid transport, carnitine palmitoyltransferase 1A (CPT1A), as a direct HIF target gene. CPT1A is repressed by HIF1 and HIF2, reducing fatty acid transport into the mitochondria, and forcing fatty acids to lipid droplets for storage. Droplet formation occurs independent of lipid source, but only when CPT1A is repressed. Functionally, repression of CPT1A is critical for tumor formation, as elevated CPT1A expression limits tumor growth. In human tumors, CPT1A expression and activity are decreased versus normal kidney; and poor patient outcome associates with lower expression of CPT1A in tumors in TCGA. Together, our studies identify HIF control of fatty acid metabolism as essential for ccRCC tumorigenesis
Community detection in graphs
The modern science of networks has brought significant advances to our
understanding of complex systems. One of the most relevant features of graphs
representing real systems is community structure, or clustering, i. e. the
organization of vertices in clusters, with many edges joining vertices of the
same cluster and comparatively few edges joining vertices of different
clusters. Such clusters, or communities, can be considered as fairly
independent compartments of a graph, playing a similar role like, e. g., the
tissues or the organs in the human body. Detecting communities is of great
importance in sociology, biology and computer science, disciplines where
systems are often represented as graphs. This problem is very hard and not yet
satisfactorily solved, despite the huge effort of a large interdisciplinary
community of scientists working on it over the past few years. We will attempt
a thorough exposition of the topic, from the definition of the main elements of
the problem, to the presentation of most methods developed, with a special
focus on techniques designed by statistical physicists, from the discussion of
crucial issues like the significance of clustering and how methods should be
tested and compared against each other, to the description of applications to
real networks.Comment: Review article. 103 pages, 42 figures, 2 tables. Two sections
expanded + minor modifications. Three figures + one table + references added.
Final version published in Physics Report
Optimasi Portofolio Resiko Menggunakan Model Markowitz MVO Dikaitkan dengan Keterbatasan Manusia dalam Memprediksi Masa Depan dalam Perspektif Al-Qur`an
Risk portfolio on modern finance has become increasingly technical, requiring the use of sophisticated mathematical tools in both research and practice. Since companies cannot insure themselves completely against risk, as human incompetence in predicting the future precisely that written in Al-Quran surah Luqman verse 34, they have to manage it to yield an optimal portfolio. The objective here is to minimize the variance among all portfolios, or alternatively, to maximize expected return among all portfolios that has at least a certain expected return. Furthermore, this study focuses on optimizing risk portfolio so called Markowitz MVO (Mean-Variance Optimization). Some theoretical frameworks for analysis are arithmetic mean, geometric mean, variance, covariance, linear programming, and quadratic programming. Moreover, finding a minimum variance portfolio produces a convex quadratic programming, that is minimizing the objective function ðð¥with constraintsð ð 𥠥 ðandð´ð¥ = ð. The outcome of this research is the solution of optimal risk portofolio in some investments that could be finished smoothly using MATLAB R2007b software together with its graphic analysis
- …