3,749 research outputs found
CoaCor: Code Annotation for Code Retrieval with Reinforcement Learning
To accelerate software development, much research has been performed to help
people understand and reuse the huge amount of available code resources. Two
important tasks have been widely studied: code retrieval, which aims to
retrieve code snippets relevant to a given natural language query from a code
base, and code annotation, where the goal is to annotate a code snippet with a
natural language description. Despite their advancement in recent years, the
two tasks are mostly explored separately. In this work, we investigate a novel
perspective of Code annotation for Code retrieval (hence called `CoaCor'),
where a code annotation model is trained to generate a natural language
annotation that can represent the semantic meaning of a given code snippet and
can be leveraged by a code retrieval model to better distinguish relevant code
snippets from others. To this end, we propose an effective framework based on
reinforcement learning, which explicitly encourages the code annotation model
to generate annotations that can be used for the retrieval task. Through
extensive experiments, we show that code annotations generated by our framework
are much more detailed and more useful for code retrieval, and they can further
improve the performance of existing code retrieval models significantly.Comment: 10 pages, 2 figures. Accepted by The Web Conference (WWW) 201
Interactive Semantic Parsing for If-Then Recipes via Hierarchical Reinforcement Learning
Given a text description, most existing semantic parsers synthesize a program
in one shot. However, it is quite challenging to produce a correct program
solely based on the description, which in reality is often ambiguous or
incomplete. In this paper, we investigate interactive semantic parsing, where
the agent can ask the user clarification questions to resolve ambiguities via a
multi-turn dialogue, on an important type of programs called "If-Then recipes."
We develop a hierarchical reinforcement learning (HRL) based agent that
significantly improves the parsing performance with minimal questions to the
user. Results under both simulation and human evaluation show that our agent
substantially outperforms non-interactive semantic parsers and rule-based
agents.Comment: 13 pages, 2 figures, accepted by AAAI 201
Leveraging 2-hop Distant Supervision from Table Entity Pairs for Relation Extraction
Distant supervision (DS) has been widely used to automatically construct
(noisy) labeled data for relation extraction (RE). Given two entities, distant
supervision exploits sentences that directly mention them for predicting their
semantic relation. We refer to this strategy as 1-hop DS, which unfortunately
may not work well for long-tail entities with few supporting sentences. In this
paper, we introduce a new strategy named 2-hop DS to enhance distantly
supervised RE, based on the observation that there exist a large number of
relational tables on the Web which contain entity pairs that share common
relations. We refer to such entity pairs as anchors for each other, and collect
all sentences that mention the anchor entity pairs of a given target entity
pair to help relation prediction. We develop a new neural RE method REDS2 in
the multi-instance learning paradigm, which adopts a hierarchical model
structure to fuse information respectively from 1-hop DS and 2-hop DS.
Extensive experimental results on a benchmark dataset show that REDS2 can
consistently outperform various baselines across different settings by a
substantial margin
Cleared for Takeoff? Compositional & Conditional Reasoning may be the Achilles Heel to (Flight-Booking) Language Agents
The rapid progress of large language models (LLMs) has seen them excel and
frequently surpass human performance on standard benchmarks. This has enabled
many downstream applications, such as LLM agents, to rely on their
sophisticated reasoning to navigate complex task requirements. However, LLMs
are known to unexpectedly falter in simple tasks and under seemingly
straightforward circumstances - underscoring the need for better and more
diverse evaluation setups to measure their true capabilities. To this end, we
choose to study compositional and conditional reasoning, two cornerstones of
human cognition, and introduce GroundCocoa - a lexically diverse benchmark
connecting these reasoning skills to the real-world problem of flight booking.
Our task involves aligning detailed user preferences with available flight
options presented in a multiple-choice format. Results indicate a significant
disparity in performance among current state-of-the-art LLMs with even the best
performing model, GPT-4 Turbo, not exceeding 67% accuracy despite advanced
prompting techniques.Comment: 18 pages, 17 figures, 3 tables. Paper under revie
RobustSTL: A Robust Seasonal-Trend Decomposition Algorithm for Long Time Series
Decomposing complex time series into trend, seasonality, and remainder
components is an important task to facilitate time series anomaly detection and
forecasting. Although numerous methods have been proposed, there are still many
time series characteristics exhibiting in real-world data which are not
addressed properly, including 1) ability to handle seasonality fluctuation and
shift, and abrupt change in trend and reminder; 2) robustness on data with
anomalies; 3) applicability on time series with long seasonality period. In the
paper, we propose a novel and generic time series decomposition algorithm to
address these challenges. Specifically, we extract the trend component robustly
by solving a regression problem using the least absolute deviations loss with
sparse regularization. Based on the extracted trend, we apply the the non-local
seasonal filtering to extract the seasonality component. This process is
repeated until accurate decomposition is obtained. Experiments on different
synthetic and real-world time series datasets demonstrate that our method
outperforms existing solutions.Comment: Accepted to the thirty-third AAAI Conference on Artificial
Intelligence (AAAI 2019), 9 pages, 5 figure
- …