4 research outputs found
WebWISE: Web Interface Control and Sequential Exploration with Large Language Models
The paper investigates using a Large Language Model (LLM) to automatically
perform web software tasks using click, scroll, and text input operations.
Previous approaches, such as reinforcement learning (RL) or imitation learning,
are inefficient to train and task-specific. Our method uses filtered Document
Object Model (DOM) elements as observations and performs tasks step-by-step,
sequentially generating small programs based on the current observations. We
use in-context learning, either benefiting from a single manually provided
example, or an automatically generated example based on a successful zero-shot
trial. We evaluate the proposed method on the MiniWob++ benchmark. With only
one in-context example, our WebWISE method achieves similar or better
performance than other methods that require many demonstrations or trials
Efficient Title Reranker for Fast and Improved Knowledge-Intense NLP
In recent RAG approaches, rerankers play a pivotal role in refining retrieval
accuracy with the ability of revealing logical relations for each pair of query
and text. However, existing rerankers are required to repeatedly encode the
query and a large number of long retrieved text. This results in high
computational costs and limits the number of retrieved text, hindering
accuracy. As a remedy of the problem, we introduce the Efficient Title Reranker
via Broadcasting Query Encoder, a novel technique for title reranking that
achieves a 20x-40x speedup over the vanilla passage reranker. Furthermore, we
introduce Sigmoid Trick, a novel loss function customized for title reranking.
Combining both techniques, we empirically validated their effectiveness,
achieving state-of-the-art results on all four datasets we experimented with
from the KILT knowledge benchmark