26 research outputs found
Re-Temp: Relation-Aware Temporal Representation Learning for Temporal Knowledge Graph Completion
Temporal Knowledge Graph Completion (TKGC) under the extrapolation setting
aims to predict the missing entity from a fact in the future, posing a
challenge that aligns more closely with real-world prediction problems.
Existing research mostly encodes entities and relations using sequential graph
neural networks applied to recent snapshots. However, these approaches tend to
overlook the ability to skip irrelevant snapshots according to entity-related
relations in the query and disregard the importance of explicit temporal
information. To address this, we propose our model, Re-Temp (Relation-Aware
Temporal Representation Learning), which leverages explicit temporal embedding
as input and incorporates skip information flow after each timestamp to skip
unnecessary information for prediction. Additionally, we introduce a two-phase
forward propagation method to prevent information leakage. Through the
evaluation on six TKGC (extrapolation) datasets, we demonstrate that our model
outperforms all eight recent state-of-the-art models by a significant margin.Comment: Findings of EMNLP 202
DDI-MuG: Multi-aspect graphs for drug-drug interaction extraction
IntroductionDrug-drug interaction (DDI) may lead to adverse reactions in patients, thus it is important to extract such knowledge from biomedical texts. However, previously proposed approaches typically focus on capturing sentence-aspect information while ignoring valuable knowledge concerning the whole corpus. In this paper, we propose a Multi-aspect Graph-based DDI extraction model, named DDI-MuG.MethodsWe first employ a bio-specific pre-trained language model to obtain the token contextualized representations. Then we use two graphs to get syntactic information from input instance and word co-occurrence information within the entire corpus, respectively. Finally, we combine the representations of drug entities and verb tokens for the final classificationResultsTo validate the effectiveness of the proposed model, we perform extensive experiments on two widely used DDI extraction dataset, DDIExtraction-2013 and TAC 2018. It is encouraging to see that our model outperforms all twelve state-of-the-art models.DiscussionIn contrast to the majority of earlier models that rely on the black-box approach, our model enables visualization of crucial words and their interrelationships by utilizing edge information from two graphs. To the best of our knowledge, this is the first model that explores multi-aspect graphs to the DDI extraction task, and we hope it can establish a foundation for more robust multi-aspect works in the future
SG-Shuffle: Multi-aspect Shuffle Transformer for Scene Graph Generation
Scene Graph Generation (SGG) serves a comprehensive representation of the
images for human understanding as well as visual understanding tasks. Due to
the long tail bias problem of the object and predicate labels in the available
annotated data, the scene graph generated from current methodologies can be
biased toward common, non-informative relationship labels. Relationship can
sometimes be non-mutually exclusive, which can be described from multiple
perspectives like geometrical relationships or semantic relationships, making
it even more challenging to predict the most suitable relationship label. In
this work, we proposed the SG-Shuffle pipeline for scene graph generation with
3 components: 1) Parallel Transformer Encoder, which learns to predict object
relationships in a more exclusive manner by grouping relationship labels into
groups of similar purpose; 2) Shuffle Transformer, which learns to select the
final relationship labels from the category-specific feature generated in the
previous step; and 3) Weighted CE loss, used to alleviate the training bias
caused by the imbalanced dataset
A Survey of Large Language Models in Finance (FinLLMs)
Large Language Models (LLMs) have shown remarkable capabilities across a wide
variety of Natural Language Processing (NLP) tasks and have attracted attention
from multiple domains, including financial services. Despite the extensive
research into general-domain LLMs, and their immense potential in finance,
Financial LLM (FinLLM) research remains limited. This survey provides a
comprehensive overview of FinLLMs, including their history, techniques,
performance, and opportunities and challenges. Firstly, we present a
chronological overview of general-domain Pre-trained Language Models (PLMs)
through to current FinLLMs, including the GPT-series, selected open-source
LLMs, and financial LMs. Secondly, we compare five techniques used across
financial PLMs and FinLLMs, including training methods, training data, and
fine-tuning methods. Thirdly, we summarize the performance evaluations of six
benchmark tasks and datasets. In addition, we provide eight advanced financial
NLP tasks and datasets for developing more sophisticated FinLLMs. Finally, we
discuss the opportunities and the challenges facing FinLLMs, such as
hallucination, privacy, and efficiency. To support AI research in finance, we
compile a collection of accessible datasets and evaluation benchmarks on
GitHub.Comment: More information on https://github.com/adlnlp/FinLLM
PDF-VQA: A New Dataset for Real-World VQA on PDF Documents
Document-based Visual Question Answering examines the document understanding
of document images in conditions of natural language questions. We proposed a
new document-based VQA dataset, PDF-VQA, to comprehensively examine the
document understanding from various aspects, including document element
recognition, document layout structural understanding as well as contextual
understanding and key information extraction. Our PDF-VQA dataset extends the
current scale of document understanding that limits on the single document page
to the new scale that asks questions over the full document of multiple pages.
We also propose a new graph-based VQA model that explicitly integrates the
spatial and hierarchically structural relationships between different document
elements to boost the document structural understanding. The performances are
compared with several baselines over different question types and
tasks\footnote{The full dataset will be released after paper acceptance
StockEmotions: Discover Investor Emotions for Financial Sentiment Analysis and Multivariate Time Series
There has been growing interest in applying NLP techniques in the financial
domain, however, resources are extremely limited. This paper introduces
StockEmotions, a new dataset for detecting emotions in the stock market that
consists of 10,000 English comments collected from StockTwits, a financial
social media platform. Inspired by behavioral finance, it proposes 12
fine-grained emotion classes that span the roller coaster of investor emotion.
Unlike existing financial sentiment datasets, StockEmotions presents granular
features such as investor sentiment classes, fine-grained emotions, emojis, and
time series data. To demonstrate the usability of the dataset, we perform a
dataset analysis and conduct experimental downstream tasks. For financial
sentiment/emotion classification tasks, DistilBERT outperforms other baselines,
and for multivariate time series forecasting, a Temporal Attention LSTM model
combining price index, text, and emotion features achieves the best performance
than using a single feature.Comment: Preprint for the AAAI-23 Bridge Program (AI for Financial Services
Doc-GCN: Heterogeneous Graph Convolutional Networks for Document Layout Analysis
Recognizing the layout of unstructured digital documents is crucial when
parsing the documents into the structured, machine-readable format for
downstream applications. Recent studies in Document Layout Analysis usually
rely on computer vision models to understand documents while ignoring other
information, such as context information or relation of document components,
which are vital to capture. Our Doc-GCN presents an effective way to harmonize
and integrate heterogeneous aspects for Document Layout Analysis. We first
construct graphs to explicitly describe four main aspects, including syntactic,
semantic, density, and appearance/visual information. Then, we apply graph
convolutional networks for representing each aspect of information and use
pooling to integrate them. Finally, we aggregate each aspect and feed them into
2-layer MLPs for document layout component classification. Our Doc-GCN achieves
new state-of-the-art results in three widely used DLA datasets.Comment: Accepted by COLING 202