28 research outputs found
MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language Models
Language Models (LMs) have shown impressive performance in various natural
language tasks. However, when it comes to natural language reasoning, LMs still
face challenges such as hallucination, generating incorrect intermediate
reasoning steps, and making mathematical errors. Recent research has focused on
enhancing LMs through self-improvement using feedback. Nevertheless, existing
approaches relying on a single generic feedback source fail to address the
diverse error types found in LM-generated reasoning chains. In this work, we
propose Multi-Aspect Feedback, an iterative refinement framework that
integrates multiple feedback modules, including frozen LMs and external tools,
each focusing on a specific error category. Our experimental results
demonstrate the efficacy of our approach to addressing several errors in the
LM-generated reasoning chain and thus improving the overall performance of an
LM in several reasoning tasks. We see a relative improvement of up to 20% in
Mathematical Reasoning and up to 18% in Logical Entailment.Comment: Accepted at EMNLP 2023 Main Conference, Camera Read
Automatically Correcting Large Language Models: Surveying the landscape of diverse self-correction strategies
Large language models (LLMs) have demonstrated remarkable performance across
a wide array of NLP tasks. However, their efficacy is undermined by undesired
and inconsistent behaviors, including hallucination, unfaithful reasoning, and
toxic content. A promising approach to rectify these flaws is self-correction,
where the LLM itself is prompted or guided to fix problems in its own output.
Techniques leveraging automated feedback -- either produced by the LLM itself
or some external system -- are of particular interest as they are a promising
way to make LLM-based solutions more practical and deployable with minimal
human feedback. This paper presents a comprehensive review of this emerging
class of techniques. We analyze and taxonomize a wide array of recent work
utilizing these strategies, including training-time, generation-time, and
post-hoc correction. We also summarize the major applications of this strategy
and conclude by discussing future directions and challenges.Comment: Work in Progress. Version
Learning Attention-based Embeddings for Relation Prediction in Knowledge Graphs
The recent proliferation of knowledge graphs
(KGs) coupled with incomplete or partial information, in the form of missing relations
(links) between entities, has fueled a lot of
research on knowledge base completion (also
known as relation prediction). Several recent works suggest that convolutional neural
network (CNN) based models generate richer
and more expressive feature embeddings and
hence also perform well on relation prediction.
However, we observe that these KG embeddings treat triples independently and thus fail
to cover the complex and hidden information
that is inherently implicit in the local neighborhood surrounding a triple. To this effect, our
paper proposes a novel attention-based feature
embedding that captures both entity and relation features in any given entity’s neighborhood. Additionally, we also encapsulate relation clusters and multi-hop relations in our
model. Our empirical study offers insights
into the efficacy of our attention-based model
and we show marked performance gains in
comparison to state-of-the-art methods on all
datasets
FEW-SHOT LEARNING ON GRAPHS VIA SUPERCLASSES BASED ON GRAPH SPECTRAL MEASURES
We propose to study the problem of few-shot graph classification in graph neural networks (GNNs) to recognize unseen classes, given limited labeled graph
examples. Despite several interesting GNN variants being proposed recently for
node and graph classification tasks, when faced with scarce labeled examples in
the few-shot setting, these GNNs exhibit significant loss in classification performance. Here, we present an approach where a probability measure is assigned
to each graph based on the spectrum of the graph’s normalized Laplacian. This
enables us to accordingly cluster the graph base-labels associated with each graph
into super-classes, where the L
p Wasserstein distance serves as our underlying distance metric. Subsequently, a super-graph constructed based on the super-classes
is then fed to our proposed GNN framework which exploits the latent inter-class
relationships made explicit by the super-graph to achieve better class label separation among the graphs. We conduct exhaustive empirical evaluations of our
proposed method and show that it outperforms both the adaptation of state-ofthe-art graph classification methods to few-shot scenario and our naive baseline
GNNs. Additionally, we also extend and study the behavior of our method to
semi-supervised and active learning scenarios
Derek Utley, Intercultural Resource Pack 5. Intercultural Communication Resources for Language Teachers
L’intérêt de l’enseignement interculturel en anglais des affaires n’est plus à démontrer, comme en témoignent les nombreux ouvrages, séminaires et formations diverses proposées aux entreprises pour sensibiliser leur personnel aux différences culturelles. Mais comment l’enseigner ? Il semblait jusque là que la démarche ethnographique proposée par Fons Trompenaars entre autres soit la plus efficace. Une phase de sensibilisation, au cours de laquelle des situations problème sont proposées aux pa..
Learning attention-based embeddings for relation prediction in knowledge graphs
The recent proliferation of knowledge graphs (KGs) coupled with incomplete or partial information, in the form of missing relations (links) between entities, has fueled a lot of research on knowledge base completion (also known as relation prediction). Several recent works suggest that convolutional neural network (CNN) based models generate richer and more expressive feature embeddings and hence also perform well on relation prediction. However, we observe that these KG embeddings treat triples independently and thus fail to cover the complex and hidden information that is inherently implicit in the local neighborhood surrounding a triple. To this effect, our paper proposes a novel attention-based feature embedding that captures both entity and relation features in any given entity's neighborhood. Additionally, we also encapsulate relation clusters and multi-hop relations in our model. Our empirical study offers insights into the efficacy of our attention-based model and we show marked performance gains in comparison to state-of-the-art methods on all datasets
Tetralogy of Fallot with restrictive ventricular septal defect by accessory tricuspid leaflet tissue
In tetralogy of Fallot septal defect is usually large because of malalignment of outlet septum, restrictive defect has been reported rarely. We present a case of tetralogy of Fallot with accessory tricuspid leaflet tissue restricting ventricular septal defect. The report includes echocardiographic and catheter images of this rare presentation of tetralogy of Fallot