3 research outputs found
Optimizing for Measure of Performance in Max-Margin Parsing
Many statistical learning problems in the area of natural language processing
including sequence tagging, sequence segmentation and syntactic parsing has
been successfully approached by means of structured prediction methods. An
appealing property of the corresponding discriminative learning algorithms is
their ability to integrate the loss function of interest directly into the
optimization process, which potentially can increase the resulting performance
accuracy. Here, we demonstrate on the example of constituency parsing how to
optimize for F1-score in the max-margin framework of structural SVM. In
particular, the optimization is with respect to the original (not binarized)
trees
Worst-Case Polynomial-Time Exact MAP Inference on Discrete Models with Global Dependencies
Considering the worst-case scenario, junction tree algorithm remains the most
efficient and general solution for exact MAP inference on discrete graphical
models. Unfortunately, its main tractability assumption requires the treewidth
of a corresponding MRF to be bounded strongly limiting the range of admissible
applications. In fact, many practical problems in the area of structured
prediction require modelling of global dependencies by either directly
introducing global factors or enforcing global constraints on the prediction
variables. This, however, always results in a fully-connected graph making
exact inference by means of this algorithm intractable. Nevertheless, depending
on the structure of the global factors, we can further relax the conditions for
an efficient inference. In this paper we reformulate the work in [1] and
present a better way to establish the theory also extending the set of
handleable problem instances for free - since it requires only a simple
modification of the originally presented algorithm. To demonstrate that this
extension is not of a purely theoretical interest we identify one further use
case in the context of generalisation bounds for structured learning which
cannot be handled by the previous formulation. Finally, we accordingly adjust
the theoretical guarantees that the modified algorithm always finds an optimal
solution in polynomial time
A Survey on Multi-output Learning
Multi-output learning aims to simultaneously predict multiple outputs given
an input. It is an important learning problem due to the pressing need for
sophisticated decision making in real-world applications. Inspired by big data,
the 4Vs characteristics of multi-output imposes a set of challenges to
multi-output learning, in terms of the volume, velocity, variety and veracity
of the outputs. Increasing number of works in the literature have been devoted
to the study of multi-output learning and the development of novel approaches
for addressing the challenges encountered. However, it lacks a comprehensive
overview on different types of challenges of multi-output learning brought by
the characteristics of the multiple outputs and the techniques proposed to
overcome the challenges. This paper thus attempts to fill in this gap to
provide a comprehensive review on this area. We first introduce different
stages of the life cycle of the output labels. Then we present the paradigm on
multi-output learning, including its myriads of output structures, definitions
of its different sub-problems, model evaluation metrics and popular data
repositories used in the study. Subsequently, we review a number of
state-of-the-art multi-output learning methods, which are categorized based on
the challenges.Comment: Paper accepted by IEEE Transactions on Neural Networks and Learning
System