2 research outputs found
Human-In-The-Loop Learning of Qualitative Preference Models
In this work, we present a novel human-in-the-loop framework to help the
human user understand the decision making process that involves choosing
preferred options. We focus on qualitative preference models over alternatives
from combinatorial domains. This framework is interactive: the user provides
her behavioral data to the framework, and the framework explains the learned
model to the user. It is iterative: the framework collects feedback on the
learned model from the user and tries to improve it accordingly till the user
terminates the iteration. In order to communicate the learned preference model
to the user, we develop visualization of intuitive and explainable graphic
models, such as lexicographic preference trees and forests, and conditional
preference networks. To this end, we discuss key aspects of our framework for
lexicographic preference models.Comment: Published in the Proceedings of the 32nd International Florida
Artificial Intelligence Research Society Conference, 201
The Complexity of Learning Acyclic Conditional Preference Networks
Learning of user preferences, as represented by, for example, Conditional
Preference Networks (CP-nets), has become a core issue in AI research. Recent
studies investigate learning of CP-nets from randomly chosen examples or from
membership and equivalence queries. To assess the optimality of learning
algorithms as well as to better understand the combinatorial structure of
classes of CP-nets, it is helpful to calculate certain learning-theoretic
information complexity parameters. This article focuses on the frequently
studied case of learning from so-called swap examples, which express
preferences among objects that differ in only one attribute. It presents bounds
on or exact values of some well-studied information complexity parameters,
namely the VC dimension, the teaching dimension, and the recursive teaching
dimension, for classes of acyclic CP-nets. We further provide algorithms that
learn tree-structured and general acyclic CP-nets from membership queries.
Using our results on complexity parameters, we assess the optimality of our
algorithms as well as that of another query learning algorithm for acyclic
CP-nets presented in the literature. Our algorithms are near-optimal, and can,
under certain assumptions, be adapted to the case when the membership oracle is
faulty.Comment: 57 page