367 research outputs found
Answering questions about archived, annotated meetings
Retrieving information from archived meetings is a new domain of information retrieval that has received increasing attention in the past few years. Search in spontaneous spoken conversations has been recognized as more difficult than text-based document retrieval because meeting discussions contain two levels of information: the content itself, i.e. what topics are discussed, but also the argumentation process, i.e. what conflicts are resolved and what decisions are made. To capture the richness of information in meetings, current research focuses on recording meetings in Smart-Rooms, transcribing meeting discussion into text and annotating discussion with semantic higher-level structures to allow for efficient access to the data. However, it is not yet clear what type of user interface is best suited for searching and browsing such archived, annotated meetings. Content-based retrieval with keyword search is too naive and does not take into account the semantic annotations on the data. The objective of this thesis is to assess the feasibility and usefulness of a natural language interface to meeting archives that allows users to ask complex questions about meetings and retrieve episodes of meeting discussions based on semantic annotations. The particular issues that we address are: the need of argumentative annotation to answer questions about meetings; the linguistic and domain-specific natural language understanding techniques required to interpret such questions; and the use of visual overviews of meeting annotations to guide users in formulating questions. To meet the outlined objectives, we have annotated meetings with argumentative structure and built a prototype of a natural language understanding engine that interprets questions based on those annotations. Further, we have performed two sets of user experiments to study what questions users ask when faced with a natural language interface to annotated meeting archives. For this, we used a simulation method called Wizard of Oz, to enable users to express questions in their own terms without being influenced by limitations in speech recognition technology. Our experimental results show that technically it is feasible to annotate meetings and implement a deep-linguistic NLU engine for questions about meetings, but in practice users do not consistently take advantage of these features. Instead they often search for keywords in meetings. When visual overviews of the available annotations are provided, users refer to those annotations in their questions, but the complexity of questions remains simple. Users search with a breadth-first approach, asking questions in sequence instead of a single complex question. We conclude that natural language interfaces to meeting archives are useful, but that more experimental work is needed to find ways to incent users to take advantage of the expressive power of natural language when asking questions about meetings
Recommended from our members
Dialog as a vehicle for lifelong learning of grounded language understanding systems
Natural language interfaces have the potential to make various forms of technology, including mobile phones and computers as well as robots or other machines such as ATMs and self-checkout counters, more accessible and less intimidating to users who are unfamiliar or uncomfortable with other types of interfaces. In particular, natural language understanding systems on physical robots face a number of challenges, including the need to ground language in perception, the ability to adapt to changes in the environment and novel uses of language, and to deal with uncertainty in understanding. To effectively handle these challenges, such systems need to perform lifelong learning - continually updating the scope and predictions of the model with user interactions. In this thesis, we discuss ways in which dialog interaction with users can be used to improve grounded natural language understanding systems, motivated by service robot applications.
We focus on two types of queries that can be used in such dialog systems – active learning queries to elicit knowledge about the environment that can be used to improve perceptual models, and clarification questions that confirm the system’s hypotheses, or elicit specific information required to complete a task. Our goal is to build a system that can learn how to interact with users balancing a quick completion of tasks desired by the user with asking additional active learning questions viiito improve the underlying grounded language understanding components.
We present work on jointly improving semantic parsers from and learning a dialog policy for clarification dialogs, that improve a robot’s ability to understand natural language commands. We introduce the framework of opportunistic active learning, where a robot introduces opportunistic queries, that may not be immediately relevant, into an interaction in the hope of improving performance in future interactions. We demonstrate the usefulness of this framework in learning to ground natural language descriptions of objects, and learn a dialog policy for such interactions. We also learn dialog policies that balance task completion, opportunistic active learning, and attribute-based clarification questions. Finally, we attempt to expand this framework to different types of underlying models of grounded language understanding.Computer Science
Analysis and Design of Speech-Recognition Grammars
Currently, most commercial speech-enabled products are constructed using grammar-based technology. Grammar design is a critical issue for good recognition accuracy. Two methods are commonly used for creating grammars: 1) to generate them automatically from a large corpus of input data which is very costly to acquire, or 2) to construct them using an iterative process involving manual design, followed by testing with end-user speech input. This is a time-consuming and very expensive process requiring expert knowledge of language design, as well as the application area. Another hurdle to the creation and use of speech-enabled applications is that expertise is also required to integrate the speech capability with the application code and to deploy the application for wide-scale use. An alternative approach, which we propose, is 1) to construct them using the iterative process described above, but to replace end-user testing by analysis of the recognition grammars using a set of grammar metrics which have been shown to be good indicators of recognition accuracy, 2) to improve recognition accuracy in the design process by encoding semantic constraints in the syntax rules of the grammar, 3) to augment the above process by generating recognition grammars automatically from specifications of the application, and 4) to use tools for creating speech-enabled applications together with an architecture for their deployment which enables expert users, as well as users who do not have expertise in language processing, to easily build speech applications and add them to the web
Robust Dialog Management Through A Context-centric Architecture
This dissertation presents and evaluates a method of managing spoken dialog interactions with a robust attention to fulfilling the human user’s goals in the presence of speech recognition limitations. Assistive speech-based embodied conversation agents are computer-based entities that interact with humans to help accomplish a certain task or communicate information via spoken input and output. A challenging aspect of this task involves open dialog, where the user is free to converse in an unstructured manner. With this style of input, the machine’s ability to communicate may be hindered by poor reception of utterances, caused by a user’s inadequate command of a language and/or faults in the speech recognition facilities. Since a speech-based input is emphasized, this endeavor involves the fundamental issues associated with natural language processing, automatic speech recognition and dialog system design. Driven by ContextBased Reasoning, the presented dialog manager features a discourse model that implements mixed-initiative conversation with a focus on the user’s assistive needs. The discourse behavior must maintain a sense of generality, where the assistive nature of the system remains constant regardless of its knowledge corpus. The dialog manager was encapsulated into a speech-based embodied conversation agent platform for prototyping and testing purposes. A battery of user trials was performed on this agent to evaluate its performance as a robust, domain-independent, speech-based interaction entity capable of satisfying the needs of its users
CRT-based dialogs: Theory and design
CRT (cathode ray tube) based, direct selection dialogs for computing machines and systems were apparently a cure for issues like ease of learning and ease of use. But unforeseen ~ and probably unforeseeable problems arose as increasingly sophisticated systems and dialogs were developed. This paper describes some of the emerging problems in CRT-based dialog design, develops theories about why they occur, and discusses potential solutions for them as a basis for future research. This investigation also provides a survey of the research into what makes programming and programming languages difficult, and what makes them simple
Ontologies, VoiceXML and Reasoners
Despite the inherent limitation of accessing information serially, voice applications are increasingly growing in popularity as computing technologies advance. This is a positive development, because voice communication offers a number of benefits over other forms of communication. For example, voice may be better for delivering services to users whose eyes and hands may be engaged in other activities (e.g. driving) or to semi-literate or illiterate users. This thesis proposes a knowledge based architecture for building voice applications to help reduce the limitations of serial access to information. The proposed architecture, called OVR (Ontologies, VoiceXML and Reasoners), uses a rich backend that represents knowledge via ontologies and utilises reasoning engines to reason with it, in order to generate intelligent behaviour. Ontologies were chosen over other knowledge representation formalisms because of their expressivity and executable format, and because current trends suggest a general shift towards the use of ontologies in many systems used for information storing and sharing. For the frontend, this architecture uses VoiceXML, the emerging, and de facto standard for voice automated applications. A functional prototype was built for an initial validation of the architecture. The system is a simple voice application to help locate information about service providers that offer HIV (Human Immunodeficiency Virus) testing. We called this implementation HTLS (HIV Testing Locator System). The functional prototype was implemented using a number of technologies. OWL API, a Java interface designed to facilitate manipulation of ontologies authored in OWL was used to build a customised query interface for HTLS. Pellet reasoner was used for supporting queries to the knowledge base and Drools (JBoss rule engine) was used for processing dialog rules. VXI was used as the VoiceXML browser and an experimental softswitch called iLanga as the bridge to the telephony system. (At the heart of iLanga is Asterisk, a well known PBX-in-a-box.) HTLS behaved properly under system testing, providing the sought initial validation of OVR.LaTeX with hyperref packag
I-centric User Interaction
Die Vision I-centric Communications bedeutet, einen uneingeschränkten Blick auf das menschliche Kommunikationsverhalten zu werfen, um Kommunikationssysteme entsprechend daran angepasst zu entwickeln. Diese Vision definiert einen benutzerorientierten Ansatz zur Erstellung von Diensten und Anwendungen. Dies setzt zunächst eine Analyse der Benutzeranforderungen voraus, um geeignete Systeme und Dienste zu entwerfen. Anstatt Technologie-fokussierte Lösungen ohne jegliche Anpassung an die jeweiligen Personen anzubieten, sollte ein I-centric System seine Dienste ohne sichtbare technische Details und unter Berücksichtigung von Benutzerpräferenzen sowie der Benutzerumgebung darbieten. Der Vision von I-centric Communications folgend, stellt die vorliegende Arbeit einen Ansatz zur Realisierung der Idee von I-centric User Interaction vor. Dieser Ansatz erweitert und vervollständigt die Vision durch verbesserte Benutzerinteraktionsfähigkeiten. Diese Dissertation zeigt, dass es möglich ist, Kommunikationssysteme zu realisieren, die die Interaktion zwischen Benutzern und Diensten ohne Einschränkung auf bestimmte Technologien für Benutzerschnittstellen sowie in personalisierter und umgebungsberücksichtigende Art und Weise unterstützen. Derartig verbesserte Benutzerinteraktion wird die Akzeptanz und die Benutzung von Diensten erhöhen. Einerseits soll die Benutzerinteraktion verschiedene Arten von Technologien für Benutzerschnittstellen unterstützen, durch die die Geräte-Unabhängigkeit und der ständige Zugang zu den Diensten ermöglicht werden. Entsprechend dem aktuellem Kontext und der Absicht können die Benutzer die bevorzugte und geeignete Art der Interaktion wählen. Andererseits soll die Interaktion selbst den Benutzerpräferenzen sowie der jeweiligen Umgebung angepasst werden. Dementsprechend diskutiert die vorliegende Arbeit diese unterschiedlichen Problembereiche, identifiziert die notwendigen Funktionen und bietet entsprechende Lösungsansätze jeweils. Die Arbeit präsentiert und analysiert zunächst die Vision I-centric Communications mit Hinblick auf den Aspekt der Benutzerinteraktion. Basierend auf den identifizierten Anforderungen wurde ein Ansatz zur Realisierung von I-centric User Interaction entwickelt. Dieser Ansatz, der in dieser Arbeit vorgestellt wird, spezifiziert ein Service Adaptation Framework und einzelne Modelle für Generische Benutzerinteraktion , für Personalisierung sowie für Ambient Awareness , die sich jeweils auf die identifizierten Problembereiche konzentrieren. Abschließend präsentiert die vorliegende Arbeit Ergebnisse einer prototypischen Realisierung des dargelegten Ansatzes. Die Ergebnisse demonstrieren die Einsetzbarkeit der entwickelten Konzepte und die Erfüllung der Vision von I-centric User Interaction . Die Forschungsarbeit im Bereich I-centric Communications wurde in Kooperation zwischen dem Lehrstuhl für Offene Kommunikationssysteme (OKS) der Technischen Universität Berlin (TUB) und dem Fraunhofer Institut FOKUS durchgeführt. Die Vision sowie das Referenzmodell für I-centric Communications , die in der vorliegenden Arbeit vorgestellt werden, sind Ergebnisse dieser Kooperation. Die Forschungsschwerpunkte der Kooperation zwischen TUB und FOKUS waren das Generelle Modell für I-centric Dienste , die Dienstplattform für I-centric Dienste sowie ein Ansatz zur Interaktion zwischen Nutzern und I-centric Dienste . Die vorliegende Arbeit konzentriert sich auf den Ansatz I-centric User Interaction , der die Interaktion zwischen Nutzern und den Diensten betrachtet. Die Aspekte der I-centric Dienste werden in der vorliegenden Arbeit nicht betrachtet. Diese Aspekte wurden in einer zweiten Dissertation von Stefan Arbanowski, Fraunhofer FOKUS, analysiert und ausgearbeitet. Die Ergebnisse dieser Arbeit wurden in verschiedenen nationalen und internationalen Forschungsprojekten (BMBF LiveFutura, BMBF PI-AVIda, BMBF VHE-UD, IST WSI, IST WWRI), Standardisierungsgremien (OMG, WWRF), Konferenzpapieren sowie Zeitschriften eingebracht, um die Vision von I-centric Communications einem größeren Auditorium vorzustellen.The vision of I-centric Communications means to take an unlimited look at human communication behavior and to adapt the activities of communication systems to it. This vision defines a user-centered approach for the realization of services and applications. It requires to start analyzing user demands to design suitable systems and services. Instead of just providing technology-focused solutions without any adaptation to individuals, an I-centric system should provide services hiding technical details and considering the individual s preferences as well as the individual s environment. Following the vision of I-centric Communications, this thesis introduces an approach to realize I-centric User Interaction. This approach enhances and completes the vision by providing advanced user interaction capabilities. It answers the question whether it is possible to realize a communication system, which allows the interaction between user and services without any restriction to specific user interface technologies and in a personalized as well as ambient aware manner. Such enhanced user interaction will lead to a higher acceptance and increased usage of services. On the one hand, the user interaction shall support different kinds of user interface technologies enabling Device Independence and ubiquitous access to the services. According to their current context and intended action, users can select the preferred and suitable way of interaction. On the other hand, the user interaction shall be adapted to the user s preferences and to the user s environment. Accordingly, this work discusses these different areas of concern, identifies necessary functions, and provides suitable solutions for each. First, the thesis introduces and analyses the vision of I-centric Communications with special regard to the aspect of user interaction. Based on the identified requirements and areas of concern, an approach to realize I-centric User Interaction was developed. The approach, presented in this thesis, specifies a Service Adaptation Framework and individual models for Personalization, for Ambient Awareness, and for Generic User Interaction focusing on the respective areas of concern. Finally, the thesis illustrates the results from the prototypical implementation of the presented approach, which has been pursued in several projects in parallel. These results demonstrate the applicability of the developed concepts and the fulfillment of the vision of I-centric User Interaction. The work in the area of I-centric Communications was carried out in cooperation of the Department for Open Communication Systems (OKS) at the Technical University Berlin (TUB) and the Fraunhofer Institute FOKUS. The vision and the reference model for I-centric Communications, introduced in this thesis, are results of this cooperation. The main research directions for the cooperation between TUB and FOKUS have been a general model for I-centric services, the service platform for I-centric services, and an approach for the interaction of users with I-centric services. This thesis focuses on an approach for I-centric User Interaction. The general aspects of I-centric services as defined by the vision are out of scope of this thesis. Nevertheless, these aspects have been analyzed by Stefan Arbanowski, researcher at Fraunhofer FOKUS, in a second PhD thesis in parallel. The results of this work have been contributed to different national and international projects (BMBF LiveFutura, BMBF PI-AVIda, BMBF VHE-UD, IST WSI, IST WWRI), standardization bodies (OMG, WWRF), conferences papers, and journals by introducing the vision of I-centric Communications to a larger auditorium, and by exploiting parts of the developed I-centric systems
Innovations for Requirements Analysis, From Stakeholders' Needs to Formal Designs
14th MontereyWorkshop 2007
Monterey, CA, USA, September 10-13, 2007
Revised Selected PapersWe are pleased to present the proceedings of the 14thMontereyWorkshop, which
took place September 10–13, 2007 in Monterey, CA, USA. In this preface, we give
the reader an overview of what took place at the workshop and introduce the
contributions in this Lecture Notes in Computer Science volume. A complete
introduction to the theme of the workshop, as well as to the history of the
Monterey Workshop series, can be found in Luqi and Kordon’s “Advances in
Requirements Engineering: Bridging the Gap between Stakeholders’ Needs and
Formal Designs” in this volume. This paper also contains the case study that
many participants used as a problem to frame their analyses, and a summary of
the workshop’s results
- …