165,361 research outputs found

    A Mouth Full of Words: Visually Consistent Acoustic Redubbing

    Get PDF
    This paper introduces a method for automatic redubbing of video that exploits the many-to-many mapping of phoneme sequences to lip movements modelled as dynamic visemes [1]. For a given utterance, the corresponding dynamic viseme sequence is sampled to construct a graph of possible phoneme sequences that synchronize with the video. When composed with a pronunciation dictionary and language model, this produces a vast number of word sequences that are in sync with the original video, literally putting plausible words into the mouth of the speaker. We demonstrate that traditional, one-to-many, static visemes lack flexibility for this application as they produce significantly fewer word sequences. This work explores the natural ambiguity in visual speech and offers insight for automatic speech recognition and the importance of language modeling

    Some Remarks on the Model Theory of Epistemic Plausibility Models

    Full text link
    Classical logics of knowledge and belief are usually interpreted on Kripke models, for which a mathematically well-developed model theory is available. However, such models are inadequate to capture dynamic phenomena. Therefore, epistemic plausibility models have been introduced. Because these are much richer structures than Kripke models, they do not straightforwardly inherit the model-theoretical results of modal logic. Therefore, while epistemic plausibility structures are well-suited for modeling purposes, an extensive investigation of their model theory has been lacking so far. The aim of the present paper is to fill exactly this gap, by initiating a systematic exploration of the model theory of epistemic plausibility models. Like in 'ordinary' modal logic, the focus will be on the notion of bisimulation. We define various notions of bisimulations (parametrized by a language L) and show that L-bisimilarity implies L-equivalence. We prove a Hennesy-Milner type result, and also two undefinability results. However, our main point is a negative one, viz. that bisimulations cannot straightforwardly be generalized to epistemic plausibility models if conditional belief is taken into account. We present two ways of coping with this issue: (i) adding a modality to the language, and (ii) putting extra constraints on the models. Finally, we make some remarks about the interaction between bisimulation and dynamic model changes.Comment: 19 pages, 3 figure

    Towards a flexible open-source software library for multi-layered scholarly textual studies: An Arabic case study dealing with semi-automatic language processing

    Get PDF
    This paper presents both the general model and a case study of the Computational and Collaborative Philology Library (CoPhiLib), an ongoing initiative underway at the Institute for Computational Linguistics (ILC) of the National Research Council (CNR), Pisa, Italy. The library, designed and organized as a reusable, abstract and open-source software component, aims at solving the needs of multi-lingual and cross-lingual analysis by exposing common Application Programming Interfaces (APIs). The core modules, coded by the Java programming language, constitute the groundwork of a Web platform designed to deal with textual scholarly needs. The Web application, implemented according to the Java Enterprise specifications, focuses on multi-layered analysis for the study of literary documents and related multimedia sources. This ambitious challenge seeks to obtain the management of textual resources, on the one hand by abstracting from current language, on the other hand by decoupling from the specific requirements of single projects. This goal is achieved thanks to methodologies declared by the 'agile process', and by putting into effect suitable use case modeling, design patterns, and component-based architectures. The reusability and flexibility of the system have been tested on an Arabic case study: the system allows users to choose the morphological engine (such as AraMorph or Al-Khalil), along with linguistic granularity (i.e. with or without declension). Finally, the application enables the construction of annotated resources for further statistical engines (training set). © 2014 IEEE

    A method for mapping XML-based specifications between development methodologies

    Get PDF
    The Unified Modeling Language (UML) is widely used by software engineers as the basis of analysis and design in software development. However, UML ignores human factors in the course of software development because of its strong emphasis on the internal structure and functionality of the application. This thesis presents a method of mapping human-computer interaction (HCI) requirement specifications generated by usability engineering (UE) methodologies (e.g. Putting Usability First (PUF)) into UML specifications. These two sets of requirement specification are specified, using Extensible Markup Language (XML) so that HCI requirement specifications can be integrated into UML ones. A Mapping Tool was developed to facilitate the creation of mappings between PUF XML tags and XMI tags. The Mapping Tool was used to create mappings between PUF and UML requirement specifications. This mapping process and its outputs were evaluated to demonstrate that the tool worked. The results of the evaluation show that the HCI requirement specification represented by the PUF XML tags can improve the UML specification by adding them into the XMI tags

    TEACHING AND LEARNING PROCESS TO IMPROVE READING SKILLS FOR THE FIFTH GRADE STUDENTS OF SD NEGERI 11 NGRINGO, JATEN, KARANGANYAR

    Get PDF
    This final project report is written based on the job training that which was carried out by the writer at SD Negeri 11 Ngringo, Jaten. The purposes of the project are to describe the process, method and the problems and the solution in teaching and learning process to improve reading skill for the fifth grade students of SD Negeri 11 Ngringo, Jaten, including its problems faced by the students and also the solutions for those problems. In writing this final report, the writer collected the data by observing the class, interviewing to the students, library study and directly teaching in the classroom. The writer found difficulties aced by students in the way to read the text covering the pronunciation and understanding the text. The writer used direct method in teaching and learning process. Moreover, the writer taught them systematically based on the lesson plan. It consist of three main steps namely, Pre Teaching, the process of teaching implementation and closing activity of the lesson. Pre-teaching was beginning activity of the lesson. The process of teaching implementation was the main steps of teacher duties that consisted of four steps namely, Building Knowledge of Field, Modeling Text, Join Construction and Individual Construction. The third step was closing activity where the writer could assess whether the objective of the lesson was adequately reached or not. Based on the job training experience, it can be concluded that the problems faced by fifth grade students in learning reading were from the students itself such as the low students’ interest and motivation. So, they were difficult to pronounce and understand the text. The writer tried to solve this problem by creating fun and communicative situation in the class related on the skill learned

    Non-functional requirements: size measurement and testing with COSMIC-FFP

    Get PDF
    The non-functional requirements (NFRs) of software systems are well known to add a degree of uncertainty to process of estimating the cost of any project. This paper contributes to the achievement of more precise project size measurement through incorporating NFRs into the functional size quantification process. We report on an initial solution proposed to deal with the problem of quantitatively assessing the NFR modeling process early in the project, and of generating test cases for NFR verification purposes. The NFR framework has been chosen for the integration of NFRs into the requirements modeling process and for their quantitative assessment. Our proposal is based on the functional size measurement method, COSMIC-FFP, adopted in 2003 as the ISO/IEC 19761 standard. Also in this paper, we extend the use of COSMIC-FFP for NFR testing purposes. This is an essential step for improving NFR development and testing effort estimates, and consequently for managing the scope of NFRs. We discuss the merits of the proposed approach and the open questions related to its design
    • …
    corecore