8 research outputs found

    Implementasi Screenplay dan Audio Foley Effect pada Pembuatan Film Animasi 3D "Si Molek"

    Get PDF
    Screenplay adalah kisah yang diceritakan dengan gambar, dalam dialog dan deskripsi, dan ditempatkan dalam konteks struktur dramatis, yang menguraikan urut-urutan adegan, tempat, dialog dan keadaan. Foley merupakan jenis suara yang dibuat untuk mengisi noise pemain dalam film pada waktu yang sebenarnya atau nyata. Teknologi animasi sendiri sudah menyebar hingga pada pembuatan cerita daerah/cerita rakyat yang awalnya hanya diceritakan secara lisan/tertulis. Cerita rakyat adalah suatu cerita yang pada dasarnya disampaikan oleh seseorang pada orang lain melalui penuturan lisan. Hampir disetiap daerah di Indonesia memiliki cerita rakyatnya masing-masing. Salah satu cerita rakyat dari provinsi Riau adalah cerita rakyat Si Molek. Pada penelitian ini, akan dibuat sebuah film animasi 3D cerita rakyat Si Molek dengan menggunakan teknik Screenplay dan juga teknik Foley Effect. Teknik Screenplay digunakan sebagai acuan atau pedoman dalam proses pembuatan animasi 3D Si Molek. Sedangkan teknik Foley Effect digunakan karena tanpa adanya Foley Effect sebuah film akan terasa kurang realistis dan natural. Dengan film animasi 3D Si Molek yang menerapkan teknik Screenplay dan Foley Effect ini diharapkan menghasilkan sebuah film animasi 3D yang baik dan juga sebagai sarana memperkenalkan cerita rakyat ini kepada masyarakat. Berdasarkan hasil pengujian yang telah dilakukan teknik Screenplay dan teknik Foley Effect berhasil diimplementasikan dengan benar

    Agents meet OKR: An Object and Key Results Driven Agent System with Hierarchical Self-Collaboration and Self-Evaluation

    Full text link
    In this study, we introduce the concept of OKR-Agent designed to enhance the capabilities of Large Language Models (LLMs) in task-solving. Our approach utilizes both self-collaboration and self-correction mechanism, facilitated by hierarchical agents, to address the inherent complexities in task-solving. Our key observations are two-fold: first, effective task-solving demands in-depth domain knowledge and intricate reasoning, for which deploying specialized agents for individual sub-tasks can markedly enhance LLM performance. Second, task-solving intrinsically adheres to a hierarchical execution structure, comprising both high-level strategic planning and detailed task execution. Towards this end, our OKR-Agent paradigm aligns closely with this hierarchical structure, promising enhanced efficacy and adaptability across a range of scenarios. Specifically, our framework includes two novel modules: hierarchical Objects and Key Results generation and multi-level evaluation, each contributing to more efficient and robust task-solving. In practical, hierarchical OKR generation decomposes Objects into multiple sub-Objects and assigns new agents based on key results and agent responsibilities. These agents subsequently elaborate on their designated tasks and may further decompose them as necessary. Such generation operates recursively and hierarchically, culminating in a comprehensive set of detailed solutions. The multi-level evaluation module of OKR-Agent refines solution by leveraging feedback from all associated agents, optimizing each step of the process. This ensures solution is accurate, practical, and effectively address intricate task requirements, enhancing the overall reliability and quality of the outcome. Experimental results also show our method outperforms the previous methods on several tasks. Code and demo are available at https://okr-agent.github.io

    A Human-Computer Duet System for Music Performance

    Full text link
    Virtual musicians have become a remarkable phenomenon in the contemporary multimedia arts. However, most of the virtual musicians nowadays have not been endowed with abilities to create their own behaviors, or to perform music with human musicians. In this paper, we firstly create a virtual violinist, who can collaborate with a human pianist to perform chamber music automatically without any intervention. The system incorporates the techniques from various fields, including real-time music tracking, pose estimation, and body movement generation. In our system, the virtual musician's behavior is generated based on the given music audio alone, and such a system results in a low-cost, efficient and scalable way to produce human and virtual musicians' co-performance. The proposed system has been validated in public concerts. Objective quality assessment approaches and possible ways to systematically improve the system are also discussed

    A Review of Text-to-Animation Systems

    Get PDF
    Text-to-graphics systems encompass three types of tools: text-to-picture, text-to-scene and text-to-animation. They are an artificial intelligence application wherein users can create 2D and 3D scenes or animations and recently immersive environments from natural language. These complex tasks require the collaboration of various fields, such as natural language processing, computational linguistics and computer graphics. Text-to-animation systems have received more interest than their counterparts, and have been developed for various domains, including theatrical pre-production, education or training. In this survey we focus on text-to-animation systems, discussing their requirements, challenges and proposing solutions, and investigate the natural language understanding approaches adopted in previous research works to solve the challenge of animation generation. We review text-to-animation systems developed over the period 2001-2021, and investigate their recent trends in order to paint the current landscape of the field

    Generating Animations from Screenplays

    No full text
    Automatically generating animation from natural language text finds application in a number of areas e.g. movie script writing, instructional videos, and public safety. However, translating natural language text into animation is a challenging task. Existing text-to-animation systems can handle only very simple sentences, which limits their applications. In this paper, we develop a text-to-animation system which is capable of handling complex sentences. We achieve this by introducing a text simplification step into the process. Building on an existing animation generation system for screenwriting, we create a robust NLP pipeline to extract information from screenplays and map them to the system’s knowledge base. We develop a set of linguistic transformation rules that simplify complex sentences. Information extracted from the simplified sentences is used to generate a rough storyboard and video depicting the text. Our sentence simplification module outperforms existing systems in terms of BLEU and SARI metrics.We further evaluated our system via a user study: 68% participants believe that our system generates reasonable animation from input screenplays
    corecore