1,448 research outputs found

    Leveraging Image based Prior for Visual Place Recognition

    Full text link
    In this study, we propose a novel scene descriptor for visual place recognition. Unlike popular bag-of-words scene descriptors which rely on a library of vector quantized visual features, our proposed descriptor is based on a library of raw image data, such as publicly available photo collections from Google StreetView and Flickr. The library images need not to be associated with spatial information regarding the viewpoint and orientation of the scene. As a result, these images are cheaper than the database images; in addition, they are readily available. Our proposed descriptor directly mines the image library to discover landmarks (i.e., image patches) that suitably match an input query/database image. The discovered landmarks are then compactly described by their pose and shape (i.e., library image ID, bounding boxes) and used as a compact discriminative scene descriptor for the input image. We evaluate the effectiveness of our scene description framework by comparing its performance to that of previous approaches.Comment: 8 pages, 6 figures, preprint. Accepted for publication in MVA2015 (oral presentation

    Data-Driven Shape Analysis and Processing

    Full text link
    Data-driven methods play an increasingly important role in discovering geometric, structural, and semantic relationships between 3D shapes in collections, and applying this analysis to support intelligent modeling, editing, and visualization of geometric data. In contrast to traditional approaches, a key feature of data-driven approaches is that they aggregate information from a collection of shapes to improve the analysis and processing of individual shapes. In addition, they are able to learn models that reason about properties and relationships of shapes without relying on hard-coded rules or explicitly programmed instructions. We provide an overview of the main concepts and components of these techniques, and discuss their application to shape classification, segmentation, matching, reconstruction, modeling and exploration, as well as scene analysis and synthesis, through reviewing the literature and relating the existing works with both qualitative and numerical comparisons. We conclude our report with ideas that can inspire future research in data-driven shape analysis and processing.Comment: 10 pages, 19 figure

    Augmented Reality and Health Informatics: A Study based on Bibliometric and Content Analysis of Scholarly Communication and Social Media

    Get PDF
    Healthcare outcomes have been shown to improve when technology is used as part of patient care. Health Informatics (HI) is a multidisciplinary study of the design, development, adoption, and application of IT-based innovations in healthcare services delivery, management, and planning. Augmented Reality (AR) is an emerging technology that enhances the userโ€™s perception and interaction with the real world. This study aims to illuminate the intersection of the field of AR and HI. The domains of AR and HI by themselves are areas of significant research. However, there is a scarcity of research on augmented reality as it applies to health informatics. Given both scholarly research and social media communication having contributed to the domains of AR and HI, research methodologies of bibliometric and content analysis on scholarly research and social media communication were employed to investigate the salient features and research fronts of the field. The study used Scopus data (7360 scholarly publications) to identify the bibliometric features and to perform content analysis of the identified research. The Altmetric database (an aggregator of data sources) was used to determine the social media communication for this field. The findings from this study included Publication Volumes, Top Authors, Affiliations, Subject Areas and Geographical Locations from scholarly publications as well as from a social media perspective. The highest cited 200 documents were used to determine the research fronts in scholarly publications. Content Analysis techniques were employed on the publication abstracts as a secondary technique to determine the research themes of the field. The study found the research frontiers in the scholarly communication included emerging AR technologies such as tracking and computer vision along with Surgical and Learning applications. There was a commonality between social media and scholarly communication themes from an applications perspective. In addition, social media themes included applications of AR in Healthcare Delivery, Clinical Studies and Mental Disorders. Europe as a geographic region dominates the research field with 50% of the articles and North America and Asia tie for second with 20% each. Publication volumes show a steep upward slope indicating continued research. Social Media communication is still in its infancy in terms of data extraction, however aggregators like Altmetric are helping to enhance the outcomes. The findings from the study revealed that the frontier research in AR has made an impact in the surgical and learning applications of HI and has the potential for other applications as new technologies are adopted

    ์˜๋ฏธ๋ก ์  ํ™˜๊ฒฝ ์ดํ•ด ๊ธฐ๋ฐ˜ ์ธ๊ฐ„ ๋กœ๋ด‡ ํ˜‘์—…

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ •๋ณด๊ณตํ•™๋ถ€,2020. 2. ์ด๋ฒ”ํฌ.Human-robot cooperation is unavoidable in various applications ranging from manufacturing to field robotics owing to the advantages of adaptability and high flexibility. Especially, complex task planning in large, unconstructed, and uncertain environments can employ the complementary capabilities of human and diverse robots. For a team to be effectives, knowledge regarding team goals and current situation needs to be effectively shared as they affect decision making. In this respect, semantic scene understanding in natural language is one of the most fundamental components for information sharing between humans and heterogeneous robots, as robots can perceive the surrounding environment in a form that both humans and other robots can understand. Moreover, natural-language-based scene understanding can reduce network congestion and improve the reliability of acquired data. Especially, in field robotics, transmission of raw sensor data increases network bandwidth and decreases quality of service. We can resolve this problem by transmitting information in the form of natural language that has encoded semantic representations of environments. In this dissertation, I introduce a human and heterogeneous robot cooperation scheme based on semantic scene understanding. I generate sentences and scene graphs, which is a natural language grounded graph over the detected objects and their relationships, with the graph map generated using a robot mapping algorithm. Subsequently, a framework that can utilize the results for cooperative mission planning of humans and robots is proposed. Experiments were performed to verify the effectiveness of the proposed methods. This dissertation comprises two parts: graph-based scene understanding and scene understanding based on the cooperation between human and heterogeneous robots. For the former, I introduce a novel natural language processing method using a semantic graph map. Although semantic graph maps have been widely applied to study the perceptual aspects of the environment, such maps do not find extensive application in natural language processing tasks. Several studies have been conducted on the understanding of workspace images in the field of computer vision; in these studies, the sentences were automatically generated, and therefore, multiple scenes have not yet been utilized for sentence generation. A graph-based convolutional neural network, which comprises spectral graph convolution and graph coarsening, and a recurrent neural network are employed to generate sentences attention over graphs. The proposed method outperforms the conventional methods on a publicly available dataset for single scenes and can be utilized for sequential scenes. Recently, deep learning has demonstrated impressive developments in scene understanding using natural language. However, it has not been extensively applied to high-level processes such as causal reasoning, analogical reasoning, or planning. The symbolic approach that calculates the sequence of appropriate actions by combining the available skills of agents outperforms in reasoning and planning; however, it does not entirely consider semantic knowledge acquisition for human-robot information sharing. An architecture that combines deep learning techniques and symbolic planner for human and heterogeneous robots to achieve a shared goal based on semantic scene understanding is proposed for scene understanding based on human-robot cooperation. In this study, graph-based perception is used for scene understanding. A planning domain definition language (PDDL) planner and JENA-TDB are utilized for mission planning and data acquisition storage, respectively. The effectiveness of the proposed method is verified in two situations: a mission failure, in which the dynamic environment changes, and object detection in a large and unseen environment.์ธ๊ฐ„๊ณผ ์ด์ข… ๋กœ๋ด‡ ๊ฐ„์˜ ํ˜‘์—…์€ ๋†’์€ ์œ ์—ฐ์„ฑ๊ณผ ์ ์‘๋ ฅ์„ ๋ณด์ผ ์ˆ˜ ์žˆ๋‹ค๋Š” ์ ์—์„œ ์ œ์กฐ์—…์—์„œ ํ•„๋“œ ๋กœ๋ณดํ‹ฑ์Šค๊นŒ์ง€ ๋‹ค์–‘ํ•œ ๋ถ„์•ผ์—์„œ ํ•„์—ฐ์ ์ด๋‹ค. ํŠนํžˆ, ์„œ๋กœ ๋‹ค๋ฅธ ๋Šฅ๋ ฅ์„ ์ง€๋‹Œ ๋กœ๋ด‡๋“ค๊ณผ ์ธ๊ฐ„์œผ๋กœ ๊ตฌ์„ฑ๋œ ํ•˜๋‚˜์˜ ํŒ€์€ ๋„“๊ณ  ์ •ํ˜•ํ™”๋˜์ง€ ์•Š์€ ๊ณต๊ฐ„์—์„œ ์„œ๋กœ์˜ ๋Šฅ๋ ฅ์„ ๋ณด์™„ํ•˜๋ฉฐ ๋ณต์žกํ•œ ์ž„๋ฌด ์ˆ˜ํ–‰์„ ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•œ๋‹ค๋Š” ์ ์—์„œ ํฐ ์žฅ์ ์„ ๊ฐ–๋Š”๋‹ค. ํšจ์œจ์ ์ธ ํ•œ ํŒ€์ด ๋˜๊ธฐ ์œ„ํ•ด์„œ๋Š”, ํŒ€์˜ ๊ณตํ†ต๋œ ๋ชฉํ‘œ ๋ฐ ๊ฐ ํŒ€์›์˜ ํ˜„์žฌ ์ƒํ™ฉ์— ๊ด€ํ•œ ์ •๋ณด๋ฅผ ์‹ค์‹œ๊ฐ„์œผ๋กœ ๊ณต์œ ํ•  ์ˆ˜ ์žˆ์–ด์•ผ ํ•˜๋ฉฐ ํ•จ๊ป˜ ์˜์‚ฌ ๊ฒฐ์ •์„ ํ•  ์ˆ˜ ์žˆ์–ด์•ผ ํ•œ๋‹ค. ์ด๋Ÿฌํ•œ ๊ด€์ ์—์„œ, ์ž์—ฐ์–ด๋ฅผ ํ†ตํ•œ ์˜๋ฏธ๋ก ์  ํ™˜๊ฒฝ ์ดํ•ด๋Š” ์ธ๊ฐ„๊ณผ ์„œ๋กœ ๋‹ค๋ฅธ ๋กœ๋ด‡๋“ค์ด ๋ชจ๋‘ ์ดํ•ดํ•  ์ˆ˜ ์žˆ๋Š” ํ˜•ํƒœ๋กœ ํ™˜๊ฒฝ์„ ์ธ์ง€ํ•œ๋‹ค๋Š” ์ ์—์„œ ๊ฐ€์žฅ ํ•„์ˆ˜์ ์ธ ์š”์†Œ์ด๋‹ค. ๋˜ํ•œ, ์šฐ๋ฆฌ๋Š” ์ž์—ฐ์–ด ๊ธฐ๋ฐ˜ ํ™˜๊ฒฝ ์ดํ•ด๋ฅผ ํ†ตํ•ด ๋„คํŠธ์›Œํฌ ํ˜ผ์žก์„ ํ”ผํ•จ์œผ๋กœ์จ ํš๋“ํ•œ ์ •๋ณด์˜ ์‹ ๋ขฐ์„ฑ์„ ๋†’์ผ ์ˆ˜ ์žˆ๋‹ค. ํŠนํžˆ, ๋Œ€๋Ÿ‰์˜ ์„ผ์„œ ๋ฐ์ดํ„ฐ ์ „์†ก์— ์˜ํ•ด ๋„คํŠธ์›Œํฌ ๋Œ€์—ญํญ์ด ์ฆ๊ฐ€ํ•˜๊ณ  ํ†ต์‹  QoS (Quality of Service) ์‹ ๋ขฐ๋„๊ฐ€ ๊ฐ์†Œํ•˜๋Š” ๋ฌธ์ œ๊ฐ€ ๋นˆ๋ฒˆํžˆ ๋ฐœ์ƒํ•˜๋Š” ํ•„๋“œ ๋กœ๋ณดํ‹ฑ์Šค ์˜์—ญ์—์„œ๋Š” ์˜๋ฏธ๋ก ์  ํ™˜๊ฒฝ ์ •๋ณด์ธ ์ž์—ฐ์–ด๋ฅผ ์ „์†กํ•จ์œผ๋กœ์จ ํ†ต์‹  ๋Œ€์—ญํญ์„ ๊ฐ์†Œ์‹œํ‚ค๊ณ  ํ†ต์‹  QoS ์‹ ๋ขฐ๋„๋ฅผ ์ฆ๊ฐ€์‹œํ‚ฌ ์ˆ˜ ์žˆ๋‹ค. ๋ณธ ํ•™์œ„ ๋…ผ๋ฌธ์—์„œ๋Š” ํ™˜๊ฒฝ์˜ ์˜๋ฏธ๋ก ์  ์ดํ•ด ๊ธฐ๋ฐ˜ ์ธ๊ฐ„ ๋กœ๋ด‡ ํ˜‘๋™ ๋ฐฉ๋ฒ•์— ๋Œ€ํ•ด ์†Œ๊ฐœํ•œ๋‹ค. ๋จผ์ €, ๋กœ๋ด‡์˜ ์ง€๋„ ์ž‘์„ฑ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ํ†ตํ•ด ํš๋“ํ•œ ๊ทธ๋ž˜ํ”„ ์ง€๋„๋ฅผ ์ด์šฉํ•˜์—ฌ ์ž์—ฐ์–ด ๋ฌธ์žฅ๊ณผ ๊ฒ€์ถœํ•œ ๊ฐ์ฒด ๋ฐ ๊ฐ ๊ฐ์ฒด ๊ฐ„์˜ ๊ด€๊ณ„๋ฅผ ์ž์—ฐ์–ด ๋‹จ์–ด๋กœ ํ‘œํ˜„ํ•˜๋Š” ๊ทธ๋ž˜ํ”„๋ฅผ ์ƒ์„ฑํ•œ๋‹ค. ๊ทธ๋ฆฌ๊ณ  ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ ๊ฒฐ๊ณผ๋ฅผ ์ด์šฉํ•˜์—ฌ ์ธ๊ฐ„๊ณผ ๋‹ค์–‘ํ•œ ๋กœ๋ด‡๋“ค์ด ํ•จ๊ป˜ ํ˜‘์—…ํ•˜์—ฌ ์ž„๋ฌด๋ฅผ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜๋Š” ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ ์ œ์•ˆํ•œ๋‹ค. ๋ณธ ํ•™์œ„ ๋…ผ๋ฌธ์€ ํฌ๊ฒŒ ๊ทธ๋ž˜ํ”„๋ฅผ ์ด์šฉํ•œ ์˜๋ฏธ๋ก ์  ํ™˜๊ฒฝ ์ดํ•ด์™€ ์˜๋ฏธ๋ก ์  ํ™˜๊ฒฝ ์ดํ•ด๋ฅผ ํ†ตํ•œ ์ธ๊ฐ„๊ณผ ์ด์ข… ๋กœ๋ด‡ ๊ฐ„์˜ ํ˜‘์—… ๋ฐฉ๋ฒ•์œผ๋กœ ๊ตฌ์„ฑ๋œ๋‹ค. ๋จผ์ €, ๊ทธ๋ž˜ํ”„๋ฅผ ์ด์šฉํ•œ ์˜๋ฏธ๋ก ์  ํ™˜๊ฒฝ ์ดํ•ด ๋ถ€๋ถ„์—์„œ๋Š” ์˜๋ฏธ๋ก ์  ๊ทธ๋ž˜ํ”„ ์ง€๋„๋ฅผ ์ด์šฉํ•œ ์ƒˆ๋กœ์šด ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ ๋ฐฉ๋ฒ•์— ๋Œ€ํ•ด ์†Œ๊ฐœํ•œ๋‹ค. ์˜๋ฏธ๋ก ์  ๊ทธ๋ž˜ํ”„ ์ง€๋„ ์ž‘์„ฑ ๋ฐฉ๋ฒ•์€ ๋กœ๋ด‡์˜ ํ™˜๊ฒฝ ์ธ์ง€ ์ธก๋ฉด์—์„œ ๋งŽ์ด ์—ฐ๊ตฌ๋˜์—ˆ์ง€๋งŒ ์ด๋ฅผ ์ด์šฉํ•œ ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ ๋ฐฉ๋ฒ•์€ ๊ฑฐ์˜ ์—ฐ๊ตฌ๋˜์ง€ ์•Š์•˜๋‹ค. ๋ฐ˜๋ฉด ์ปดํ“จํ„ฐ ๋น„์ „ ๋ถ„์•ผ์—์„œ๋Š” ์ด๋ฏธ์ง€๋ฅผ ์ด์šฉํ•œ ํ™˜๊ฒฝ ์ดํ•ด ์—ฐ๊ตฌ๊ฐ€ ๋งŽ์ด ์ด๋ฃจ์–ด์กŒ์ง€๋งŒ, ์—ฐ์†์ ์ธ ์žฅ๋ฉด๋“ค์€ ๋‹ค๋ฃจ๋Š”๋ฐ๋Š” ํ•œ๊ณ„์ ์ด ์žˆ๋‹ค. ๋”ฐ๋ผ์„œ ์šฐ๋ฆฌ๋Š” ๊ทธ๋ž˜ํ”„ ์ŠคํŽ™ํŠธ๋Ÿผ ์ด๋ก ์— ๊ธฐ๋ฐ˜ํ•œ ๊ทธ๋ž˜ํ”„ ์ปจ๋ณผ๋ฃจ์…˜๊ณผ ๊ทธ๋ž˜ํ”„ ์ถ•์†Œ ๋ ˆ์ด์–ด๋กœ ๊ตฌ์„ฑ๋œ ๊ทธ๋ž˜ํ”„ ์ปจ๋ณผ๋ฃจ์…˜ ์‹ ๊ฒฝ๋ง ๋ฐ ์ˆœํ™˜ ์‹ ๊ฒฝ๋ง์„ ์ด์šฉํ•˜์—ฌ ๊ทธ๋ž˜ํ”„๋ฅผ ์„ค๋ช…ํ•˜๋Š” ๋ฌธ์žฅ์„ ์ƒ์„ฑํ•œ๋‹ค. ์ œ์•ˆํ•œ ๋ฐฉ๋ฒ•์€ ๊ธฐ์กด์˜ ๋ฐฉ๋ฒ•๋“ค๋ณด๋‹ค ํ•œ ์žฅ๋ฉด์— ๋Œ€ํ•ด ํ–ฅ์ƒ๋œ ์„ฑ๋Šฅ์„ ๋ณด์˜€์œผ๋ฉฐ ์—ฐ์†๋œ ์žฅ๋ฉด๋“ค์— ๋Œ€ํ•ด์„œ๋„ ์„ฑ๊ณต์ ์œผ๋กœ ์ž์—ฐ์–ด ๋ฌธ์žฅ์„ ์ƒ์„ฑํ•œ๋‹ค. ์ตœ๊ทผ ๋”ฅ๋Ÿฌ๋‹์€ ์ž์—ฐ์–ด ๊ธฐ๋ฐ˜ ํ™˜๊ฒฝ ์ธ์ง€์— ์žˆ์–ด ๊ธ‰์†๋„๋กœ ํฐ ๋ฐœ์ „์„ ์ด๋ฃจ์—ˆ๋‹ค. ํ•˜์ง€๋งŒ ์ธ๊ณผ ์ถ”๋ก , ์œ ์ถ”์  ์ถ”๋ก , ์ž„๋ฌด ๊ณ„ํš๊ณผ ๊ฐ™์€ ๋†’์€ ์ˆ˜์ค€์˜ ํ”„๋กœ์„ธ์Šค์—๋Š” ์ ์šฉ์ด ํž˜๋“ค๋‹ค. ๋ฐ˜๋ฉด ์ž„๋ฌด๋ฅผ ์ˆ˜ํ–‰ํ•˜๋Š” ๋ฐ ์žˆ์–ด ๊ฐ ์—์ด์ „ํŠธ์˜ ๋Šฅ๋ ฅ์— ๋งž๊ฒŒ ํ–‰์œ„๋“ค์˜ ์ˆœ์„œ๋ฅผ ๊ณ„์‚ฐํ•ด์ฃผ๋Š” ์ƒ์ง•์  ์ ‘๊ทผ๋ฒ•(symbolic approach)์€ ์ถ”๋ก ๊ณผ ์ž„๋ฌด ๊ณ„ํš์— ์žˆ์–ด ๋›ฐ์–ด๋‚œ ์„ฑ๋Šฅ์„ ๋ณด์ด์ง€๋งŒ ์ธ๊ฐ„๊ณผ ๋กœ๋ด‡๋“ค ์‚ฌ์ด์˜ ์˜๋ฏธ๋ก ์  ์ •๋ณด ๊ณต์œ  ๋ฐฉ๋ฒ•์— ๋Œ€ํ•ด์„œ๋Š” ๊ฑฐ์˜ ๋‹ค๋ฃจ์ง€ ์•Š๋Š”๋‹ค. ๋”ฐ๋ผ์„œ, ์ธ๊ฐ„๊ณผ ์ด์ข… ๋กœ๋ด‡ ๊ฐ„์˜ ํ˜‘์—… ๋ฐฉ๋ฒ• ๋ถ€๋ถ„์—์„œ๋Š” ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ฒ•๋“ค๊ณผ ์ƒ์ง•์  ํ”Œ๋ž˜๋„ˆ(symbolic planner)๋ฅผ ์—ฐ๊ฒฐํ•˜๋Š” ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ ์ œ์•ˆํ•˜์—ฌ ์˜๋ฏธ๋ก ์  ์ดํ•ด๋ฅผ ํ†ตํ•œ ์ธ๊ฐ„ ๋ฐ ์ด์ข… ๋กœ๋ด‡ ๊ฐ„์˜ ํ˜‘์—…์„ ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•œ๋‹ค. ์šฐ๋ฆฌ๋Š” ์˜๋ฏธ๋ก ์  ์ฃผ๋ณ€ ํ™˜๊ฒฝ ์ดํ•ด๋ฅผ ์œ„ํ•ด ์ด์ „ ๋ถ€๋ถ„์—์„œ ์ œ์•ˆํ•œ ๊ทธ๋ž˜ํ”„ ๊ธฐ๋ฐ˜ ์ž์—ฐ์–ด ๋ฌธ์žฅ ์ƒ์„ฑ์„ ์ˆ˜ํ–‰ํ•œ๋‹ค. PDDL ํ”Œ๋ž˜๋„ˆ์™€ JENA-TDB๋Š” ๊ฐ๊ฐ ์ž„๋ฌด ๊ณ„ํš ๋ฐ ์ •๋ณด ํš๋“ ์ €์žฅ์†Œ๋กœ ์‚ฌ์šฉํ•œ๋‹ค. ์ œ์•ˆํ•œ ๋ฐฉ๋ฒ•์˜ ํšจ์šฉ์„ฑ์€ ์‹œ๋ฎฌ๋ ˆ์ด์…˜์„ ํ†ตํ•ด ๋‘ ๊ฐ€์ง€ ์ƒํ™ฉ์— ๋Œ€ํ•ด์„œ ๊ฒ€์ฆํ•œ๋‹ค. ํ•˜๋‚˜๋Š” ๋™์  ํ™˜๊ฒฝ์—์„œ ์ž„๋ฌด ์‹คํŒจ ์ƒํ™ฉ์ด๋ฉฐ ๋‹ค๋ฅธ ํ•˜๋‚˜๋Š” ๋„“์€ ๊ณต๊ฐ„์—์„œ ๊ฐ์ฒด๋ฅผ ์ฐพ๋Š” ์ƒํ™ฉ์ด๋‹ค.1 Introduction 1 1.1 Background and Motivation 1 1.2 Literature Review 5 1.2.1 Natural Language-Based Human-Robot Cooperation 5 1.2.2 Artificial Intelligence Planning 5 1.3 The Problem Statement 10 1.4 Contributions 11 1.5 Dissertation Outline 12 2 Natural Language-Based Scene Graph Generation 14 2.1 Introduction 14 2.2 Related Work 16 2.3 Scene Graph Generation 18 2.3.1 Graph Construction 19 2.3.2 Graph Inference 19 2.4 Experiments 22 2.5 Summary 25 3 Language Description with 3D Semantic Graph 26 3.1 Introduction 26 3.2 Related Work 26 3.3 Natural Language Description 29 3.3.1 Preprocess 29 3.3.2 Graph Feature Extraction 33 3.3.3 Natural Language Description with Graph Features 34 3.4 Experiments 35 3.5 Summary 42 4 Natural Question with Semantic Graph 43 4.1 Introduction 43 4.2 Related Work 45 4.3 Natural Question Generation 47 4.3.1 Preprocess 49 4.3.2 Graph Feature Extraction 50 4.3.3 Natural Question with Graph Features 51 4.4 Experiments 52 4.5 Summary 58 5 PDDL Planning with Natural Language 59 5.1 Introduction 59 5.2 Related Work 60 5.3 PDDL Planning with Incomplete World Knowledge 61 5.3.1 Natural Language Process for PDDL Planning 63 5.3.2 PDDL Planning System 64 5.4 Experiments 65 5.5 Summary 69 6 PDDL Planning with Natural Language-Based Scene Understanding 70 6.1 Introduction 70 6.2 Related Work 74 6.3 A Framework for Heterogeneous Multi-Agent Cooperation 77 6.3.1 Natural Language-Based Cognition 78 6.3.2 Knowledge Engine 80 6.3.3 PDDL Planning Agent 81 6.4 Experiments 82 6.4.1 Experiment Setting 82 6.4.2 Scenario 84 6.4.3 Results 87 6.5 Summary 91 7 Conclusion 92Docto

    {CurveFusion}: {R}econstructing Thin Structures from {RGBD} Sequences

    Get PDF
    We introduce CurveFusion, the first approach for high quality scanning of thin structures at interactive rates using a handheld RGBD camera. Thin filament-like structures are mathematically just 1D curves embedded in R^3, and integration-based reconstruction works best when depth sequences (from the thin structure parts) are fused using the object's (unknown) curve skeleton. Thus, using the complementary but noisy color and depth channels, CurveFusion first automatically identifies point samples on potential thin structures and groups them into bundles, each being a group of a fixed number of aligned consecutive frames. Then, the algorithm extracts per-bundle skeleton curves using L1 axes, and aligns and iteratively merges the L1 segments from all the bundles to form the final complete curve skeleton. Thus, unlike previous methods, reconstruction happens via integration along a data-dependent fusion primitive, i.e., the extracted curve skeleton. We extensively evaluate CurveFusion on a range of challenging examples, different scanner and calibration settings, and present high fidelity thin structure reconstructions previously just not possible from raw RGBD sequences

    Computational time analysis in extended kalman filter based simultaneous localization and mapping

    Get PDF
    The simultaneous localization and mapping (SLAM) of a mobile robot is one of the applications that use estimation techniques. SLAM is a navigation technique that allows a mobile robot to navigate around autonomously while observing its surroundings in an unfamiliar environment. SLAM does not require a priori map, instead the mobile robot creates a map of the area incrementally with the help of sensors on board and uses this map to localize its location Due to its relatively easy algorithm and efficiency of estimation via the representation of the belief by a multivariate Gaussian distribution and a unimodal distribution, with a single mean annotated and corresponding covariance uncertainty, the extended Kalman filter (EKF) has become one of the most preferred estimators in mobile robot SLAM. However, due to the update process of the covariance matrix, EKF-based SLAM has high computational time. In SLAM, if more observation is being made by mobile robot, the state covariance size will be increasing. This eventually requires more memory and processing time due to excessive computation needs to be calculated over time. Therefore there is a need of enhancing the estimation performance by reducing the computational time in SLAM. Three phases involve in this research methodology which the first is theoretical formulation of the mobile robot model. This is followed by the environment and estimation method used to solve the SLAM of mobile robot. Simulation analysis was used to verify the findings. This research attempts to introduce a new approach to simplify the structure of the covariance matrix using the eigenvalues matrix diagonalization method. Through simulation result it is proved that time taken to complete the SLAM process using diagonalized covariance was reduced as compared to the normal covariance. However, there is one limitation encountered from this method in which the covariance values become too small, that indicates an optimistic estimation. For this reason, second objective is motivated to improve the optimistic problem. Addition of new element into the diagonal matrix, which is known as a pseudo element, is also investigated in this study. Via mathematical approach, these problems are discussed and explored from estimation-theoretic point of view. Through adding the pseudo noise element into diagonalized covariance, the optimistic condition of covariance matrix can be improved. This was shown through the increased size of covariance ellipses at the end of simulation process. Based on the findings it can be concluded that the addition of pseudo matrix in the updated state covariance can further improved the computational time for mobile robot estimation
    • โ€ฆ
    corecore