2,385 research outputs found

    SIFT Flow: Dense Correspondence across Scenes and its Applications

    Get PDF
    While image alignment has been studied in different areas of computer vision for decades, aligning images depicting different scenes remains a challenging problem. Analogous to optical flow where an image is aligned to its temporally adjacent frame, we propose SIFT flow, a method to align an image to its nearest neighbors in a large image corpus containing a variety of scenes. The SIFT flow algorithm consists of matching densely sampled, pixel-wise SIFT features between two images, while preserving spatial discontinuities. The SIFT features allow robust matching across different scene/object appearances, whereas the discontinuity-preserving spatial model allows matching of objects located at different parts of the scene. Experiments show that the proposed approach robustly aligns complex scene pairs containing significant spatial differences. Based on SIFT flow, we propose an alignment-based large database framework for image analysis and synthesis, where image information is transferred from the nearest neighbors to a query image according to the dense scene correspondence. This framework is demonstrated through concrete applications, such as motion field prediction from a single image, motion synthesis via object transfer, satellite image registration and face recognition

    Evaluation of process-structure-property relationships of carbon nanotube forests using simulation and deep learning

    Get PDF
    This work is aimed to explore process-structure-property relationships of carbon nanotube (CNT) forests. CNTs have superior mechanical, electrical and thermal properties that make them suitable for many applications. Yet, due to lack of manufacturing control, there is a huge performance gap between promising properties of individual CNTs and CNT forest properties that hinders their adoption into potential industrial applications. In this research, computational modelling, in-situ electron microscopy for CNT synthesis, and data-driven and high-throughput deep convolutional neural networks are employed to not only accelerate implementing CNTs in various applications but also to establish a framework to make validated predictive models that can be easily extended to achieve application-tailored synthesis of any materials. A time-resolved and physics-based finite-element simulation tool is modelled in MATLAB to investigate synthesis of CNT forests, specially to study the CNT-CNT interactions and generated mechanical forces and their role in ensemble structure and properties. A companion numerical model with similar construct is then employed to examine forest mechanical properties in compression. In addition, in-situ experiments are carried out inside Environmental Scanning Electron Microscope (ESEM) to nucleate and synthesize CNTs. Findings may primarily be used to expand the forest growth and self-assembly knowledge and to validate the assumptions of simulation package. Also, SEM images can be used as feed database to construct a deep learning model to grow CNTs by design. The chemical vapor deposition parameter space of CNT synthesis is so vast that it is not possible to investigate all conceivable combinations in terms of time and costs. Hence, simulated CNT forest morphology images are used to train machine learning and learning algorithms that are able to predict CNT synthesis conditions based on desired properties. Exceptionally high prediction accuracies of R2 > 0.94 is achieved for buckling load and stiffness, as well as accuracies of > 0.91 for the classification task. This high classification accuracy promotes discovering the CNT forest synthesis-structure relationships so that their promising performance can be adopted in real world applications. We foresee this work as a meaningful step towards creating an unsupervised simulation using machine learning techniques that can seek out the desired CNT forest synthesis parameters to achieve desired property sets for diverse applications.Includes bibliographical reference

    CT diagnosis of early stroke : the initial approach to the new CAD tool based on multiscale estimation of ischemia

    Get PDF
    Background: Computer aided diagnosis (CAD) becomes one of the most important diagnostic tools for urgent states in cerebral stroke and other life-threatening conditions where time plays a crucial role. Routine CT is still diagnostically insufficient in hyperacute stage of stroke that is in the therapeutic window for thrombolytic therapy. Authors present computer assistant of early ischemic stroke diagnosis that supports the radiologic interpretations. A new semantic-visualization system of ischemic symptoms applied to noncontrast, routine CT examination was based on multiscale image processing and diagnostic content estimation. Material/Methods: Evaluation of 95 sets of examinations in patients admitted to a hospital with symptoms suggesting stroke was undertaken by four radiologists from two medical centers unaware of the final clinical findings. All of the consecutive cases were considered as having no CT direct signs of hyperacute ischemia. At the first test stage only the CTs performed at the admission were evaluated independently by radiologists. Next, the same early scans were evaluated again with additional use of multiscale computer-assistant of stroke (MulCAS). Computerized suggestion with increased sensitivity to the subtle image manifestations of cerebral ischemia was constructed as additional view representing estimated diagnostic content with enhanced stroke symptoms synchronized to routine CT data preview. Follow-up CT examinations and clinical features confirmed or excluded the diagnosis of stroke constituting 'gold standard' to verify stroke detection performance. Results: Higher AUC (area under curve) values were found for MulCAS -aided radiological diagnosis for all readers and the differences were statistically significant for random readers-random cases parametric and non-parametric DBM MRMC analysis. Sensitivity and specificity of acute stroke detection for the readers was increased by 30% and 4%, respectively. Conclusions: Routine CT completed with proposed method of computer assisted diagnosis provided noticeable better diagnosis efficiency of acute stroke according to the rates and opinions of all test readers. Further research includes fully automatic detection of hypodense regions to complete assisted indications and formulate the suggestions of stroke cases more objectively. Planned prospective studies will let evaluate more accurately the impact of this CAD tool on diagnosis and further treatment in patients suffered from stroke. It is necessary to determine whether this method is possible to be applied widely

    Human Performance Modeling and Rendering via Neural Animated Mesh

    Full text link
    We have recently seen tremendous progress in the neural advances for photo-real human modeling and rendering. However, it's still challenging to integrate them into an existing mesh-based pipeline for downstream applications. In this paper, we present a comprehensive neural approach for high-quality reconstruction, compression, and rendering of human performances from dense multi-view videos. Our core intuition is to bridge the traditional animated mesh workflow with a new class of highly efficient neural techniques. We first introduce a neural surface reconstructor for high-quality surface generation in minutes. It marries the implicit volumetric rendering of the truncated signed distance field (TSDF) with multi-resolution hash encoding. We further propose a hybrid neural tracker to generate animated meshes, which combines explicit non-rigid tracking with implicit dynamic deformation in a self-supervised framework. The former provides the coarse warping back into the canonical space, while the latter implicit one further predicts the displacements using the 4D hash encoding as in our reconstructor. Then, we discuss the rendering schemes using the obtained animated meshes, ranging from dynamic texturing to lumigraph rendering under various bandwidth settings. To strike an intricate balance between quality and bandwidth, we propose a hierarchical solution by first rendering 6 virtual views covering the performer and then conducting occlusion-aware neural texture blending. We demonstrate the efficacy of our approach in a variety of mesh-based applications and photo-realistic free-view experiences on various platforms, i.e., inserting virtual human performances into real environments through mobile AR or immersively watching talent shows with VR headsets.Comment: 18 pages, 17 figure
    corecore