297 research outputs found
Novel Analytical Continuation Based Shape Reconstruction Methods For Perfect Electric Conducting Targets
Tez (Doktora) -- İstanbul Teknik Üniversitesi, Fen Bilimleri Enstitüsü, 2009Thesis (PhD) -- İstanbul Technical University, Institute of Science and Technology, 2009Bu çalışmanın amacı yanına erişilemeyen, elektriksel açıdan mükemmel iletken cisimlerin şekillerinin elektromagnetik dalgalar kullanılarak belirlenebilmesi için yeni ve etkin yöntemlerin geliştirilmesidir. Bu çerçevede, problemin kötü-kurulmuş ve doğrusal olmayan kısımlarını ayrı ayrı ele alan iki farklı yöntem sunulmuştur. Her iki yöntemin de ilk adımında, gürültülü uzak alan verisinden bilinmeyen cismin yakınındaki saçılan alan belirlenmeye çalışılır. Bu amaçla her iki yöntem için de tek katman potansiyel yaklaşımından faydalanılmış ve uzak alan verisi dairesel bir bölge üzerine bir tek katman potansiyel yoğunluğu anlamında devam ettirilmiştir. İlk yöntemde, bu dairesel bölgenin bilinmeyen cismi minimum yarıçapla kapsadığı varsayılmaktadır. Minimum dairenin içerisinde alan, bilinmeyen cismin yüzeyine kadar, daha önce bulunan potansiyel yoğunluğu aracılığıyla hesaplanmış saçılan alanın Taylor serisi açılımı kullanılarak analitik olarak devam ettirilir. İkinci yöntem içinse, dairesel bölge cismin içerisine yerleştirilmiş ve bilinmeyen cismi, bilinmeyen yüzeyin dışındaki bölgede cisimle aynı alan dağılımını yaratacak homojen olmayan bir yüzey empedansına sahip bir empedans silindiri olarak modellemek için kullanılmıştır. Her iki yöntemin de son aşamasında, bilinmeyen cismin üzerinde toplam alanın sıfıra gitmesi biçimindeki sınır koşulu kullanılarak şekil bulma problemi Gauss-Newton algoritmasıyla yinelemeli olarak çözülen doğrusal olmayan bir eşitliğin köklerinin bulunmasına indirgenir.The aim of this study is to develop new and efficient methods to reconstruct the shape of inaccessible, perfect electric conducting targets through the use of electromagnetic waves. Within this framework, two different shape reconstruction methods, which handle the ill-posedness and the nonlinearity of the underlying inverse problem separately, are presented. At the initial step, both methods deal with the reconstruction of the scattered field in the vicinity of the unknown target from the noise corrupted far field pattern. To this aim, the single layer potential approach is utilized for both methods and the far field pattern is backpropagated to a circular domain in terms of a single layer potential density. In the first method, the circular domain is assumed to cover the unknown target with minimum radius. Inside the minimum circle, the Taylor series expansion of the scattered field calculated from the potential density is exploited to analytically continue the field to the unknown boundary. For the second method, the circular domain is located inside the target and it is used to model the unknown target as an impedance cylinder having inhomogeneous surface impedance which generates the same field distribution with the target outside of the unknown boundary. At the final step of both methods, by exploiting the boundary condition that the total electric field on the unknown boundary vanishes, the shape reconstruction problem is reduced to finding the roots of a non-linear equation which is solved iteratively via Gauss-Newton Method.DoktoraPh
Through-the-Wall Imaging and Multipath Exploitation
We consider the problem of using electromagnetic sensing to estimate targets in complex environments, such as when they are hidden behind walls and other opaque objects. The often unknown electromagnetic interactions between the target and the surrounding area, make the problem challenging. To improve our results, we exploit information in the multipath of the objects surrounding both the target and the sensors. First, we estimate building layouts by using the jump-diffusion algorithm and employing prior knowledge about typical building layouts. We also take advantage of a detailed physical model that captures the scattering by the inner walls and efficiently utilizes the frequency bandwidth. We then localize targets hidden behind reinforced concrete walls. The sensing signals reflected from the targets are significantly distorted and attenuated by the embedded metal bars. Using the surface formulation of the method of moments, we model the response of the reinforced walls, and incorporate their transmission coefficients into the beamforming method to achieve better estimation accuracy. In a related effort, we utilize the sparsity constraint to improve electromagnetic imaging of hidden conducting targets, assuming that a set of equivalent sources can be substituted for the targets. We derive a linear measurement model and employ l1 regularization to identify the equivalent sources in the vicinity of the target surfaces. The proposed inverse method reconstructs the target shape in one or two steps, using single-frequency data. Our results are experimentally verified. Finally, we exploit the multipath from sensor-array platforms to facilitate direction finding. This in contrast to the usual approach, which utilizes the scattering close to the targets. We analyze the effect of the multipath in a statistical signal processing framework, and compute the Cramer-Rao bound to obtain the system resolution. We conduct experiments on a simple array platform to support our theoretical approach
CaSPR: Learning Canonical Spatiotemporal Point Cloud Representations
We propose CaSPR, a method to learn object-centric Canonical Spatiotemporal
Point Cloud Representations of dynamically moving or evolving objects. Our goal
is to enable information aggregation over time and the interrogation of object
state at any spatiotemporal neighborhood in the past, observed or not.
Different from previous work, CaSPR learns representations that support
spacetime continuity, are robust to variable and irregularly spacetime-sampled
point clouds, and generalize to unseen object instances. Our approach divides
the problem into two subtasks. First, we explicitly encode time by mapping an
input point cloud sequence to a spatiotemporally-canonicalized object space. We
then leverage this canonicalization to learn a spatiotemporal latent
representation using neural ordinary differential equations and a generative
model of dynamically evolving shapes using continuous normalizing flows. We
demonstrate the effectiveness of our method on several applications including
shape reconstruction, camera pose estimation, continuous spatiotemporal
sequence reconstruction, and correspondence estimation from irregularly or
intermittently sampled observations.Comment: NeurIPS 202
3D CNN methods in biomedical image segmentation
A definite trend in Biomedical Imaging is the one towards the integration of increasingly complex interpretative layers to the pure data acquisition process. One of the most interesting and looked-forward goals in the field is the automatic segmentation of objects of interest in extensive acquisition data, target that would allow Biomedical Imaging to look beyond its use as a purely assistive tool to become a cornerstone in ambitious large-scale challenges like the extensive quantitative study of the Human Brain.
In 2019 Convolutional Neural Networks represent the state of the art in Biomedical Image segmentation and scientific interests from a variety of fields, spacing from automotive to natural resource exploration, converge to their development. While most of the applications of CNNs are focused on single-image segmentation, biomedical image data -being it MRI, CT-scans, Microscopy, etc- often benefits from three-dimensional volumetric expression.
This work explores a reformulation of the CNN segmentation problem that is native to the 3D nature of the data, with particular interest to the applications to Fluorescence Microscopy volumetric data produced at the European Laboratories for Nonlinear Spectroscopy in the context of two different large international human brain study projects: the Human Brain Project and the White House BRAIN Initiative
Recommended from our members
A Sparsity-Inducing Optimization-Based Algorithm for Planar Patches Extraction from Noisy Point-Cloud Data
Currently, much of the manual labor needed to generate as-built Building Information Models (BIMs) of existing facilities is spent converting raw Point Cloud Datasets (PCDs) to BIMs descriptions. Automating the PCD conversion process can drastically reduce the cost of generating as-built BIMs. Due to the widespread existence of planar structures in civil infrastructures, detecting and extracting planar patches from raw PCDs is a fundamental step in the conversion pipeline from PCDs to BIMs. However, existing methods cannot effectively address both automatically detecting and extracting planar patches from infrastructure PCDs. The existing methods cannot resolve the problem due to the large scale and model complexity of civil infrastructure, or due to the requirements of extra constraints or known information. To address the problem, this paper presents a novel framework for automatically detecting and extracting planar patches from large-scale and noisy raw PCDs. The proposed method automatically detects planar structures, estimates the parametric plane models, and determines the boundaries of the planar patches. The first step recovers existing linear dependence relationships amongst points in the PCD by solving a group-sparsity inducing optimization problem. Next, a spectral clustering procedure based on the recovered linear dependence relationships segments the PCD. Then, for each segmented group, model parameters of the extracted planes are estimated via Singular Value Decomposition (SVD) and Maximum Likelihood Estimation Sample Consensus (MLESAC). Finally, the α-shape algorithm detects the boundaries of planar structures based on a projection of the data to the planar model. The proposed approach is evaluated comprehensively by experiments on two types of PCDs from real-world infrastructures, one captured directly by laser scanners and the other reconstructed from video using structure-from-motion techniques. In order to evaluate the performance comprehensively, five evaluation metrics are proposed which measure different aspects of performance. Experimental results reveal that the proposed method outperforms the existing methods, in the sense that the method automatically and accurately extracts planar patches from large-scaled raw PCDs without any extra constraints nor user assistance.This is the accepted manuscript. The final version is available from Wiley at http://onlinelibrary.wiley.com/doi/10.1111/mice.12063/abstract
Low–High Orthoimage Pairs-Based 3D Reconstruction for Elevation Determination Using Drone
This paper presents a 3D reconstruction method for fast elevation determination on construction sites. The proposed method is intended to automatically and accurately determine construction site elevations using drone-based, low–high orthoimage pairs. This method requires fewer images than other methods for covering a large target area of a construction site. An up–forward–down path was designed to capture approximately -scale images at different altitudes over target stations. A pixel grid matching and elevation determination algorithm was developed to automatically match images in dense pixel grid-style via self-adaptive patch feature descriptors, and simultaneously determine elevations based on a virtual elevation model. The 3D reconstruction results were an elevation map and an orthoimage at each station. Then, the large-scale results of the entire site were easily stitched from adjacent results with narrow overlaps. Moreover, results alignment was automatically performed via the U-net detected ground control point. Experiments validated that in 10–20 and 20–40 orthoimage pairs, 92% of 2,500- and 4,761-pixels were matched in the strongest and strong levels, which was better than sparse reconstructions via structure from motion; moreover, the elevation measurements were as accurate as photogrammetry using multiscale overlapping images
Material Recognition Meets 3D Reconstruction : Novel Tools for Efficient, Automatic Acquisition Systems
For decades, the accurate acquisition of geometry and reflectance properties has represented one of the major objectives in computer vision and computer graphics with many applications in industry, entertainment and cultural heritage. Reproducing even the finest details of surface geometry and surface reflectance has become a ubiquitous prerequisite in visual prototyping, advertisement or digital preservation of objects. However, today's acquisition methods are typically designed for only a rather small range of material types. Furthermore, there is still a lack of accurate reconstruction methods for objects with a more complex surface reflectance behavior beyond diffuse reflectance. In addition to accurate acquisition techniques, the demand for creating large quantities of digital contents also pushes the focus towards fully automatic and highly efficient solutions that allow for masses of objects to be acquired as fast as possible. This thesis is dedicated to the investigation of basic components that allow an efficient, automatic acquisition process. We argue that such an efficient, automatic acquisition can be realized when material recognition "meets" 3D reconstruction and we will demonstrate that reliably recognizing the materials of the considered object allows a more efficient geometry acquisition. Therefore, the main objectives of this thesis are given by the development of novel, robust geometry acquisition techniques for surface materials beyond diffuse surface reflectance, and the development of novel, robust techniques for material recognition. In the context of 3D geometry acquisition, we introduce an improvement of structured light systems, which are capable of robustly acquiring objects ranging from diffuse surface reflectance to even specular surface reflectance with a sufficient diffuse component. We demonstrate that the resolution of the reconstruction can be increased significantly for multi-camera, multi-projector structured light systems by using overlappings of patterns that have been projected under different projector poses. As the reconstructions obtained by applying such triangulation-based techniques still contain high-frequency noise due to inaccurately localized correspondences established for images acquired under different viewpoints, we furthermore introduce a novel geometry acquisition technique that complements the structured light system with additional photometric normals and results in significantly more accurate reconstructions. In addition, we also present a novel method to acquire the 3D shape of mirroring objects with complex surface geometry. The aforementioned investigations on 3D reconstruction are accompanied by the development of novel tools for reliable material recognition which can be used in an initial step to recognize the present surface materials and, hence, to efficiently select the subsequently applied appropriate acquisition techniques based on these classified materials. In the scope of this thesis, we therefore focus on material recognition for scenarios with controlled illumination as given in lab environments as well as scenarios with natural illumination that are given in photographs of typical daily life scenes. Finally, based on the techniques developed in this thesis, we provide novel concepts towards efficient, automatic acquisition systems
Recommended from our members
3D multiple description coding for error resilience over wireless networks
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Mobile communications has gained a growing interest from both customers and service providers alike in the last 1-2 decades. Visual information is used in many application domains such as remote health care, video –on demand, broadcasting, video surveillance etc. In order to enhance the visual effects of digital video content, the depth perception needs to be provided with the actual visual content. 3D video has earned a significant interest from the research community in recent years, due to the tremendous impact it leaves on viewers and its enhancement of the user’s quality of experience (QoE). In the near future, 3D video is likely to be used in most video applications, as it offers a greater sense of immersion and perceptual experience. When 3D video is compressed and transmitted over error prone channels, the associated packet loss leads to visual quality degradation. When a picture is lost or corrupted so severely that the concealment result is not acceptable, the receiver typically pauses video playback and waits for the next INTRA picture to resume decoding. Error propagation caused by employing predictive coding may degrade the video quality severely. There are several ways used to mitigate the effects of such transmission errors. One widely used technique in International Video Coding Standards is error resilience.
The motivation behind this research work is that, existing schemes for 2D colour video compression such as MPEG, JPEG and H.263 cannot be applied to 3D video content. 3D video signals contain depth as well as colour information and are bandwidth demanding, as they require the transmission of multiple high-bandwidth 3D video streams. On the other hand, the capacity of wireless channels is limited and wireless links are prone to various types of errors caused by noise, interference, fading, handoff, error burst and network congestion. Given the maximum bit rate budget to represent the 3D scene, optimal bit-rate allocation between texture and depth information rendering distortion/losses should be minimised. To mitigate the effect of these errors on the perceptual 3D video quality, error resilience video coding needs to be investigated further to offer better quality of experience (QoE) to end users.
This research work aims at enhancing the error resilience capability of compressed 3D video, when transmitted over mobile channels, using Multiple Description Coding (MDC) in order to improve better user’s quality of experience (QoE).
Furthermore, this thesis examines the sensitivity of the human visual system (HVS) when employed to view 3D video scenes. The approach used in this study is to use subjective testing in order to rate people’s perception of 3D video under error free and error prone conditions through the use of a carefully designed bespoke questionnaire.Petroleum Technology Development Fund (PTDF
Codage de cartes de profondeur par deformation de courbes elastiques
In multiple-view video plus depth, depth maps can be represented by means of grayscale images and the corresponding temporal sequence can be thought as a standard grayscale video sequence. However depth maps have different properties from natural images: they present large areas of smooth surfaces separated by sharp edges. Arguably the most important information lies in object contours, as a consequence an interesting approach consists in performing a lossless coding of the contour map, possibly followed by a lossy coding of per-object depth values.In this context, we propose a new technique for the lossless coding of object contours, based on the elastic deformation of curves. A continuous evolution of elastic deformations between two reference contour curves can be modelled, and an elastically deformed version of the reference contours can be sent to the decoder with an extremely small coding cost and used as side information to improve the lossless coding of the actual contour. After the main discontinuities have been captured by the contour description, the depth field inside each region is rather smooth. We proposed and tested two different techniques for the coding of the depth field inside each region. The first technique performs the shape-adaptive wavelet transform followed by the shape-adaptive version of SPIHT. The second technique performs a prediction of the depth field from its subsampled version and the set of coded contours. It is generally recognized that a high quality view rendering at the receiver side is possible only by preserving the contour information, since distortions on edges during the encoding step would cause a sensible degradation on the synthesized view and on the 3D perception. We investigated this claim by conducting a subjective quality assessment test to compare an object-based technique and a hybrid block-based techniques for the coding of depth maps.Dans le format multiple-view video plus depth, les cartes de profondeur peuvent être représentées comme des images en niveaux de gris et la séquence temporelle correspondante peut être considérée comme une séquence vidéo standard en niveaux de gris. Cependant les cartes de profondeur ont des propriétés différentes des images naturelles: ils présentent de grandes surfaces lisses séparées par des arêtes vives. On peut dire que l'information la plus importante réside dans les contours de l'objet, en conséquence une approche intéressante consiste à effectuer un codage sans perte de la carte de contour, éventuellement suivie d'un codage lossy des valeurs de profondeur par-objet.Dans ce contexte, nous proposons une nouvelle technique pour le codage sans perte des contours de l'objet, basée sur la déformation élastique des courbes. Une évolution continue des déformations élastiques peut être modélisée entre deux courbes de référence, et une version du contour déformée élastiquement peut être envoyé au décodeur avec un coût de codage très faible et utilisé comme information latérale pour améliorer le codage sans perte du contour réel. Après que les principales discontinuités ont été capturés par la description du contour, la profondeur à l'intérieur de chaque région est assez lisse. Nous avons proposé et testé deux techniques différentes pour le codage du champ de profondeur à l'intérieur de chaque région. La première technique utilise la version adaptative à la forme de la transformation en ondelette, suivie par la version adaptative à la forme de SPIHT.La seconde technique effectue une prédiction du champ de profondeur à partir de sa version sous-échantillonnée et l'ensemble des contours codés. Il est généralement reconnu qu'un rendu de haute qualité au récepteur pour un nouveau point de vue est possible que avec la préservation de l'information de contour, car des distorsions sur les bords lors de l'étape de codage entraînerait une dégradation évidente sur la vue synthétisée et sur la perception 3D. Nous avons étudié cette affirmation en effectuant un test d'évaluation de la qualité perçue en comparant, pour le codage des cartes de profondeur, une technique basée sur la compression d'objects et une techniques de codage vidéo hybride à blocs
- …