95 research outputs found
Automatic pharynx and larynx cancer segmentation framework (PLCSF) on contrast enhanced MR images
A novel and effective pharynx and larynx cancer segmentation framework (PLCSF) is presented for automatic base of tongue and larynx cancer segmentation from gadolinium-enhanced T1-weighted magnetic resonance images (MRI). The aim of the proposed PLCSF is to assist clinicians in radiotherapy treatment planning. The initial processing of MRI data in PLCSF includes cropping of region of interest; reduction of artefacts and detection of the throat region for the location prior. Further, modified fuzzy c-means clustering is developed to robustly separate candidate cancer pixels from other tissue types. In addition, region-based level set method is evolved to ensure spatial smoothness for the final segmentation boundary after noise removal using non-linear and morphological filtering. Validation study of PLCSF on 102 axial MRI slices demonstrate mean dice similarity coefficient of 0.79 and mean modified Hausdorff distance of 2.2 mm when compared with manual segmentations. Comparison of PLCSF with other algorithms validates the robustness of the PLCSF. Inter- and intra-variability calculations from manual segmentations suggest that PLCSF can help to reduce the human subjectivity
BEMDEC: An Adaptive and Robust Methodology for Digital Image Feature Extraction
The intriguing study of feature extraction, and edge detection in particular, has, as a result of the increased use of imagery, drawn even more attention not just from the field of computer science but also from a variety of scientific fields. However, various challenges surrounding the formulation of feature extraction operator, particularly of edges, which is capable of satisfying the necessary properties of low probability of error (i.e., failure of marking true edges), accuracy, and consistent response to a single edge, continue to persist. Moreover, it should be pointed out that most of the work in the area of feature extraction has been focused on improving many of the existing approaches rather than devising or adopting new ones. In the image processing subfield, where the needs constantly change, we must equally change the way we think.
In this digital world where the use of images, for variety of purposes, continues to increase, researchers, if they are serious about addressing the aforementioned limitations, must be able to think outside the box and step away from the usual in order to overcome these challenges. In this dissertation, we propose an adaptive and robust, yet simple, digital image features detection methodology using bidimensional empirical mode decomposition (BEMD), a sifting process that decomposes a signal into its two-dimensional (2D) bidimensional intrinsic mode functions (BIMFs). The method is further extended to detect corners and curves, and as such, dubbed as BEMDEC, indicating its ability to detect edges, corners and curves. In addition to the application of BEMD, a unique combination of a flexible envelope estimation algorithm, stopping criteria and boundary adjustment made the realization of this multi-feature detector possible. Further application of two morphological operators of binarization and thinning adds to the quality of the operator
Adaptive kernel estimation for enhanced filtering and pattern classification of magnetic resonance imaging: novel techniques for evaluating the biomechanics and pathologic conditions of the lumbar spine
This dissertation investigates the contribution the lumbar spine musculature has on etiological and pathogenic characteristics of low back pain and lumbar spondylosis. This endeavor necessarily required a two-step process: 1) design of an accurate post-processing method for extracting relevant information via magnetic resonance images and 2) determine pathological trends by elucidating high-dimensional datasets through multivariate pattern classification. The lumbar musculature was initially evaluated by post-processing and segmentation of magnetic resonance (MR) images of the lumbar spine, which characteristically suffer from nonlinear corruption of the signal intensity. This so called intensity inhomogeneity degrades the efficacy of traditional intensity-based segmentation algorithms. Proposed in this dissertation is a solution for filtering individual MR images by extracting a map of the underlying intensity inhomogeneity to adaptively generate local estimates of the kernel’s optimal bandwidth. The adaptive kernel is implemented and tested within the structure of the non-local means filter, but also generalized and extended to the Gaussian and anisotropic diffusion filters. Testing of the proposed filters showed that the adaptive kernel significantly outperformed their non-adaptive counterparts. A variety of performance metrics were utilized to measure either fine feature preservation or accuracy of post-processed segmentation. Based on these metrics the adaptive filters proposed in this dissertation significantly outperformed the non-adaptive versions. Using the proposed filter, the MR data was semi-automatically segmented to delineate between adipose and lean muscle tissues. Two important findings were reached utilizing this data. First, a clear distinction between the musculature of males and females was established that provided 100% accuracy in being able to predict gender. Second, degenerative lumbar spines were accurately predicted at a rate of up to 92% accuracy. These results solidify prior assumptions made regarding sexual dimorphic anatomy and the pathogenic nature of degenerative spine disease
Faster and better: a machine learning approach to corner detection
The repeatability and efficiency of a corner detector determines how likely
it is to be useful in a real-world application. The repeatability is importand
because the same scene viewed from different positions should yield features
which correspond to the same real-world 3D locations [Schmid et al 2000]. The
efficiency is important because this determines whether the detector combined
with further processing can operate at frame rate.
Three advances are described in this paper. First, we present a new heuristic
for feature detection, and using machine learning we derive a feature detector
from this which can fully process live PAL video using less than 5% of the
available processing time. By comparison, most other detectors cannot even
operate at frame rate (Harris detector 115%, SIFT 195%). Second, we generalize
the detector, allowing it to be optimized for repeatability, with little loss
of efficiency. Third, we carry out a rigorous comparison of corner detectors
based on the above repeatability criterion applied to 3D scenes. We show that
despite being principally constructed for speed, on these stringent tests, our
heuristic detector significantly outperforms existing feature detectors.
Finally, the comparison demonstrates that using machine learning produces
significant improvements in repeatability, yielding a detector that is both
very fast and very high quality.Comment: 35 pages, 11 figure
Constructing 3D faces from natural language interface
This thesis presents a system by which 3D images of human faces can be constructed
using a natural language interface. The driving force behind the project was the need to
create a system whereby a machine could produce artistic images from verbal or
composed descriptions. This research is the first to look at constructing and modifying
facial image artwork using a natural language interface.
Specialised modules have been developed to control geometry of 3D polygonal head
models in a commercial modeller from natural language descriptions. These modules
were produced from research on human physiognomy, 3D modelling techniques and
tools, facial modelling and natural language processing. [Continues.
Scaling Multidimensional Inference for Big Structured Data
In information technology, big data is a collection of data sets so large and complex that it becomes difficult to process using traditional data processing applications [151]. In a
world of increasing sensor modalities, cheaper storage, and more data oriented questions, we are quickly passing the limits of tractable computations using traditional statistical analysis
methods. Methods which often show great results on simple data have difficulties processing complicated multidimensional data. Accuracy alone can no longer justify unwarranted memory
use and computational complexity. Improving the scaling properties of these methods for multidimensional data is the only way to make these methods relevant. In this work we explore methods for improving the scaling properties of parametric and nonparametric
models. Namely, we focus on the structure of the data to lower the complexity of a specific family of problems. The two types of structures considered in this work are distributive
optimization with separable constraints (Chapters 2-3), and scaling Gaussian processes for multidimensional lattice input (Chapters 4-5). By improving the scaling of these methods, we can expand their use to a wide range of applications which were previously intractable
open the door to new research questions
Recommended from our members
An investigation on the framework of dressing virtual humans
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.Realistic human models are widely used in variety of applications. Much research has been carried out on improving realism of virtual humans from various aspects, such as body shapes, hair, and facial expressions and so on. In most occasions, these virtual humans need to wear garments. However, it is time-consuming and tedious to dress a human model using current software packages [Maya2004]. Several methods for dressing virtual humans have been proposed recently [Bourguignon2001, Turquin2004, Turquin2007 and Wang2003B]. The method proposed by Bourguignon et al [Bourguignon2001] can only generate 3D garment contour instead of 3D surface. The method presented by Turquin et al. [Turquin2004, Turquin2007] could generate various kinds of garments from sketches but their garments followed the shape of the body and the side of a garment looked not convincing because of using simple linear interpolation. The method proposed by Wang et al. [Wang2003B] lacked interactivity from users, so users had very limited control on the garment shape.This thesis proposes a framework for dressing virtual humans to obtain convincing dressing results, which overcomes problems existing in previous papers mentioned above by using nonlinear interpolation, level set-based shape modification, feature constraints and so on. Human models used in this thesis are reconstructed from real human body data obtained using a body scanning system. Semantic information is then extracted from human models to assist in generation of 3 dimensional (3D) garments. The proposed framework allows users to dress virtual humans using garment patterns and sketches. The proposed dressing method is based on semantic virtual humans. A semantic human model is a human body with semantic information represented by certain of structure and body features. The semantic human body is reconstructed from body scanned data from a real human body. After segmenting the human model into six parts some key features are extracted. These key features are used as constraints for garment construction.Simple 3D garment patterns are generated using the techniques of sweep and offset. To dress a virtual human, users just choose a garment pattern, which is put on the human body at the default position with a default size automatically. Users are allowed to change simple parameters to specify some sizes of a garment by sketching the desired position on the human body.To enable users to dress virtual humans by their own design styles in an intuitive way, this thesis proposes an approach for garment generation from user-drawn sketches. Users can directly draw sketches around reconstructed human bodies and then generates 3D garments based on user-drawn strokes. Some techniques for generating 3D garments and dressing virtual humans are proposed. The specific focus of the research lies in generation of 3D geometric garments, garment shape modification, local shape modification, garment surface processing and decoration creation. A sketch-based interface has been developed allowing users to draw garment contour representing the front-view shape of a garment, and the system can generate a 3D geometric garment surface accordingly. To improve realism of a garment surface, this thesis presents three methods as follows. Firstly, the procedure of garment vertices generation takes key body features as constraints. Secondly, an optimisation algorithm is carried out after generation of garment vertices to optimise positions of garment vertices. Finally, some mesh processing schemes are applied to further process the garment surface. Then, an elaborate 3D geometric garment surface can be obtained through this series of processing. Finally, this thesis proposes some modification and editing methods. The user-drawn sketches are processed into spline curves, which allow users to modify the existing garment shape by dragging the control points into desired positions. This makes it easy for users to obtain a more satisfactory garment shape compared with the existing one. Three decoration tools including a 3D pen, a brush and an embroidery tool, are provided letting users decorate the garment surface by adding some small 3D details such as brand names, symbols and so on. The prototype of the framework is developed using Microsoft Visual Studio C++,OpenGL and GPU programming
- …