8,709 research outputs found

    Neural network-based colonoscopic diagnosis using on-line learning and differential evolution

    Get PDF
    In this paper, on-line training of neural networks is investigated in the context of computer-assisted colonoscopic diagnosis. A memory-based adaptation of the learning rate for the on-line back-propagation (BP) is proposed and used to seed an on-line evolution process that applies a differential evolution (DE) strategy to (re-) adapt the neural network to modified environmental conditions. Our approach looks at on-line training from the perspective of tracking the changing location of an approximate solution of a pattern-based, and thus, dynamically changing, error function. The proposed hybrid strategy is compared with other standard training methods that have traditionally been used for training neural networks off-line. Results in interpreting colonoscopy images and frames of video sequences are promising and suggest that networks trained with this strategy detect malignant regions of interest with accuracy

    A Comparison of Nature Inspired Algorithms for Multi-threshold Image Segmentation

    Full text link
    In the field of image analysis, segmentation is one of the most important preprocessing steps. One way to achieve segmentation is by mean of threshold selection, where each pixel that belongs to a determined class islabeled according to the selected threshold, giving as a result pixel groups that share visual characteristics in the image. Several methods have been proposed in order to solve threshold selectionproblems; in this work, it is used the method based on the mixture of Gaussian functions to approximate the 1D histogram of a gray level image and whose parameters are calculated using three nature inspired algorithms (Particle Swarm Optimization, Artificial Bee Colony Optimization and Differential Evolution). Each Gaussian function approximates thehistogram, representing a pixel class and therefore a threshold point. Experimental results are shown, comparing in quantitative and qualitative fashion as well as the main advantages and drawbacks of each algorithm, applied to multi-threshold problem.Comment: 16 pages, this is a draft of the final version of the article sent to the Journa

    A hierarchical genetic disparity estimation algorithm for multiview image synthesis

    Get PDF

    Implementation, integration, and optimization of a fuzzy foreground segmentation system

    Get PDF
    Foreground segmentation is often an important preliminary step for various video processing systems. By improving the accuracy of the foreground segmentation process, the overall performance of a video processing system has the potential for improvement. This work introduces a Fuzzy Foreground Segmentation System (FFSS) that uses Mamdani-type Fuzzy Inference Systems (FIS) to control pixel-level accumulated statistics. The error of the FFSS is quantified by comparing its output with hand-segmented ground-truth images from a set of image sequences that specifically model canonical problems of foreground segmentation. Optimization of the FFSS parameters is achieved using a Real-Coded Genetic Algorithm (RCGA). Additionally, multiple central composite designed experiments used to analyze the performance of RCGA under selected schemes and their respective parameters. The RCGA schemes and parameters are chosen as to reduce variation and execution time for a set of known multi-dimensional test functions. The selected multi-dimensional test functions represent assorted function landscapes. To demonstrate accuracy of the FFSS and implicate the importance of the foreground segmentation process, the system is applied to real-time human detection from a single-camera security system. The Human Detection System (HDS) is composed of an IP Camera networked to multiple heterogeneous computers for distributed parallel processing. The implementation of the HDS, adheres to a System of Systems (SoS) architecture which standardizes data/communication, reduces overall complexity, and maintains a high level of interoperability

    Learning spatio-temporal representations for action recognition: A genetic programming approach

    Get PDF
    Extracting discriminative and robust features from video sequences is the first and most critical step in human action recognition. In this paper, instead of using handcrafted features, we automatically learn spatio-temporal motion features for action recognition. This is achieved via an evolutionary method, i.e., genetic programming (GP), which evolves the motion feature descriptor on a population of primitive 3D operators (e.g., 3D-Gabor and wavelet). In this way, the scale and shift invariant features can be effectively extracted from both color and optical flow sequences. We intend to learn data adaptive descriptors for different datasets with multiple layers, which makes fully use of the knowledge to mimic the physical structure of the human visual cortex for action recognition and simultaneously reduce the GP searching space to effectively accelerate the convergence of optimal solutions. In our evolutionary architecture, the average cross-validation classification error, which is calculated by an support-vector-machine classifier on the training set, is adopted as the evaluation criterion for the GP fitness function. After the entire evolution procedure finishes, the best-so-far solution selected by GP is regarded as the (near-)optimal action descriptor obtained. The GP-evolving feature extraction method is evaluated on four popular action datasets, namely KTH, HMDB51, UCF YouTube, and Hollywood2. Experimental results show that our method significantly outperforms other types of features, either hand-designed or machine-learned
    • 

    corecore