590 research outputs found

    Sensor Scheduling with Intelligent Optimization Algorithm Based on Quantum Theory

    Get PDF
    The particle swarm optimization (PSO) algorithm superiority exists in convergence rate, but it tends to get stuck in local optima. An improved PSO algorithm is proposed using a best dimension mutation technique based on quantum theory, and it was applied to sensor scheduling problem for target tracking. The dynamics of the target are assumed as linear Gaussian model, and the sensor measurements show a linear correlation with the state of the target. This paper discusses the single target tracking problem with multiple sensors using the proposed best dimension mutation particle swarm optimization (BDMPSO) algorithm for various cases. Our experimental results verify that the proposed algorithm is able to track the target more reliably and accurately than previous ones

    A New High-Speed Foreign Fiber Detection System with Machine Vision

    Get PDF
    A new high-speed foreign fiber detection system with machine vision is proposed for removing foreign fibers from raw cotton using optimal hardware components and appropriate algorithms designing. Starting from a specialized lens of 3-charged couple device (CCD) camera, the system applied digital signal processor (DSP) and field-programmable gate array (FPGA) on image acquisition and processing illuminated by ultraviolet light, so as to identify transparent objects such as polyethylene and polypropylene fabric from cotton tuft flow by virtue of the fluorescent effect, until all foreign fibers that have been blown away safely by compressed air quality can be achieved. An image segmentation algorithm based on fast wavelet transform is proposed to identify block-like foreign fibers, and an improved canny detector is also developed to segment wire-like foreign fibers from raw cotton. The procedure naturally provides color image segmentation method with region growing algorithm for better adaptability. Experiments on a variety of images show that the proposed algorithms can effectively segment foreign fibers from test images under various circumstances

    Mathematical Optimisation of Magnetic Nanoparticle Diffusion in the Brain White Matter

    Get PDF
    European Unions Horizon 2020 research and innovation programme under Grant Agreement No. 688279. EPSRC Established Career Fellowship Grant No. EP/N025954/1. Children with Cancer UK Grant No. 16-224.Peer reviewedPublisher PD

    A Method for Recognizing Fatigue Driving Based on Dempster-Shafer Theory and Fuzzy Neural Network

    Get PDF
    This study proposes a method based on Dempster-Shafer theory (DST) and fuzzy neural network (FNN) to improve the reliability of recognizing fatigue driving. This method measures driving states using multifeature fusion. First, FNN is introduced to obtain the basic probability assignment (BPA) of each piece of evidence given the lack of a general solution to the definition of BPA function. Second, a modified algorithm that revises conflict evidence is proposed to reduce unreasonable fusion results when unreliable information exists. Finally, the recognition result is given according to the combination of revised evidence based on Dempster’s rule. Experiment results demonstrate that the recognition method proposed in this paper can obtain reasonable results with the combination of information given by multiple features. The proposed method can also effectively and accurately describe driving states

    MAT: Mask-Aware Transformer for Large Hole Image Inpainting

    Full text link
    Recent studies have shown the importance of modeling long-range interactions in the inpainting problem. To achieve this goal, existing approaches exploit either standalone attention techniques or transformers, but usually under a low resolution in consideration of computational cost. In this paper, we present a novel transformer-based model for large hole inpainting, which unifies the merits of transformers and convolutions to efficiently process high-resolution images. We carefully design each component of our framework to guarantee the high fidelity and diversity of recovered images. Specifically, we customize an inpainting-oriented transformer block, where the attention module aggregates non-local information only from partial valid tokens, indicated by a dynamic mask. Extensive experiments demonstrate the state-of-the-art performance of the new model on multiple benchmark datasets. Code is released at https://github.com/fenglinglwb/MAT.Comment: Accepted to CVPR2022 Ora

    A novel regional-minima image segmentation method for fluid transport simulations in unresolved rock images

    Get PDF
    This study would not be possible without digital rock images provided by Digital Rocks Portal and its contributors (https://www.digitalrocksportal.org/). .Peer reviewe

    BLIVA: A Simple Multimodal LLM for Better Handling of Text-Rich Visual Questions

    Full text link
    Vision Language Models (VLMs), which extend Large Language Models (LLM) by incorporating visual understanding capability, have demonstrated significant advancements in addressing open-ended visual question-answering (VQA) tasks. However, these models cannot accurately interpret images infused with text, a common occurrence in real-world scenarios. Standard procedures for extracting information from images often involve learning a fixed set of query embeddings. These embeddings are designed to encapsulate image contexts and are later used as soft prompt inputs in LLMs. Yet, this process is limited to the token count, potentially curtailing the recognition of scenes with text-rich context. To improve upon them, the present study introduces BLIVA: an augmented version of InstructBLIP with Visual Assistant. BLIVA incorporates the query embeddings from InstructBLIP and also directly projects encoded patch embeddings into the LLM, a technique inspired by LLaVA. This approach assists the model to capture intricate details potentially missed during the query decoding process. Empirical evidence demonstrates that our model, BLIVA, significantly enhances performance in processing text-rich VQA benchmarks (up to 17.76% in OCR-VQA benchmark) and in undertaking general (not particularly text-rich) VQA benchmarks (up to 7.9% in Visual Spatial Reasoning benchmark), and achieved 17.72% overall improvement in a comprehensive multimodal LLM benchmark (MME), comparing to our baseline InstructBLIP. BLIVA demonstrates significant capability in decoding real-world images, irrespective of text presence. To demonstrate the broad industry applications enabled by BLIVA, we evaluate the model using a new dataset comprising YouTube thumbnails paired with question-answer sets across 11 diverse categories. Our code and models are freely accessible at https://github.com/mlpc-ucsd/BLIVA.Comment: Accepted at AAAI Conference on Artificial Intelligence (AAAI-24
    corecore