234 research outputs found

    Model migration neural network for predicting battery aging trajectories

    Get PDF
    Accurate prediction of batteries’ future degradation is a key solution to relief users’ anxiety on battery lifespan and electric vehicle’s driving range. Technical challenges arise from the highly nonlinear dynamics of battery aging. In this paper, a feed-forward migration neural network is proposed to predict the batteries’ aging trajectories. Specifically, a base model that describes the capacity decay over time is first established from the existed battery aging dataset. This base model is then transformed by an input-output slope-and-bias-correction (SBC) method structure to capture the degradation of target cell. To enhance the model’s nonlinear transfer capability, the SBC-model is further integrated into a four-layer neural network, and easily trained via the gradient correlation algorithm. The proposed migration neural network is experimentally verified with four different commercial batteries. The predicted RMSEs are all lower than 2.5% when using only the first 30% of aging trajectories for neural network training. In addition, illustrative results demonstrate that a small size feed-forward neural network (down to 1-5-5-1) is sufficient for battery aging trajectory prediction

    Determination of Elevations for Excavation Operations Using Drone Technologies

    Get PDF
    Using deep learning technology to rapidly estimate depth information from a single image has been studied in many situations, but it is new in construction site elevation determinations, and challenges are not limited to the lack of datasets. This dissertation presents the research results of utilizing drone ortho-imaging and deep learning to estimate construction site elevations for excavation operations. It provides two flexible options of fast elevation determination including a low-high-ortho-image-pair-based method and a single-frame-ortho-image-based method. The success of this research project advanced the ortho-imaging utilization in construction surveying, strengthened CNNs (convolutional neural networks) to work with large scale images, and contributed dense image pixel matching with different scales.This research project has three major tasks. First, the high-resolution ortho-image and elevation-map datasets were acquired using the low-high ortho-image pair-based 3D-reconstruction method. In detail, a vertical drone path is designed first to capture a 2:1 scale ortho-image pair of a construction site at two different altitudes. Then, to simultaneously match the pixel pairs and determine elevations, the developed pixel matching and virtual elevation algorithm provides the candidate pixel pairs in each virtual plane for matching, and the four-scaling patch feature descriptors are used to match them. Experimental results show that 92% of pixels in the pixel grid were strongly matched, where the accuracy of elevations was within ±5 cm.Second, the acquired high-resolution datasets were applied to train and test the ortho-image encoder and elevation-map decoder, where the max-pooling and up-sampling layers link the ortho-image and elevation-map in the same pixel coordinate. This convolutional encoder-decoder was supplemented with an input ortho-image overlapping disassembling and output elevation-map assembling algorithm to crop the high-resolution datasets into multiple small-patch datasets for model training and testing. Experimental results indicated 128×128-pixel small-patch had the best elevation estimation performance, where 21.22% of the selected points were exactly matched with “ground truth,” 31.21% points were accurately matched within ±5 cm. Finally, vegetation was identified in high-resolution ortho-images and removed from corresponding elevation-maps using the developed CNN-based image classification model and the vegetation removing algorithm. Experimental results concluded that the developed CNN model using 32×32-pixel ortho-image and class-label small-patch datasets had 93% accuracy in identifying objects and localizing objects’ edges

    Early stopping by correlating online indicators in neural networks

    Get PDF
    Financiado para publicación en acceso aberto: Universidade de Vigo/CISUGinfo:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2013-2016/TIN2017-85160-C2-2-R/ES/AVANCES EN NUEVOS SISTEMAS DE EXTRACCION DE RESPUESTAS CON ANALISIS SEMANTICO Y APRENDIZAJE PROFUNDOinfo:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2017-2020/PID2020-113230RB-C22/ES/SEQUENCE LABELING MULTITASK MODELS FOR LINGUISTICALLY ENRICHED NER: SEMANTICS AND DOMAIN ADAPTATION (SCANNER-UVIGO)In order to minimize the generalization error in neural networks, a novel technique to identify overfitting phenomena when training the learner is formally introduced. This enables support of a reliable and trustworthy early stopping condition, thus improving the predictive power of that type of modeling. Our proposal exploits the correlation over time in a collection of online indicators, namely characteristic functions for indicating if a set of hypotheses are met, associated with a range of independent stopping conditions built from a canary judgment to evaluate the presence of overfitting. That way, we provide a formal basis for decision making in terms of interrupting the learning process. As opposed to previous approaches focused on a single criterion, we take advantage of subsidiarities between independent assessments, thus seeking both a wider operating range and greater diagnostic reliability. With a view to illustrating the effectiveness of the halting condition described, we choose to work in the sphere of natural language processing, an operational continuum increasingly based on machine learning. As a case study, we focus on parser generation, one of the most demanding and complex tasks in the domain. The selection of cross-validation as a canary function enables an actual comparison with the most representative early stopping conditions based on overfitting identification, pointing to a promising start toward an optimal bias and variance control.Agencia Estatal de Investigación | Ref. TIN2017-85160-C2-2-RAgencia Estatal de Investigación | Ref. PID2020-113230RB-C22Xunta de Galicia | Ref. ED431C 2018/5

    Human inspired pattern recognition via local invariant features

    Get PDF
    Vision is increasingly becoming a vital element in the manufacturing industry. As complex as it already is, vision is becoming even more difficult to implement in a pattern recognition environment as it converges toward the level of what humans visualize. Relevant brain work technologies are allowing vision systems to add capability and tasks that were long reserved for humans. The ability to recognize patterns like humans do is a good goal in terms of performance metrics for manufacturing activities. To achieve this goal, we created a neural network that achieves pattern recognition analogous to the human visual cortex using high quality keypoints by optimizing the scale space and pairing keypoints with edges as input into the model. This research uses the Taguchi Design of Experiments approach to find optimal values for the SIFT parameters with respect to finding correct matches between images that vary in rotation and scale. The approach used the Taguchi L18 matrix to determine the optimal parameter set. The performance obtained from SIFT using the optimal solution was compared with the performance from the original SIFT algorithm parameters. It is shown that correct matches between an original image and a scaled, rotated, or scaled and rotated version of that image improves by 17% using the optimal values of the SIFT. A human inspired approach was used to create a CMAC based neural network capable of pattern recognition. A comparison of 3 object, 30 object, and 50 object scenes were examined using edge and optimized SIFT based features as inputs and produced extensible results from 3 to 50 objects based on classification performance. The classification results prove that we achieve a high level of pattern recognition that ranged from 96.1% to 100% for objects under consideration. The result is a pattern recognition model capable of locally based classification based on invariant information without the need for sets of information that include input sensory data that is not necessarily invariant (background data, raw pixel data, viewpoint angles) that global models rely on in pattern recognition

    LEyes: A Lightweight Framework for Deep Learning-Based Eye Tracking using Synthetic Eye Images

    Full text link
    Deep learning has bolstered gaze estimation techniques, but real-world deployment has been impeded by inadequate training datasets. This problem is exacerbated by both hardware-induced variations in eye images and inherent biological differences across the recorded participants, leading to both feature and pixel-level variance that hinders the generalizability of models trained on specific datasets. While synthetic datasets can be a solution, their creation is both time and resource-intensive. To address this problem, we present a framework called Light Eyes or "LEyes" which, unlike conventional photorealistic methods, only models key image features required for video-based eye tracking using simple light distributions. LEyes facilitates easy configuration for training neural networks across diverse gaze-estimation tasks. We demonstrate that models trained using LEyes are consistently on-par or outperform other state-of-the-art algorithms in terms of pupil and CR localization across well-known datasets. In addition, a LEyes trained model outperforms the industry standard eye tracker using significantly more cost-effective hardware. Going forward, we are confident that LEyes will revolutionize synthetic data generation for gaze estimation models, and lead to significant improvements of the next generation video-based eye trackers.Comment: 32 pages, 8 figure

    A Data-driven Approach for Detecting Stress in Plants Using Hyperspectral Imagery

    Get PDF
    A phenotype is an observable characteristic of an individual and is a function of its genotype and its growth environment. Individuals with different genotypes are impacted differently by exposure to the same environment. Therefore, phenotypes are often used to understand morphological and physiological changes in plants as a function of genotype and biotic and abiotic stress conditions. Phenotypes that measure the level of stress can help mitigate the adverse impacts on the growth cycle of the plant. Image-based plant phenotyping has the potential for early stress detection by means of computing responsive phenotypes in a non-intrusive manner. A large number of plants grown and imaged under a controlled environment in a high-throughput plant phenotyping (HTPP) system are increasingly becoming accessible to research communities. They can be useful to compute novel phenotypes for early stress detection. In early stages of stress induction, plants manifest responses in terms of physiological changes rather than morphological, making it difficult to detect using visible spectrum cameras which use only three wide spectral bands in the 380nm - 740 nm range. In contrast, hyperspectral imaging can capture a broad range of wavelengths (350nm - 2500nm) with narrow spectral bands (5nm). Hyperspectral imagery (HSI), therefore, provides rich spectral information which can help identify and track even small changes in plant physiology in response to stress. In this research, a data-driven approach has been developed to identify regions in plants that manifest abnormal reflectance patterns after stress induction. Reflectance patterns of age-matched unstressed plants are first characterized. The normal and stressed reflectance patterns are used to train a classifier that can predict if a point in the plant is stressed or not. Stress maps of a plant can be generated from its hyperspectral image and can be used to track the temporal propagation of stress. These stress maps are used to compute novel phenotypes that represent the level of stress in a plant and the stress trajectory over time. The data-driven approach is validated using a dataset of sorghum plants exposed to drought stress in a LemnaTec Scanalyzer 3D HTPP system. Advisers: Ashok Samal and Sruti Das Choudhur

    3D Detection and Characterisation of ALMA Sources through Deep Learning

    Full text link
    We present a Deep-Learning (DL) pipeline developed for the detection and characterization of astronomical sources within simulated Atacama Large Millimeter/submillimeter Array (ALMA) data cubes. The pipeline is composed of six DL models: a Convolutional Autoencoder for source detection within the spatial domain of the integrated data cubes, a Recurrent Neural Network (RNN) for denoising and peak detection within the frequency domain, and four Residual Neural Networks (ResNets) for source characterization. The combination of spatial and frequency information improves completeness while decreasing spurious signal detection. To train and test the pipeline, we developed a simulation algorithm able to generate realistic ALMA observations, i.e. both sky model and dirty cubes. The algorithm simulates always a central source surrounded by fainter ones scattered within the cube. Some sources were spatially superimposed in order to test the pipeline deblending capabilities. The detection performances of the pipeline were compared to those of other methods and significant improvements in performances were achieved. Source morphologies are detected with subpixel accuracies obtaining mean residual errors of 10310^{-3} pixel (0.10.1 mas) and 10110^{-1} mJy/beam on positions and flux estimations, respectively. Projection angles and flux densities are also recovered within 10%10\% of the true values for 80%80\% and 73%73\% of all sources in the test set, respectively. While our pipeline is fine-tuned for ALMA data, the technique is applicable to other interferometric observatories, as SKA, LOFAR, VLBI, and VLTI

    Generalised fourier analysis of human chromosome images

    Get PDF
    corecore