272 research outputs found

    Non-Gaussian Hybrid Transfer Functions: Memorizing Mine Survivability Calculations

    Get PDF
    Hybrid algorithms and models have received significant interest in recent years and are increasingly used to solve real-world problems. Different from existing methods in radial basis transfer function construction, this study proposes a novel nonlinear-weight hybrid algorithm involving the non-Gaussian type radial basis transfer functions. The speed and simplicity of the non-Gaussian type with the accuracy and simplicity of radial basis function are used to produce fast and accurate on-the-fly model for survivability of emergency mine rescue operations, that is, the survivability under all conditions is precalculated and used to train the neural network. The proposed hybrid uses genetic algorithm as a learning method which performs parameter optimization within an integrated analytic framework, to improve network efficiency. Finally, the network parameters including mean iteration, standard variation, standard deviation, convergent time, and optimized error are evaluated using the mean squared error. The results demonstrate that the hybrid model is able to reduce the computation complexity, increase the robustness and optimize its parameters. This novel hybrid model shows outstanding performance and is competitive over other existing models

    Controlling realism and uncertainty in reservoir models using intelligent sedimentological prior information

    Get PDF
    Forecasting reservoir production has a large associated uncertainty, since this is the final part of a very complex process, this process is based on sparse and indirect data measurements. One the methodologies used in the oil industry to predict reservoir production is based on the Baye’s theorem. Baye’s theorem applied to reservoir forecasting, samples parameters from a prior understanding of the uncertainty to generate reservoir models and updates this prior information by comparing reservoir production data with model production response. In automatic history matching it is challenging to generate reservoir models that preserve geological realism (obtain reservoir models with geological features that have been seen in nature). One way to control the geological realism in reservoir models is by controlling the realism of the geological prior information. The aim of this thesis is to encapsulate sedimentological information in order to build prior information that can control the geological realism of the history-matched models. This “intelligent” prior information is introduced into the automatic history-matching framework rejecting geologically unrealistic reservoir models. Machine Learning Techniques (MLT) were used to build realistic sedimentological prior information models. Another goal of this thesis was to include geological parameters into the automatic history-match framework that have an impact on reservoir model performance: vertical variation of facies proportions, connectivity of geobodies, and the use of multiple training images as a source of realistic sedimentological prior information. The main outcome of this thesis is that the use of “intelligent” sedimentological prior information guarantees the realism of reservoir models and reduces computing time and uncertainty in reservoir production prediction

    Context-Aware Clustering and the Optimized Whale Optimization Algorithm: An Effective Predictive Model for the Smart Grid

    Get PDF
    For customers to participate in key peak pricing, period-of-use fees, and individualized responsiveness to demand programmes taken from multi-dimensional data flows, energy use projection and analysis must be done well. However, it is a difficult study topic to ascertain the knowledge of use of electricity as recorded in the electricity records' Multi-Dimensional Data Streams (MDDS). Context-Aware Clustering (CAC) and the Optimized Whale Optimization Algorithm were suggested by researchers as a fresh power usage knowledge finding model from the multi-dimensional data streams (MDDS) to resolve issue (OWOA). The proposed CAC-OWOA framework first performs the data cleaning to handle the noisy and null elements. The predictive features are extracted from the novel context-aware group formation algorithm using the statistical context parameters from the pre-processed MDDS electricity logs. To perform the energy consumption prediction, researchers have proposed the novel Artificial Neural Network (ANN) predictive algorithm using the bio-inspired optimization algorithm called OWOA. The OWOA is the modified algorithm of the existing WOA to overcome the problems of slow convergence speed and easily falling into the local optimal solutions. The ANN training method is used in conjunction with the suggested bio-inspired OWOA algorithm to lower error rates and boost overall prediction accuracy. The efficiency of the CAC-OWOA framework is evaluated using the publicly available smart grid electricity consumption logs. The experimental results demonstrate the effectiveness of the CAC-OWOA framework in terms of forecasting accuracy, precision, recall, and duration when compared to underlying approaches

    Bounding the search space for global optimization of neural networks learning error: an interval analysis approach

    Get PDF
    Training a multilayer perceptron (MLP) with algorithms employing global search strategies has been an important research direction in the field of neural networks. Despite a number of significant results, an important matter concerning the bounds of the search region---typically defined as a box---where a global optimization method has to search for a potential global minimizer seems to be unresolved. The approach presented in this paper builds on interval analysis and attempts to define guaranteed bounds in the search space prior to applying a global search algorithm for training an MLP. These bounds depend on the machine precision and the term guaranteed denotes that the region defined surely encloses weight sets that are global minimizers of the neural network's error function. Although the solution set to the bounding problem for an MLP is in general non-convex, the paper presents the theoretical results that help deriving a box which is a convex set. This box is an outer approximation of the algebraic solutions to the interval equations resulting from the function implemented by the network nodes. An experimental study using well known benchmarks is presented in accordance with the theoretical results

    Review of Deep Learning Algorithms and Architectures

    Get PDF
    Deep learning (DL) is playing an increasingly important role in our lives. It has already made a huge impact in areas, such as cancer diagnosis, precision medicine, self-driving cars, predictive forecasting, and speech recognition. The painstakingly handcrafted feature extractors used in traditional learning, classification, and pattern recognition systems are not scalable for large-sized data sets. In many cases, depending on the problem complexity, DL can also overcome the limitations of earlier shallow networks that prevented efficient training and abstractions of hierarchical representations of multi-dimensional training data. Deep neural network (DNN) uses multiple (deep) layers of units with highly optimized algorithms and architectures. This paper reviews several optimization methods to improve the accuracy of the training and to reduce training time. We delve into the math behind training algorithms used in recent deep networks. We describe current shortcomings, enhancements, and implementations. The review also covers different types of deep architectures, such as deep convolution networks, deep residual networks, recurrent neural networks, reinforcement learning, variational autoencoders, and others.https://doi.org/10.1109/ACCESS.2019.291220

    Channel Equalization using GA Family

    Get PDF
    High speed data transmissions over communication channels distort the trans- mitted signals in both amplitude and phase due to presence of Inter Symbol Inter- ference (ISI). Other impairments like thermal noise, impulse noise and cross talk also cause further distortions to the received symbols. Adaptive equalization of the digital channels at the receiver removes/reduces the e®ects of such ISIs and attempts to recover the transmitted symbols. Basically an equalizer is an inverse ¯lter which is placed at the front end of the receiver. Its transfer function is inverse to the transfer function of the associated channel. The Least-Mean-Square (LMS), Recursive-Least-Square (RLS) and Multilayer perceptron (MLP) based adaptive equalizers aim to minimize the ISI present in the digital communication channel. These are gradient based learning algorithms and therefore there is possibility that during training of the equalizers, its weights do not reach to their optimum values due to ..

    Analysis of physiological signals using machine learning methods

    Get PDF
    Technological advances in data collection enable scientists to suggest novel approaches, such as Machine Learning algorithms, to process and make sense of this information. However, during this process of collection, data loss and damage can occur for reasons such as faulty device sensors or miscommunication. In the context of time-series data such as multi-channel bio-signals, there is a possibility of losing a whole channel. In such cases, existing research suggests imputing the missing parts when the majority of data is available. One way of understanding and classifying complex signals is by using deep neural networks. The hyper-parameters of such models have been optimised using the process of back propagation. Over time, improvements have been suggested to enhance this algorithm. However, an essential drawback of the back propagation can be the sensitivity to noisy data. This thesis proposes two novel approaches to address the missing data challenge and back propagation drawbacks: First, suggesting a gradient-free model in order to discover the optimal hyper-parameters of a deep neural network. The complexity of deep networks and high-dimensional optimisation parameters presents challenges to find a suitable network structure and hyper-parameter configuration. This thesis proposes the use of a minimalist swarm optimiser, Dispersive Flies Optimisation(DFO), to enable the selected model to achieve better results in comparison with the traditional back propagation algorithm in certain conditions such as limited number of training samples. The DFO algorithm offers a robust search process for finding and determining the hyper-parameter configurations. Second, imputing whole missing bio-signals within a multi-channel sample. This approach comprises two experiments, namely the two-signal and five-signal imputation models. The first experiment attempts to implement and evaluate the performance of a model mapping bio-signals from A toB and vice versa. Conceptually, this is an extension to transfer learning using CycleGenerative Adversarial Networks (CycleGANs). The second experiment attempts to suggest a mechanism imputing missing signals in instances where multiple data channels are available for each sample. The capability to map to a target signal through multiple source domains achieves a more accurate estimate for the target domain. The results of the experiments performed indicate that in certain circumstances, such as having a limited number of samples, finding the optimal hyper-parameters of a neural network using gradient-free algorithms outperforms traditional gradient-based algorithms, leading to more accurate classification results. In addition, Generative Adversarial Networks could be used to impute the missing data channels in multi-channel bio-signals, and the generated data used for further analysis and classification tasks

    Machine Learning based Signal Generation Strategies for High-Speed Optical Transmitters

    Get PDF
    Optical communication is the only viable solution to respond to the demand for a high bit rate and long transmission distance. Directly modulated lasers (DMLs) are a cheap solution for modulating the light in optical fibre. Moreover, their hardware is simpler than externally modulated lasers. However, DML is inherently chirped and the transmission length with high bit rate is limited. This work explores and implements neural networks based signal predistortion schemes to create transmitters
    corecore