111 research outputs found

    Forecasting model for short-term wind speed using robust local mean decomposition, deep neural networks, intelligent algorithm, and error correction

    Get PDF
    Wind power generation has aroused widespread concern worldwide. Accurate prediction of wind speed is very important for the safe and economic operation of the power grid. This paper presents a short-term wind speed prediction model which includes data decomposition, deep learning, intelligent algorithm optimization, and error correction modules. First, the robust local mean decomposition (RLMD) is applied to the original wind speed data to reduce the non-stationarity of the data. Then, the salp swarm algorithm (SSA) is used to determine the optimal parameter combination of the bidirectional gated recurrent unit (BiGRU) to ensure prediction quality. In order to eliminate the predictable components of the error further, a correction module based on the improved salp swarm algorithm (ISSA) and deep extreme learning machine (DELM) is constructed. The exploration and exploitation capability of the original SSA is enhanced by introducing a crazy operator and dynamic learning strategy, and the input weights and thresholds in the DELM are optimized by the ISSA to improve the generalization ability of the model. The actual data of wind farms are used to verify the advancement of the proposed model. Compared with other models, the results show that the proposed model has the best prediction performance. As a powerful tool, the developed forecasting system is expected to be further used in the energy system

    Machine learning applications in search algorithms for gravitational waves from compact binary mergers

    Get PDF
    Gravitational waves from compact binary mergers are now routinely observed by Earth-bound detectors. These observations enable exciting new science, as they have opened a new window to the Universe. However, extracting gravitational-wave signals from the noisy detector data is a challenging problem. The most sensitive search algorithms for compact binary mergers use matched filtering, an algorithm that compares the data with a set of expected template signals. As detectors are upgraded and more sophisticated signal models become available, the number of required templates will increase, which can make some sources computationally prohibitive to search for. The computational cost is of particular concern when low-latency alerts should be issued to maximize the time for electromagnetic follow-up observations. One potential solution to reduce computational requirements that has started to be explored in the last decade is machine learning. However, different proposed deep learning searches target varying parameter spaces and use metrics that are not always comparable to existing literature. Consequently, a clear picture of the capabilities of machine learning searches has been sorely missing. In this thesis, we closely examine the sensitivity of various deep learning gravitational-wave search algorithms and introduce new methods to detect signals from binary black hole and binary neutron star mergers at previously untested statistical confidence levels. By using the sensitive distance as our core metric, we allow for a direct comparison of our algorithms to state-of-the-art search pipelines. As part of this thesis, we organized a global mock data challenge to create a benchmark for machine learning search algorithms targeting compact binaries. This way, the tools developed in this thesis are made available to the greater community by publishing them as open source software. Our studies show that, depending on the parameter space, deep learning gravitational-wave search algorithms are already competitive with current production search pipelines. We also find that strategies developed for traditional searches can be effectively adapted to their machine learning counterparts. In regions where matched filtering becomes computationally expensive, available deep learning algorithms are also limited in their capability. We find reduced sensitivity to long duration signals compared to the excellent results for short-duration binary black hole signals

    Remaining Useful Life Prediction of Lithium-ion Batteries using Spatio-temporal Multimodal Attention Networks

    Full text link
    Lithium-ion batteries are widely used in various applications, including electric vehicles and renewable energy storage. The prediction of the remaining useful life (RUL) of batteries is crucial for ensuring reliable and efficient operation, as well as reducing maintenance costs. However, determining the life cycle of batteries in real-world scenarios is challenging, and existing methods have limitations in predicting the number of cycles iteratively. In addition, existing works often oversimplify the datasets, neglecting important features of the batteries such as temperature, internal resistance, and material type. To address these limitations, this paper proposes a two-stage remaining useful life prediction scheme for Lithium-ion batteries using a spatio-temporal multimodal attention network (ST-MAN). The proposed model is designed to iteratively predict the number of cycles required for the battery to reach the end of its useful life, based on available data. The proposed ST-MAN is to capture the complex spatio-temporal dependencies in the battery data, including the features that are often neglected in existing works. Experimental results demonstrate that the proposed ST-MAN model outperforms existing CNN and LSTM-based methods, achieving state-of-the-art performance in predicting the remaining useful life of Li-ion batteries. The proposed method has the potential to improve the reliability and efficiency of battery operations and is applicable in various industries, including automotive and renewable energy

    Salp Swarm Optimized Hybrid Elman Recurrent Neural Network (SSO-ERNN) based MPPT Controller for Solar PV

    Get PDF
    Renewable energy technologies provide clean and abundant energy that can be self-renewed from natural sources; more support from the public to replace fossil fuels with various renewable energy sources to protect the environment. Although solar energy has less impact on the environment than other renewable sources, the output efficiency is lower due to the different weather conditions. So to overcome that, the MPPT controller is used for tracking peak power and better efficiency. Some conventional methods in MPPT controllers provide less tracking efficiency, and steady-state oscillations occur in maximum power tracking due to the sudden variations in solar irradiance. Thus, in this work salp swarm optimized (SSO) based Elman recurrent neural network (ERNN) controller is proposed to track the maximum power form PV with high efficiency. The weight parameter of ERNN layer is optimized with the help of SSO, which solve the complex problems and give maximum efficiency. The proposed method is performed in MATLAB/Simulink environment, which differs from existing plans and gives a better output efficiency. Using this proposed controller, the system can achieve high tracking efficiency of 99.74% compared to conventional processes

    Self-Learning Longitudinal Control for On-Road Vehicles

    Get PDF
    Fahrerassistenzsysteme (Advanced Driver Assistance Systems) sind ein wichtiges Verkaufsargument für PKWs, fordern jedoch hohe Entwicklungskosten. Insbesondere die Parametrierung für Längsregelung, die einen wichtigen Baustein für Fahrerassistenzsysteme darstellt, benötigt viel Zeit und Geld, um die richtige Balance zwischen Insassenkomfort und Regelgüte zu treffen. Reinforcement Learning scheint ein vielversprechender Ansatz zu sein, um dies zu automatisieren. Diese Klasse von Algorithmen wurde bislang allerdings vorwiegend auf simulierte Aufgaben angewendet, die unter idealen Bedingungen stattfinden und nahezu unbegrenzte Trainingszeit ermöglichen. Unter den größten Herausforderungen für die Anwendung von Reinforcement Learning in einem realen Fahrzeug sind Trajektorienfolgeregelung und unvollständige Zustandsinformationen aufgrund von nur teilweise beobachteter Dynamik. Darüber hinaus muss ein Algorithmus, der in realen Systemen angewandt wird, innerhalb von Minuten zu einem Ergebnis kommen. Außerdem kann das Regelziel sich während der Laufzeit beliebig ändern, was eine zusätzliche Schwierigkeit für Reinforcement Learning Methoden darstellt. Diese Arbeit stellt zwei Algorithmen vor, die wenig Rechenleistung benötigen und diese Hürden überwinden. Einerseits wird ein modellfreier Reinforcement Learning Ansatz vorgeschlagen, der auf der Actor-Critic-Architektur basiert und eine spezielle Struktur in der Zustandsaktionswertfunktion verwendet, um mit teilweise beobachteten Systemen eingesetzt werden zu können. Um eine Vorsteuerung zu lernen, wird ein Regler vorgeschlagen, der sich auf eine Projektion und Trainingsdatenmanipulation stützt. Andererseits wird ein modellbasierter Algorithmus vorgeschlagen, der auf Policy Search basiert. Diesem wird eine automatisierte Entwurfsmethode für eine inversionsbasierte Vorsteuerung zur Seite gestellt. Die vorgeschlagenen Algorithmen werden in einer Reihe von Szenarien verglichen, in denen sie online, d.h. während der Fahrt und bei geschlossenem Regelkreis, in einem realen Fahrzeug lernen. Obwohl die Algorithmen etwas unterschiedlich auf verschiedene Randbedingungen reagieren, lernen beide robust und zügig und sind in der Lage, sich an verschiedene Betriebspunkte, wie zum Beispiel Geschwindigkeiten und Gänge, anzupassen, auch wenn Störungen während des Trainings einwirken. Nach bestem Wissen des Autors ist dies die erste erfolgreiche Anwendung eines Reinforcement Learning Algorithmus, der online in einem realen Fahrzeug lernt

    Simulation of carbon peaking process of high energy consuming manufacturing industry in Shaanxi Province: A hybrid model based on LMDI and TentSSA-ENN

    Get PDF
    To achieve the goals of carbon peaking and carbon neutrality in Shaanxi, the high energy consuming manufacturing industry (HMI), as an important contributor, is a key link and important channel for energy conservation. In this paper, the logarithmic mean Divisia index (LMDI) method is applied to determine the driving factors of carbon emissions from the aspects of economy, energy and society, and the contribution of these factors was analyzed. Meanwhile, the improved sparrow search algorithm is used to optimize Elman neural network (ENN) to construct a new hybrid prediction model. Finally, three different development scenarios are designed using scenario analysis method to explore the potential of HMI in Shaanxi Province to achieve carbon peak in the future. The results show that: (1) The biggest promoting factor is industrial structure, and the biggest inhibiting factor is energy intensity among the drivers of carbon emissions, which are analyzed effectively in HMI using the LMDI method. (2) Compared with other neural network models, the proposed hybrid prediction model has higher accuracy and better stability in predicting industrial carbon emissions, it is more suitable for simulating the carbon peaking process of HMI. (3) Only in the coordinated development scenario, the HMI in Shaanxi is likely to achieve the carbon peak in 2030, and the carbon emission curve of the other two scenarios has not reached the peak. Then, according to the results of scenario analysis, specific and evaluable suggestions on carbon emission reduction for HMI in Shaanxi are put forward, such as optimizing energy and industrial structure and making full use of innovative resources of Shaanxi characteristic units

    Self-Learning Longitudinal Control for On-Road Vehicles

    Get PDF
    Reinforcement Learning is a promising tool to automate controller tuning. However, significant extensions are required for real-world applications to enable fast and robust learning. This work proposes several additions to the state of the art and proves their capability in a series of real world experiments

    Statistical Analysis of Delay in Time Series

    Get PDF
    This thesis focuses on delay in time series data. The first delay involves the m-delay autoregressive model. This approach considers only the first and the last previous observation of the traditional autoregressive model. Next, the delay is added to the stochastic differential equation for matching the volatility between real-world financial data and Monte Carlo simulations. Finally, a two-delay combination method is proposed to increase the prediction accuracy of the individual deep learning model
    corecore