2,514 research outputs found
Research reports: 1990 NASA/ASEE Summer Faculty Fellowship Program
Reports on the research projects performed under the NASA/ASEE Summer Faculty Fellowship Program are presented. The program was conducted by The University of Alabama and MSFC during the period from June 4, 1990 through August 10, 1990. Some of the topics covered include: (1) Space Shuttles; (2) Space Station Freedom; (3) information systems; (4) materials and processes; (4) Space Shuttle main engine; (5) aerospace sciences; (6) mathematical models; (7) mission operations; (8) systems analysis and integration; (9) systems control; (10) structures and dynamics; (11) aerospace safety; and (12) remote sensin
Wildfire Prediction to Inform Fire Management: Statistical Science Challenges
Wildfire is an important system process of the earth that occurs across a
wide range of spatial and temporal scales. A variety of methods have been used
to predict wildfire phenomena during the past century to better our
understanding of fire processes and to inform fire and land management
decision-making. Statistical methods have an important role in wildfire
prediction due to the inherent stochastic nature of fire phenomena at all
scales. Predictive models have exploited several sources of data describing
fire phenomena. Experimental data are scarce; observational data are dominated
by statistics compiled by government fire management agencies, primarily for
administrative purposes and increasingly from remote sensing observations.
Fires are rare events at many scales. The data describing fire phenomena can be
zero-heavy and nonstationary over both space and time. Users of fire modeling
methodologies are mainly fire management agencies often working under great
time constraints, thus, complex models have to be efficiently estimated. We
focus on providing an understanding of some of the information needed for fire
management decision-making and of the challenges involved in predicting fire
occurrence, growth and frequency at regional, national and global scales.Comment: Published in at http://dx.doi.org/10.1214/13-STS451 the Statistical
Science (http://www.imstat.org/sts/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Deep Learning Techniques in Extreme Weather Events: A Review
Extreme weather events pose significant challenges, thereby demanding
techniques for accurate analysis and precise forecasting to mitigate its
impact. In recent years, deep learning techniques have emerged as a promising
approach for weather forecasting and understanding the dynamics of extreme
weather events. This review aims to provide a comprehensive overview of the
state-of-the-art deep learning in the field. We explore the utilization of deep
learning architectures, across various aspects of weather prediction such as
thunderstorm, lightning, precipitation, drought, heatwave, cold waves and
tropical cyclones. We highlight the potential of deep learning, such as its
ability to capture complex patterns and non-linear relationships. Additionally,
we discuss the limitations of current approaches and highlight future
directions for advancements in the field of meteorology. The insights gained
from this systematic review are crucial for the scientific community to make
informed decisions and mitigate the impacts of extreme weather events
A deep learning method for convective weather forecasting: CNN-BiLSTM-AM (version 1.0)
This work developed a CNN-BiLSTM-AM model for convective weather forecasting using deep learning algorithms based on reanalysis and forecast data from the NCEP GFS, the performance of the model was evaluated. The results show that: (1) Compared to traditional machine learning algorithms, the CNN-BiLSTM-AM model has the ability to automatically learn deeper nonlinear features of convective weather. As a result, it exhibits higher forecasting accuracy on the convective weather dataset. Furthermore, as the forecast lead time increases, the information value provided by this model also changes. (2) In comparison to subjective forecasts by forecasters, the objective forecasting approach of the CNN-BiLSTM-AM model demonstrates advantages in metrics such as Probability of Detection (POD), False Alarm Rate (FAR), Threat Score (TS), and Missing Alarm Rate (MAR). Specifically, the average TS score for heavy precipitation reaches 0.336, which is a 33.2 % improvement compared to the forecaster's score of 0.252. Moreover, due to the CNN-BiLSTM-AM model's ability to automatically extract classification features based on a large sample dataset and consider a comprehensive range of convective parameters, it effectively reduces the FAR. (3) The interpretability study of the machine learning-based convective weather mechanism reveals that the importance ranking of convective weather forecasting factors arranged by machine learning methods largely aligns with the subjective understanding of forecasters. For example, the total precipitable water (PWAT) is identified as a critical factor for short-term heavy precipitation forecasting, regional factors have significant impacts on convective weather, and vertical motion at 300 hPa provides dynamic lifting conditions for convection. This objective analysis of factor ranking not only further confirms the effectiveness of machine learning in automatically extracting convective weather features but also validates the rationality of the sample set construction. Overall, the use of the CNN-BiLSTM-AM model in convective weather forecasting demonstrates superior performance compared to traditional machine learning algorithms and subjective forecasting methods
Icing thickness prediction model for overhead transmission lines
Failures in a large electric power system are often inevitable. Severe weather conditions are one of the main causes of transmission line failures. Using the fault data of transmission lines of Shaanxi Power Grid from 2006 to 2016, in conjunction with meteorological information, this paper analyses the relationship between the temporal-spatial distribution characteristics of meteorological disasters and the fault of transmission lines in Shaanxi Province, China.
In order to analyze the influence of micro-meteorology on ice coating, a grey correlation analysis method is proposed. This thesis calculates the grey relational between ice thickness and micro-meteorological parameters such as ambient temperature, relative humidity, wind speed and precipitation. The results show that the correlation between ambient temperature, wind speed and ice thickness is bigger than others. Based on the results of grey correlation analysis, a Multivariate Grey Model (MGM) and a Back Propagation (BP) neural network prediction model are built based on ice thickness, ambient temperature and wind speed. The prediction results of these two models are verified by the case of ice-coating of Shaanxi power grid. The results show that the prediction errors of the two models are small and satisfy the engineering requirement. Then a realistic case is carried out by using these two models. An icing risk map is drawn according to the results
Internal Fault Diagnosis of MMC-HVDC Based on Classification Algorithms in Machine Learning
With the development of the HVDC system, MMC-HVDC is now the most advanced technology that has been put into use. In power systems, faults happen during the operation due to natural reasons or devices physical issues, which would cause serious economic losses and other implications. Thus, fault detection and analysis are extremely important, especially in the HVDC system. Existing works in literature mainly focus on the faults detection and analysis on the system side such as short circuit of the AC side, and open circuit of the DC side. However, little attention has been paid to the fault detection and analysis inside the converters. With the technology development of converter devices, replacing the whole converter becomes more expensive. Thus, my research mainly focuses on the detection and classification of the faults within the internal of the MMC module.
In this research, an SPS model of MMC-HVDC is built as the example. Faults including short circuit and open circuit located inside the MMC module are simulated. Machine learning algorithms are chosen as the tool to achieve the goal of detecting faults and locating the fault position inside the MMC module precisely. After comparing the basic characteristics and properly application situations of various methods of machine learning, Coarse KNN, Complex Tree and Bagged Tree (Random Forest) are deployed to solve the problem. The performance of the methods are analyzed and compared, to get the most proper method in solving the problem
CROM: Continuous Reduced-Order Modeling of PDEs Using Implicit Neural Representations
The long runtime of high-fidelity partial differential equation (PDE) solvers
makes them unsuitable for time-critical applications. We propose to accelerate
PDE solvers using reduced-order modeling (ROM). Whereas prior ROM approaches
reduce the dimensionality of discretized vector fields, our continuous
reduced-order modeling (CROM) approach builds a smooth, low-dimensional
manifold of the continuous vector fields themselves, not their discretization.
We represent this reduced manifold using continuously differentiable neural
fields, which may train on any and all available numerical solutions of the
continuous system, even when they are obtained using diverse methods or
discretizations. We validate our approach on an extensive range of PDEs with
training data from voxel grids, meshes, and point clouds. Compared to prior
discretization-dependent ROM methods, such as linear subspace proper orthogonal
decomposition (POD) and nonlinear manifold neural-network-based autoencoders,
CROM features higher accuracy, lower memory consumption, dynamically adaptive
resolutions, and applicability to any discretization. For equal latent space
dimension, CROM exhibits 79 and 49 better accuracy, and
39 and 132 smaller memory footprint, than POD and autoencoder
methods, respectively. Experiments demonstrate 109 and 89
wall-clock speedups over unreduced models on CPUs and GPUs, respectively
A Helmholtz equation solver using unsupervised learning: Application to transcranial ultrasound
Transcranial ultrasound therapy is increasingly used for the non-invasive
treatment of brain disorders. However, conventional numerical wave solvers are
currently too computationally expensive to be used online during treatments to
predict the acoustic field passing through the skull (e.g., to account for
subject-specific dose and targeting variations). As a step towards real-time
predictions, in the current work, a fast iterative solver for the heterogeneous
Helmholtz equation in 2D is developed using a fully-learned optimizer. The
lightweight network architecture is based on a modified UNet that includes a
learned hidden state. The network is trained using a physics-based loss
function and a set of idealized sound speed distributions with fully
unsupervised training (no knowledge of the true solution is required). The
learned optimizer shows excellent performance on the test set, and is capable
of generalization well outside the training examples, including to much larger
computational domains, and more complex source and sound speed distributions,
for example, those derived from x-ray computed tomography images of the skull.Comment: 23 pages, 13 figure
Digital Signal Processing for Optical Communications and Coherent LiDAR
Internet data traffic within data centre, access and metro networks is experiencing
unprecedented growth driven by many data-intensive applications. Significant
efforts have been devoted to the design and implementation of low-complexity
digital signal processing (DSP) algorithms that are suitable for these short-reach
optical links. In this thesis, a novel low-complexity frequency-domain (FD)
multiple-input multiple-output (MIMO) equaliser with momentum-based gradient
descent algorithm is proposed, capable of mitigating both static and dynamic
impairments arising from the optical fibre. The proposed frequency-domain
equaliser (FDE) also improves the robustness of the adaptive equaliser against
feedback latencies which is the main disadvantage of FD adaptive equalisers under
rapid channel variations.
The development and maturity of optical fibre communication techniques over
the past few decades have also been beneficial to many other fields, especially
coherent light detection and ranging (LiDAR) techniques. Many applications
of coherent LiDAR are also cost-sensitive, e.g., autonomous vehicles (AVs).
Therefore, in this thesis, a low-cost and low-complexity single-photodiode-based
coherent LiDAR system is investigated. The receiver sensitivity performance of this
receiver architecture is assessed through both simulations and experiments, using
two ranging waveforms known as double-sideband (DSB) amplitude-modulated
chirp signal and single-sideband (SSB) frequency-modulated continuous-wave
(FMCW) signals. Besides, the impact of laser phase noise on the ranging precision
when operating within and beyond the laser coherence length is studied. Achievable
ranging precision beyond the laser coherence length is quantified
Lightning Prediction for Space Launch Using Machine Learning Based Off of Electric Field Mills and Lightning Detection and Ranging Data
Kennedy Space Center and Cape Canaveral Air Station, FL, where the Air Force conducts space launches, are in an area of frequent lightning strikes, which is main obstacle in meeting launch goals. The 45th Weather Squadron (45th WS) ensures that any weather safety requirements are met during pre-launch and actual space launch. Using only summer months from three years’ worth of lightning detection and ranging (LDAR) and electric field mill (EFM) data from KSC, several feedforward neural networks are constructed. Separate models are built for each EFM and trained by adjusting parameters to forecast lightning 30 minutes out in the surrounding area of each field mill
- …