873 research outputs found
Crop Yield Prediction Using Deep Neural Networks
Crop yield is a highly complex trait determined by multiple factors such as
genotype, environment, and their interactions. Accurate yield prediction
requires fundamental understanding of the functional relationship between yield
and these interactive factors, and to reveal such relationship requires both
comprehensive datasets and powerful algorithms. In the 2018 Syngenta Crop
Challenge, Syngenta released several large datasets that recorded the genotype
and yield performances of 2,267 maize hybrids planted in 2,247 locations
between 2008 and 2016 and asked participants to predict the yield performance
in 2017. As one of the winning teams, we designed a deep neural network (DNN)
approach that took advantage of state-of-the-art modeling and solution
techniques. Our model was found to have a superior prediction accuracy, with a
root-mean-square-error (RMSE) being 12% of the average yield and 50% of the
standard deviation for the validation dataset using predicted weather data.
With perfect weather data, the RMSE would be reduced to 11% of the average
yield and 46% of the standard deviation. We also performed feature selection
based on the trained DNN model, which successfully decreased the dimension of
the input space without significant drop in the prediction accuracy. Our
computational results suggested that this model significantly outperformed
other popular methods such as Lasso, shallow neural networks (SNN), and
regression tree (RT). The results also revealed that environmental factors had
a greater effect on the crop yield than genotype.Comment: 9 pages, Presented at 2018 INFORMS Conference on Business Analytics
and Operations Research (Baltimore, MD, USA). One of the winning solutions to
the 2018 Syngenta Crop Challeng
Stock Market Prediction via Deep Learning Techniques: A Survey
The stock market prediction has been a traditional yet complex problem
researched within diverse research areas and application domains due to its
non-linear, highly volatile and complex nature. Existing surveys on stock
market prediction often focus on traditional machine learning methods instead
of deep learning methods. Deep learning has dominated many domains, gained much
success and popularity in recent years in stock market prediction. This
motivates us to provide a structured and comprehensive overview of the research
on stock market prediction focusing on deep learning techniques. We present
four elaborated subtasks of stock market prediction and propose a novel
taxonomy to summarize the state-of-the-art models based on deep neural networks
from 2011 to 2022. In addition, we also provide detailed statistics on the
datasets and evaluation metrics commonly used in the stock market. Finally, we
highlight some open issues and point out several future directions by sharing
some new perspectives on stock market prediction
Representation learning in finance
Finance studies often employ heterogeneous datasets from different sources with different structures and frequencies. Some data are noisy, sparse, and unbalanced with missing values; some are unstructured, containing text or networks. Traditional techniques often struggle to combine and effectively extract information from these datasets. This work explores representation learning as a proven machine learning technique in learning informative embedding from complex, noisy, and dynamic financial data. This dissertation proposes novel factorization algorithms and network modeling techniques to learn the local and global representation of data in two specific financial applications: analystsâ earnings forecasts and asset pricing.
Financial analystsâ earnings forecast is one of the most critical inputs for security valuation and investment decisions. However, it is challenging to fully utilize this type of data due to the missing values. This work proposes one matrix-based algorithm, âCoupled Matrix Factorization,â and one tensor-based algorithm, âNonlinear Tensor Coupling and Completion Framework,â to impute missing values in analystsâ earnings forecasts and then use the imputed data to predict firmsâ future earnings. Experimental analysis shows that missing value imputation and representation learning by coupled matrix/tensor factorization from the observed entries improve the accuracy of firm earnings prediction. The results confirm that representing financial time-series in their natural third-order tensor form improves the latent representation of the data. It learns high-quality embedding by overcoming information loss of flattening data in spatial or temporal dimensions.
Traditional asset pricing models focus on linear relationships among asset pricing factors and often ignore nonlinear interaction among firms and factors. This dissertation formulates novel methods to identify nonlinear asset pricing factors and develops asset pricing models that capture global and local properties of data. First, this work proposes an artificial neural network âauto enco derâ based model to capture the latent asset pricing factors from the global representation of an equity index. It also shows that autoencoder effectively identifies communal and non-communal assets in an index to facilitate portfolio optimization. Second, the global representation is augmented by propagating information from local communities, where the network determines the strength of this information propagation. Based on the Laplacian spectrum of the equity market network, a network factor âZ-scoreâ is proposed to facilitate pertinent information propagation and capture dynamic changes in network structures. Finally, a âDynamic Graph Learning Framework for Asset Pricingâ is proposed to combine both global and local representations of data into one end-to-end asset pricing model. Using graph attention mechanism and information diffusion function, the proposed model learns new connections for implicit networks and refines connections of explicit networks. Experimental analysis shows that the proposed model incorporates information from negative and positive connections, captures the network evolution of the equity market over time, and outperforms other state-of-the-art asset pricing and predictive machine learning models in stock return prediction.
In a broader context, this is a pioneering work in FinTech, particularly in understanding complex financial market structures and developing explainable artificial intelligence models for finance applications. This work effectively demonstrates the application of machine learning to model financial networks, capture nonlinear interactions on data, and provide investors with powerful data-driven techniques for informed decision-making
Crowd Counting in Low-Resolution Crowded Scenes Using Region-Based Deep Convolutional Neural Networks
© 2013 IEEE. Crowd counting and density estimation is an important and challenging problem in the visual analysis of the crowd. Most of the existing approaches use regression on density maps for the crowd count from a single image. However, these methods cannot localize individual pedestrian and therefore cannot estimate the actual distribution of pedestrians in the environment. On the other hand, detection-based methods detect and localize pedestrians in the scene, but the performance of these methods degrades when applied in high-density situations. To overcome the limitations of pedestrian detectors, we proposed a motion-guided filter (MGF) that exploits spatial and temporal information between consecutive frames of the video to recover missed detections. Our framework is based on the deep convolution neural network (DCNN) for crowd counting in the low-to-medium density videos. We employ various state-of-the-art network architectures, namely, Visual Geometry Group (VGG16), Zeiler and Fergus (ZF), and VGGM in the framework of a region-based DCNN for detecting pedestrians. After pedestrian detection, the proposed motion guided filter is employed. We evaluate the performance of our approach on three publicly available datasets. The experimental results demonstrate the effectiveness of our approach, which significantly improves the performance of the state-of-the-art detectors
Artificial Intelligence and Machine Learning Approaches to Energy Demand-Side Response: A Systematic Review
Recent years have seen an increasing interest in Demand Response (DR) as a means to provide flexibility, and hence improve the reliability of energy systems in a cost-effective way. Yet, the high complexity of the tasks associated with DR, combined with their use of large-scale data and the frequent need for near real-time de-cisions, means that Artificial Intelligence (AI) and Machine Learning (ML) â a branch of AI â have recently emerged as key technologies for enabling demand-side response. AI methods can be used to tackle various challenges, ranging from selecting the optimal set of consumers to respond, learning their attributes and pref-erences, dynamic pricing, scheduling and control of devices, learning how to incentivise participants in the DR schemes and how to reward them in a fair and economically efficient way. This work provides an overview of AI methods utilised for DR applications, based on a systematic review of over 160 papers, 40 companies and commercial initiatives, and 21 large-scale projects. The papers are classified with regards to both the AI/ML algorithm(s) used and the application area in energy DR. Next, commercial initiatives are presented (including both start-ups and established companies) and large-scale innovation projects, where AI methods have been used for energy DR. The paper concludes with a discussion of advantages and potential limitations of reviewed AI techniques for different DR tasks, and outlines directions for future research in this fast-growing area
Ensemble Reinforcement Learning: A Survey
Reinforcement Learning (RL) has emerged as a highly effective technique for
addressing various scientific and applied problems. Despite its success,
certain complex tasks remain challenging to be addressed solely with a single
model and algorithm. In response, ensemble reinforcement learning (ERL), a
promising approach that combines the benefits of both RL and ensemble learning
(EL), has gained widespread popularity. ERL leverages multiple models or
training algorithms to comprehensively explore the problem space and possesses
strong generalization capabilities. In this study, we present a comprehensive
survey on ERL to provide readers with an overview of recent advances and
challenges in the field. First, we introduce the background and motivation for
ERL. Second, we analyze in detail the strategies that have been successfully
applied in ERL, including model averaging, model selection, and model
combination. Subsequently, we summarize the datasets and analyze algorithms
used in relevant studies. Finally, we outline several open questions and
discuss future research directions of ERL. By providing a guide for future
scientific research and engineering applications, this survey contributes to
the advancement of ERL.Comment: 42 page
Traffic Path Recommendation Model based on a Weighted Sum of Extracted Parameter
A path recommendation for vehicular traffic is important task of traffic analysis. It is a challenging problem for researchers to extract all paths and recommend the shortest path between Origin and Destination (OD) pairs. This paper comes up with a model which is established on the weighted sum of selected link references to recommend a path for OD pairs. First, to maintain spatial dependence between link references, a vehicular traffic network of roads is proposed as a rectangular coordinate system. The algorithm based on K-means and smoothing is introduced to select link references across OD pairs. A distance aggregation algorithm is proposed to evaluate all possible paths across an OD pair. Finally, out of overwhelming paths, the algorithm recommends the shortest distance path across an OD pair. Our proposed model effectively selects the link references and gets an overall shortest path recommendation. The proposed model analyzes the non-Euclidean distance of selected link references. Our experimental analysis shows that on an average, the first four link predictions lead to 77.37% distance coverage for the recommended path
- âŠ