147 research outputs found

    Local polynomial method for ensemble forecast of time series

    Get PDF
    We present a nonparametric approach based on local polynomial regression for ensemble forecast of time series. The state space is first reconstructed by embedding the univariate time series of the response variable in a space of dimension (<i>D</i>) with a delay time (&tau;). To obtain a forecast from a given time point <i>t</i>, three steps are involved: (i) the current state of the system is mapped on to the state space, known as the feature vector, (ii) a small number (<i>K</i>=&alpha;*<i>n</i>, &alpha;=fraction (0,1] of the data, <i>n</i>=data length) of neighbors (and their future evolution) to the feature vector are identified in the state space, and (iii) a polynomial of order <i>p</i> is fitted to the identified neighbors, which is then used for prediction. A suite of parameter combinations (<i>D</i>, &tau;, &alpha;, <i>p</i>) is selected based on an objective criterion, called the Generalized Cross Validation (GCV). All of the selected parameter combinations are then used to issue a T-step iterated forecast starting from the current time <i>t</i>, thus generating an ensemble forecast which can be used to obtain the forecast probability density function (PDF). The ensemble approach improves upon the traditional method of providing a single mean forecast by providing the forecast uncertainty. Further, for short noisy data it can provide better forecasts. We demonstrate the utility of this approach on two synthetic (Henon and Lorenz attractors) and two real data sets (Great Salt Lake bi-weekly volume and NINO3 index). This framework can also be used to forecast a vector of response variables based on a vector of predictors

    Local polynomial method for ensemble forecast of time series

    No full text
    International audienceWe present a nonparametric approach based on local polynomial regression for ensemble forecast of time series. The state space is first reconstructed by embedding the univariate time series of the response variable in a space of dimension (D) with a delay time (?). To obtain a forecast from a given time point t, three steps are involved: (i) the current state of the system is mapped on to the state space, known as the feature vector, (ii) a small number (K=?*n, ?=fraction (0,1] of the data, n=data length) of neighbors (and their future evolution) to the feature vector are identified in the state space, and (iii) a polynomial of order p is fitted to the identified neighbors, which is then used for prediction. A suite of parameter combinations (D, ?, ?, p) is selected based on an objective criterion, called the Generalized Cross Validation (GCV). All of the selected parameter combinations are then used to issue a T-step iterated forecast starting from the current time t, thus generating an ensemble forecast which can be used to obtain the forecast probability density function (PDF). The ensemble approach improves upon the traditional method of providing a single mean forecast by providing the forecast uncertainty. Further, for short noisy data it can provide better forecasts. We demonstrate the utility of this approach on two synthetic (Henon and Lorenz attractors) and two real data sets (Great Salt Lake bi-weekly volume and NINO3 index). This framework can also be used to forecast a vector of response variables based on a vector of predictors

    Two-Dimensional Flood Inundation Modeling in the Godavari River Basin, India—Insights on Model Output Uncertainty

    Get PDF
    Most flood inundation models do not come with an uncertainty analysis component chiefly because of the complexity associated with model calibration. Additionally, the fact that the models are both data-and compute-intensive, and since uncertainty results from multiple sources, adds another layer of complexity for model use. In the present study, flood inundation modeling was performed in the Godavari River Basin using the Hydrologic Engineering Center—River Analysis System 2D (HEC-RAS 2D) model. The model simulations were generated for six different scenarios that resulted from combinations of different geometric, hydraulic and hydrologic conditions. Thus, the resulted simulations account for multiple sources of uncertainty. The SRTM-30 m and MERIT90 m Digital elevation Model (DEM), two sets of Manning’s roughness coefficient (Manning’s n) and observed and estimated boundary conditions, were used to reflect geometric, hydraulic and hydrologic uncertainties, respectively. The HEC-RAS 2D model ran in an unsteady state mode for the abovementioned six scenarios for the selected three flood events that were observed in three different years, i.e., 1986, 2005 and 2015. The water surface elevation (H) was compared in all scenarios as well as with the observed values at selected locations. In addition, ‘H’ values were analyzed for two different structures of the computational model. The average correlation coefficient (r) between the observed and simulated H values is greater than 0.85, and the highest r, i.e., 0.95, was observed for the combination of MERIT-90 m DEM and optimized (obtained via trial and error) Manning’s n. The analysis shows uncertainty in the river geometry information, and the results highlight the varying role of geometric, hydraulic and hydrologic conditions in the water surface elevation estimates. In addition to the role of the abovementioned, the study recommends a systematic model calibration and river junction modeling to understand the hydrodynamics upstream and downstream of the junction. © 2021 by the authors. Licensee MDPI, Basel, Switzerland

    Water and Food Nexus: Role of Socio-Economic Status on Water–Food Nexus in an Urban Agglomeration Hyderabad, India Using Consumption Water Footprint

    Get PDF
    Cities are complex and evolving systems with various factors playing key roles, e.g., population increase, the migration of population, the availability of resources, and the flexibility of policies. Consumers’ socioeconomic status is also an important aspect that needs to be studied in the context of a self-reliant urban city in its resource consumption. In this regard, the association between water-food and socio-economic attributes was analyzed based on the consumer-centric approach for the Hyderabad Metro Development Authority (HMDA) region, India. In this study, the embedded water content in food consumption was estimated and analyzed for nine food groups and twelve economic classes of the HMDA region. The middle economic classes were found to correspond to ~80% of embedded water content in the HMDA region, followed by the upper and lower economic classes. Except for cereals, per capita, the water consumption of all food groups increased with the spending power of the economic class. The green, blue, and grey consumption water footprints (WFs) suggested that much of the water that is being consumed in the HMDA region is precipitation- driven, followed by surface and groundwater resources. Limited water resources, water resource variability, climate change consequences including future climate projections, uncertainty in data, WF estimates, and region’s future growth imply a detailed study in drafting policies to become a self-reliant region. © 2021 by the authors. Licensee MDPI, Basel, Switzerland

    Multi-Spatial Resolution Rainfall-Runoff Modelling—A Case Study of Sabari River Basin, India

    Get PDF
    One of the challenges in rainfall-runoff modeling is the identification of an appropriate model spatial resolution that allows streamflow estimation at customized locations of the river basin. In lumped modeling, spatial resolution is not an issue as spatial variability is not accounted for, whereas in distributed modeling grid or cell resolution can be related to spatial resolution but its application is limited because of its large data requirements. Streamflow estimation at the data-poor customized locations is not possible in lumped modeling, whereas it is challenging in distributed modeling. In this context, semi-distributed modeling offers a solution including model resolution and estimation of streamflow at customized locations of a river basins with less data requirements. In this study, the Hydrologic Engineering Center-Hydrologic Modeling System (HEC-HMS) model is employed in semi-distribution mode on river basins of six different spatial resolutions. The model was calibrated and validated for fifteen and three selected flood events, respectively, of three types, i.e., single peak (SP), double peak (DP)-and multiple peaks (MP) at six different spatial resolution of the Sabari River Basin (SRB), a sub-basin of the Godavari basin, India. Calibrated parameters were analyzed to understand hydrologic parameter variability in the context of spatial resolution and flood event aspects. Streamflow hydrographs were developed, and various verification metrics and model scores were calculated for reference-and calibration-scenarios. During the calibration phase, the median of correlation coefficient and NSE for all 15 events of all six configurations was 0.90 and 0.69, respectively. The estimated streamflow hydrographs from six configurations suggest the model’s ability to simulate the processes efficiently. Parameters obtained from the calibration phase were used to generate an ensemble of streamflow at multiple locations including basin outlet as part of the validation. The estimated ensemble of streamflows appeared to be realistic, and both single-valued and ensemble verification metrics indicated the model’s good performance. The results suggested better performance of lumped modeling followed by the semi-distributed modeling with a finer spatial resolution. Thus, the study demonstrates a method that can be applied for real-time streamflow forecast at interior locations of a basin, which are not necessarily data rich. © 2021 by the authors. Licensee MDPI, Basel, Switzerland

    Investigating the Accuracies in Short-Term Weather Forecasts and Its Impact on Irrigation Practices

    Get PDF
    The effectiveness of irrigation schedules and crop growth models is largely dependent on the accuracies associated with the most uncertain input parameter, viz. weather forecasts. This research is focused on quantifying the inaccuracies associated with the India Meteorological Department (IMD) issuing short-term weather forecasts (with 1-5 days lead time), and their propagation into irrigation models, with an objective to select the optimal parameters for use with simulation. While precipitation (P) forecasts were directly used, remaining meteorological forecasts were converted to reference evapotranspiration (ET0) forecasts. The effectiveness of the 'P' and 'ET0' forecasts was found to be low at all lead times. We applied two popular bias correction methods: linear scaling (LS) and empirical quantile mapping (EQM), and observed a marginal improvement in forecast skill. Bias-corrected, forecast-driven irrigation scenarios, along with conventional irrigation (that ignores weather forecast), and a hypothetical perfect 5-day forecast-based irrigation (as reference) system were tested on a water-intensive paddy crop for two growing seasons (S1: monsoon and S2: winter). Conventional irrigation resulted in the highest use of irrigation water (820 mm in S1, 880 mm in S2) and percolation loss (1,140 mm in S1, 680 mm in S2), while achieving a low relative yield (0.88 in S1, 0.87 in S2). LS-corrected forecasts outperformed other scenarios with 20.24%±4.21% and 1.25%±1.51% savings in irrigation costs for S1 and S2, respectively. While IMD forecasts greatly improved irrigation schedules in the monsoon season, their usage for winter crops was found to be trivial. Our findings conclude that LS-corrected IMD forecasts were moderately reliable over multiple lead times and can serve as a valuable addition to irrigation scheduling, provided the contribution of 'P' to the total water balance is significant

    Attack Prevention in IoT through Hybrid Optimization Mechanism and Deep Learning Framework

    Get PDF
    The Internet of Things (IoT) connects schemes, programs, data management, and operations, and as they continuously assist in the corporation, they may be a fresh entryway for cyber-attacks. Presently, illegal downloading and virus attacks pose significant threats to IoT security. These risks may acquire confidential material, causing reputational and financial harm. In this paper hybrid optimization mechanism and deep learning,a frame is used to detect the attack prevention in IoT. To develop a cybersecurity warning system in a huge data set, the cybersecurity warning systems index system is first constructed, then the index factors are picked and measured, and finally, the situation evaluation is done.Numerous bio-inspired techniques were used to enhance the productivity of an IDS by lowering the data dimensionality and deleting unnecessary and noisy input. The Grey Wolf Optimization algorithm (GWO) is a developed bio-inspired algorithm that improves the efficacy of the IDS in detecting both regular and abnormal congestion in the network. The smart initialization step integrates the different pre-processing strategies to make sure that informative features are incorporated in the early development stages, has been improved. Researchers pick multi-source material in a big data environment for the identification and verification of index components and present a parallel reduction approach based on the classification significance matrix to decrease data underlying data characteristics. For the simulation of this situation, grey wolf optimization and whale optimization were combined to detect the attack prevention and the deep learning approach was presented. Utilizing system software plagiarism, the TensorFlow deep neural network is intended to classify stolen software. To reduce the noise from the signal and to zoom the significance of each word in the perspective of open-source plagiarism, the tokenization and weighting feature approaches are utilized. Malware specimens have been collected from the Mailing database for testing purposes. The experimental findings show that the suggested technique for measuring cyber security hazards in IoT has superior classification results to existing methods. Hence to detect the attack prevention in IoT process Whale with Grey wolf optimization (WGWO) and deep convolution network is used

    Analysis of Cyber Security In E-Governance Utilizing Blockchain Performance

    Get PDF
    E-Government refers to the administration of Information and Communication Technologies (ICT) to the procedures and functions of the government with the objective of enhancing the transparency, efficiency and participation of the citizens. E-Government is tough systems that require distribution, protection of privacy and security and collapse of these could result in social and economic costs on a large scale. Many of the available e-government systems like electronic identity system of management (eIDs), websites are established at duplicated databases and servers. An established validation and management system could face a single failure point and the system is prone to Distributed Denial of Service Attacks (DDoS), denial of service attacks (DoS), malware and other cyber attacks. The execution of a privacy preserving and a secure decentralized system is enabled by the block chain technology. Here any third-party organizations do not have any control over the transactions of the Government. With the help of block chain technology, new and existing data are encapsulated within ledger or blocks, which are evenly distributed through the network in an enduring and sustainable way. The privacy and security of information are improved with the help of block chain technology, where distribution and encryption of data are performed through the total network. This analytical paper maps out the analysis of the security in the e-government system, utilizing the block chain technology that provides privacy and security of information and thereby enhancing the trust among the public sector. Qualitative and theoretical analysis is made for the proposed topic and implications of privacy and security of the proposed system is made

    An efficient approach for estimating streamflow forecast skill elasticity

    Get PDF
    Seasonal streamflow prediction skill can derive from catchment initial hydrological conditions (IHCs) and from the future seasonal climate forecasts (SCFs) used to produce the hydrological forecasts. Although much effort has gone into producing state-of-the-art seasonal streamflow forecasts from improving IHCs and SCFs, these developments are expensive and time consuming and the forecasting skill is still limited in most parts of the world. Hence, sensitivity analyses are crucial to funnel the resources into useful modelling and forecasting developments. It is in this context that a sensitivity analysis technique, the variational ensemble streamflow prediction assessment (VESPA) approach, was recently introduced. VESPA can be used to quantify the expected improvements in seasonal streamflow forecast skill as a result of realistic improvements in its predictability sources (i.e., the IHCs and the SCFs) - termed ‘skill elasticity’ - and to indicate where efforts should be targeted. The VESPA approach is however computationally expensive, relying on multiple hindcasts having varying levels of skill in IHCs and SCFs. This paper presents two approximations of the approach that are computationally inexpensive alternatives. These new methods were tested against the original VESPA results using 30 years of ensemble hindcasts for 18 catchments of the contiguous United States. The results suggest that one of the methods, End Point Blending, is an effective alternative for estimating the forecast skill elasticities yielded by the VESPA approach. The results also highlight the importance of the choice of verification score for a goal-oriented sensitivity analysis

    Using oceanic-atmospheric oscillations for long lead time streamflow forecasting

    Full text link
    We present a data-driven model, Support Vector Machine (SVM), for long lead time streamflow forecasting using oceanic-atmospheric oscillations. The SVM is based on statistical learning theory that uses a hypothesis space of linear functions based on Kernel approach and has been used to predict a quantity forward in time on the basis of training from past data. The strength of SVM lies in minimizing the empirical classification error and maximizing the geometric margin by solving inverse problem. The SVM model is applied to three gages, i.e., Cisco, Green River, and Lees Ferry in the Upper Colorado River Basin in the western United States. Annual oceanic-atmospheric indices, comprising Pacific Decadal Oscillation (PDO), North Atlantic Oscillation (NAO), Atlantic Multidecadal Oscillation (AMO), and El Nino–Southern Oscillations (ENSO) for a period of 1906–2001 are used to generate annual streamflow volumes with 3 years lead time. The SVM model is trained with 86 years of data (1906–1991) and tested with 10 years of data (1992–2001). On the basis of correlation coefficient, root means square error, and Nash Sutcliffe Efficiency Coefficient the model shows satisfactory results, and the predictions are in good agreement with measured streamflow volumes. Sensitivity analysis, performed to evaluate the effect of individual and coupled oscillations, reveals a strong signal for ENSO and NAO indices as compared to PDO and AMO indices for the long lead time streamflow forecast. Streamflow predictions from the SVM model are found to be better when compared with the predictions obtained from feedforward back propagation artificial neural network model and linear regression
    corecore