129 research outputs found

    Utilisation des réseaux de neurones et de la régularisation bayésienne en modélisation de la température de l’eau en rivière

    Get PDF
    Dans ce travail, nous avons élaboré un modèle de prédiction des variations de la température d’un cours d’eau en fonction de variables climatiques, telles que la température de l’air ambiant, le débit d’eau et la quantité de précipitation reçue par le cours d’eau. Les réseaux de neurones statiques ont été utilisés pour approximer la relation entre ces différentes variables avec une erreur moyenne de 0,7 °C. Par ailleurs, nous proposons un modèle de prédiction de l’évolution de la température de l’eau à court et moyen termes pour les jours (j + i, i = 1,2,..). Deux méthodes ont été appliquées : la première, de type itérative, utilise la valeur estimée du jour j pour prédire la valeur de la température de l’eau au jour j + 1; la seconde méthode, beaucoup plus simple à mettre en oeuvre, consiste à estimer la température de tous les jours considérés en une seule fois.L’optimisation de la fonction de coût par l’algorithme de Levenberg-Marquardt, disponible dans l’outil « réseaux de neurones » de MATLAB a permis d’améliorer nettement la performance des modèles. Des résultats très satisfaisants sont alors obtenus en testant la validité du modèle par la validation croisée avec des erreurs moyennes de prédiction à sept jours de 1,5 °C.Understanding and predicting water temperatures is essential in order to help prevent or forecast high temperature problems. To attain this objective, we define in this work a model that predicts temperature variations in a small stream according to climatic variables, such as air temperature, water flow and quantity of rainfall in the river catchment. Static neural networks were used as a technique for evaluation of the relations among the various variables, with a mean error of 0.7°C.In addition, we developed a forecasting model able to estimate the short-term and mid-term variations of water temperature, i.e., to forecast the temperature of days (j+i , i=1,2…) from climatic parameters of day j. Two methods were used: the first one is iterative and uses the estimated value of day j to estimate the value of the water temperature for day j+1. The second method is much simpler, involving an estimate of the temperature of all days at once. The Levenberg-Marquardt algorithm implemented in the Matlab neural network toolbox allowed a marked improvement in the performance of the model. Very satisfactory results were then obtained by testing the validity by cross validation technique with a mean error of 1.5°C for long term prediction of 7 days

    Utilisation de l'information historique en analyse hydrologique fréquentielle

    Get PDF
    L'utilisation de l'information historique dans une analyse fréquentielle permet de mieux mobiliser l'information réellement disponible et devrait donc permettre d'améliorer l'estimation des quantiles de grande période de retour. Par information historique, on entend ici de l'information relative à des grandes crues qui se sont produites avant le début de la période de mesure (dite période de jaugeage systématique) des niveaux et débits des lacs et rivières. On observe de manière générale que l'utilisation de l'information historique conduit à une diminution de l'impact des valeurs singulières dans les séries d'enregistrements systématiques et à une diminution de l'écart-type des estimations. Dans le présent article on présente les méthodes statistiques qui permettent la modélisation de l'information historique.Use of information about historical floods, i.e. extreme floods that occurred prior to systematic gauging, can often substantially improve the precision of flood quantile estimates. Such information can be retrieved from archives, newspapers, interviews with local residents, or by use of paleohydrologic and dendohydrologic traces. Various statistical techniques for incorporating historical information into frequency analyses are discussed in this review paper. The basic hypothesis in the statistical modeling of historical information is that a certain perception water level exists and that during a given historical period preceding the period of gauging, all exceedances of this level have been recorded, be it in newpapers, in people's memory, or trough traces in the catchment such as sediment deposits or traces on trees. No information is available on floods that did not exceed the perception threshold. It is further assumed that a period of systematic gauging is available. Figure 1 illustrates this situation. The U.S. Water Resources Council (1982) recommended the use of the method of adjusted moments for fitting the log Pearson type III distribution. A weighting factor is applied to the data below the threshold observed during the gauged period to account for the missing data below the threshold in the historical period. Several studies have pointed out that the method of adjusted moments is inefficient. Maximum likelihood estimators based on partially censored data have been shown to be much more efficient and to provide a practical framework for incorporating imprecise and categorical data. Unfortunately, for some of the most common 3-parameter distributions used in hydrology, the maximum likelihood method poses numerical problems. Recently, some authors have proposed use of the method of expected moments, a variant of the method of adjusted moments which gives less weight to observations below the threshold. According to preliminary studies, estimators based on expected moments are almost as efficient as maximum likelihood estimators, but have the advantage of avoiding the numerical problems related to the maximization of likelihood functions. Several studies have emphasized the potential gain in estimation accuracy with the use of historical information. Because historical floods by definition are large, their introduction in a flood frequency analysis can have a major impact on estimates of rare floods. This is particularly true when 3-parameter distributions are considered. Moreover, use of historical information is a means to increase the representativity of a outlier in the systematic data. For example, an extreme outlier will not get the same weight in the analysis if one can state with certainty that it is the largest flood in, say, 200 years, and not only the largest flood in, say, 20 years of systematic gauging.Historical data are generally imprecise, and their inaccuracy should be properly accounted for in the analysis. However, even with substantial uncertainty in the data, the use of historical information is a viable means to improve estimates of rare floods

    Revue bibliographique des méthodes de prévision des débits

    Get PDF
    Dans le domaine de la prévision des débits, une grande variété de méthodes sont disponibles: des modèles stochastiques et conceptuels mais aussi des approches plus novatrices telles que les réseaux de neurones artificiels, les modèles à base de règles floues, la méthode des k plus proches voisins, la régression floue et les splines de régression. Après avoir effectué une revue détaillée de ces méthodes et de leurs applications récentes, nous proposons une classification qui permet de mettre en lumière les différences mais aussi les ressemblances entre ces approches. Elles sont ensuite comparées pour les problèmes différents de la prévision à court, moyen et long terme. Les recommandations que nous effectuons varient aussi avec le niveau d'information a priori. Par exemple, lorsque l'on dispose de séries chronologiques stationnaires de longue durée, nous recommandons l'emploi de la méthode non paramétrique des k plus proches voisins pour les prévisions à court et moyen terme. Au contraire, pour la prévision à plus long terme à partir d'un nombre restreint d'observations, nous suggérons l'emploi d'un modèle conceptuel couplé à un modèle météorologique basé sur l'historique. Bien que l'emphase soit mise sur le problème de la prévision des débits, une grande partie de cette revue, principalement celle traitant des modèles empiriques, est aussi pertinente pour la prévision d'autres variables.A large number of models are available for streamflow forecasting. In this paper we classify and compare nine types of models for short, medium and long-term flow forecasting, according to six criteria: 1. validity of underlying hypotheses, 2. difficulties encountered when building and calibrating the model, 3. difficulties in computing the forecasts, 4. uncertainty modeling, 5. information required by each type of model, and 6. parameter updating. We first distinguish between empirical and conceptual models, the difference being that conceptual models correspond to simplified representations of the watershed, while empirical model only try to capture the structural relationships between inputs to the watershed and outputs, such as streamflow. Amongst empirical models, we distinguish between stochastic models, i.e. models based on the theory of probability, and non-stochastic models. Three types of stochastic models are presented: statistical regression models, Box-Jenkins models, and the nonparametric k-nearest neighbor method. Statistical linear regression is only applicable for long term forecasting (monthly flows, for example), since it requires independent and identically distributed observations. It is a simple method of forecasting, and its hypotheses can be validated a posteriori if sufficient data are available. Box-Jenkins models include linear autoregressive models (AR), linear moving average models (MA), linear autoregressive - moving average models (ARMA), periodic ARMA models (PARMA) and ARMA models with auxiliary inputs (ARMAX). They are more adapted for weekly or daily flow forecasting, since the yallow for the explicit modeling of time dependence. Efficient methods are available for designing the model and updating the parameters as more data become available. For both statistical linear regression and Box-Jenkins models, the inputs must be uncorrelated and linearly related to the output. Furthermore, the process must be stationary. When it is suspected that the inputs are correlated or have a nonlinear effect on the output, the k-nearest neighbor method may be considered. This data-based nonparametric approach simply consists in looking, among past observations of the process, for the k events which are most similar to the present situation. A forecast is then built from the flows which were observed for these k events. Obviously, this approach requires a large database and a stationary process. Furthermore, the time required to calibrate the model and compute the forecasts increases rapidly with the size of the database. A clear advantage of stochastic models is that forecast uncertainty may be quantified by constructing a confidence interval. Three types of non-stochastic empirical models are also discussed: artificial neural networks (ANN), fuzzy linear regression and multivariate adaptive regression splines (MARS). ANNs were originally designed as simple conceptual models of the brain. However, for forecasting purposes, these models can be thought of simply as a subset of non linear empirical models. In fact, the ANN model most commonly used in forecasting, a multi-layer feed-forward network, corresponds to a non linear autoregressive model (NAR). To capture the moving average components of a time series, it is necessary to use recurrent architectures. ANNs are difficult to design and calibrate, and the computation of forecasts is also complex. Fuzzy linear regression makes it possible to extract linear relationships from small data sets, with fewer hypotheses than statistical linear regression. It does not require the observations to be uncorrelated, nor does it ask for the error variance to be homogeneous. However, the model is very sensitive to outliers. Furthermore, a posteriori validation of the hypothesis of linearity is not possible for small data sets. MARS models are based on the hypothesis that time series are chaotic instead of stochastic. The main advantage of the method is its ability to model non-stationary processes. The approach is non-parametric, and therefore requires a large data set.Amongst conceptual models, we distinguish between physical models, hydraulic machines, and fuzzy rule-based systems. Most conceptual hydrologic models are hydraulic machines, in which the watershed is considered to behave like a network of reservoirs. Physical modeling of a watershed would imply using fundamental physical equations at a small scale, such as the law of conservation of mass. Given the complexity of a watershed, this can be done in practice only for water routing. Consequently, only short term flow forecasts can be obtained from a physical model, since the effects of precipitation, infiltration and evaporation must be negligible. Fuzzy rule-based systems make it possible to model the water cycle using fuzzy IF-THEN rules, such as IF it rains a lot in a short period of time, THEN there will be a large flow increase following the concentration time. Each fuzzy quantifier is modeled using a fuzzy number to take into account the uncertainty surrounding it. When sufficient data are available, the fuzzy quantifiers can be constructed from the data. In general, conceptual models require more effort to develop than empirical models. However, for exceptional events, conceptual models can often provide more realistic forecasts, since empirical models are not well suited for extrapolation.A fruitful approach is to combine conceptual and empirical models. One way of doing this, called extended streamflow prediction or ESP, is to combine a stochastic model for generating meteorological scenarios with a conceptual model of the watershed.Based on this review of flow forecasting models, we recommend for short term forecasting (hourly and daily flows) the use of the k-nearest neighbor method, Box-Jenkins models, water routing models or hydraulic machines. For medium term forecasting (weekly flows, for example), we recommend the k-nearest neighbor method and Box-Jenkins models, as well as fuzzy-rule based and ESP models. For long term forecasting (monthly flows), we recommend statistical and fuzzy regression, Box-Jenkins, MARS and ESP models. It is important to choose a type of model which is appropriate for the problem at hand and for which the information available is sufficient. Each type of model having its advantages, it can be more efficient to combine different approaches when forecasting streamflow

    Revue de processus ponctuels et synthèse de tests statistiques pour le choix d'un type de processus

    Get PDF
    Nous nous intéressons dans ce travail de recherche à la modélisation d'une série d'événements par la théorie des processus ponctuels temporels. Un processus ponctuel est défini comme étant un processus stochastique pour lequel chaque réalisation constitue une collection de points. Un grand nombre d'ouvrages traitent particulièrement de ces processus, cependant, il existe dans la littérature peu de travaux qui se préoccupent de l'analyse de séries d'événements. On identifie deux catégories de séries d'événements : une série d'un seul type d'événements et une série de plusieurs types d'événements.L'objectif de ce travail est de mettre en évidence les différents tests statistiques appliqués aux séries d'un seul ou de plusieurs types d'événements et de proposer une classification de ces tests. Nous présentons d'abord une revue de littérature des processus ponctuels temporels, accompagnée d'une classification de ces modèles. Par la suite, nous identifions les tests statistiques de séries d'un seul type d'événements et nous examinons leur applicabilité pour une série de deux ou de plusieurs types d'événements. Les tests statistiques identifiés sont répartis en quatre classes : analyse graphique, tests appliqués au processus de Poisson homogène et non homogène, tests appliqués au processus de renouvellement homogène et les tests de discrimination entre deux processus ponctuels. Ce travail est réalisé avec l'idée d'une application ultérieure dans le cadre de l'analyse du risque.Les résultats de cette recherche ont montré qu'il n'existe dans la littérature que des tests d'une série d'un seul type d'événements et ils sont, généralement, valables pour les processus ponctuels suivants : Poisson homogène et renouvellement homogène. L'application de ces tests aux séries de deux ou de plusieurs types d'événements est possible dans le cas où les événements sont définis par leurs nombres et leurs temps d'occurrence seulement, i.e. la durée de chaque événement n'est pas prise en considération.The design and management of hydraulic structures require a good knowledge of the characteristics of extreme hydrologic events such as floods and droughts, that may occur at the site of interest. Occurrences of such events may be modelled as temporal point processes. This modelling approach allows the derivation of various performance indices related to the design and operation of this infrastructure, as well as to the quantification and management of the associated risks. In this paper, we present statistical tests that may be applied for the modelling of a series of events by temporal point processes. A point process is defined as a stochastic process for which each realisation constitutes a series of points. Although a large body of literature dealt with temporal point processes, very few focused on the analysis of a series of events.In the present paper we identify two types of series of events: the first represents a series of only one type of event, and the second represents a series of several types of events. The main objective of this research is to comprehensively review the statistical tests applied to the series of one or several types of events and to propose a classification of these tests. This comprehensive review of statistical tests applied to point processes is carried out with the ultimate objective of applying these tests to real case studies within the framework of risk analysis. For example, an extended low-flow event constitutes a risk that may place a water resources system in a state of failure. Thus, it's important to identify and quantify this risk in order to ensure the optimal management of water resources. The modelling of the observed series of events by point processes can provide some statistical results, such as the distribution of number of events or the shape of the intensity function. These results are useful in a risk analysis framework, which includes two steps: risk evaluation and risk management. In the first part of the paper, a review and classification of the various temporal point processes are presented. These include the homogeneous and nonhomogeneous Poisson processes, the Negative Binomial process, the cluster point processes (such as the Neyman-Scott and the Bartlett-Lewis processes), the doubly stochastic Poisson processes, the self-exciting point processes, the homogeneous and nonhomogeneous renewal processes and the semi-markov processes. Also, we illustrate the various links and relationships that exist between these point processes. This classification is elaborated by considering the homogeneous Poisson process as the starting point. The simplicity and the wide use of this process in the statistical and hydrological literature justify this choice.In the second part of the paper, statistical tests of a series of one type of event are identified. A series of events may be characterised by the number of events, the occurrence times of the events or by the duration of each event. These characteristics are considered as random variables that must be represented by suitable statistical distributions. A series of events may also be characterised by the intensity function, which represents the instantaneous average rate of occurrence of an event. Clearly, the choice of the statistical distribution to model the number of events in a series or the intensity function depends on the nature of the observed data. For example, a stationary series of events may be represented by a constant intensity function. Thus, it is necessary to conduct an analysis of the observed series of the events, such as graphical analysis and statistical testing in order to select and validate the hypothesis underlying the point process model. The hypotheses that may be verified include trend analysis, homogeneity analysis, periodicity analysis, independence of intervals between events, and the adequacy of a given distribution for the number of events and for the time intervals separating events.In the third part, the applicability of the tests identified in the second part to the case of a series of two or more types of events is examined. In this part, our goal is to analyse the global point process (or the pooled output) obtained by the superposition of the p subsidiary point processes. The decomposition of the global process into p point processes necessitates an identification of each type of event, characterised generally by the number of occurrences and by the intervals between the successive events of the same type. We also examine the applicability of the statistical tests identified in the second part to the case where the global point process is characterised by the duration of each type of event. We investigate more specifically the case of two subsidiary point processes (p=2) where the two event types alternate in the time (an alternating point process). Finally, statistical tests identified in the second part are classified into four categories: tests based on graphical analysis; tests applied to the homogeneous and nonhomogeneous Poisson processes; tests applied to the homogeneous renewal process; and finally tests of discrimination between two specific processes. Theses tests of discrimination include the selection among the Poisson process and the renewal process, the Poisson process and the Binomial point process, and finally, the selection among these three point processes: Cox process, Neyman-Scott process and renewal process.The results of this research indicate that, in the past, mostly tests for a series of one type of event were presented in the literature. These tests are only valid for the following point processes: a homogenous Poisson process or a homogenous renewal process. The application of these tests to a series of two or several types of events is possible as long as these events are only described by their number and time of occurrence i.e. the duration of each event can not be taken into consideration. Otherwise, these tests are applicable to the alternating point process, which is characterised only by the number and the duration of the two types of events

    Chemical composition, antimicrobial, antioxidant and anticancer activities of essential oil from Ammodaucus leucotrichus Cosson & Durieu (Apiaceae) growing in South Algeria

    Get PDF
    ABSTRACT. The chemical composition of the essential oil obtained by hydrodistillation from aerial parts of A. leucotrichus Cosson & Durieu (Apiaceae) grown in the south of Algeria (El-Oued) was determined by GC-MS analysis. The oil was found to be rich in perilladehyde 64.66% and D-Limonene 26.99%. The biological activity of A. leucotrichus Cosson & Durieu essential oil has been investigated. The in vitro antimicrobial activity of the essential oil sample was tested on eight strains, one yeast and one fungi. The test showed interesting antimicrobial properties, especially on Salmonella enterica and E. coli, the antioxidant capacity of the oil was measured using the cyclic voltammetry, and the AAT value of A. leucotrichus essential oil was evaluated 47.84 mg α-TE/L. In addition, the antitumor activity showed that the oil of A. leucotrichus was very significant against the HCT116 colon cancer cell line.               KEY WORDS: Ammodaucus leucotrichus, Antioxidant activity, Anticancer activity, Cyclic voltammetry Bull. Chem. Soc. Ethiop. 2019, 33(3), 541-549. DOI: https://dx.doi.org/10.4314/bcse.v33i3.1

    Evaluation of a Depth-Based Multivariate k

    Get PDF
    A nonparametric simulation model (k-nearest neighbor resampling, KNNR) for water quality analysis involving geographic information is suggested to overcome the drawbacks of parametric models. Geographic information is, however, not appropriately handled in the KNNR nonparametric model. In the current study, we introduce a novel statistical notion, called a “depth function,” in the classical KNNR model to appropriately manipulate geographic information in simulating stormwater quality. An application is presented for a case study of the total suspended solids throughout the entire United States. The stormwater total suspended solids concentration data indicated that the proposed model significantly improves the simulation performance compared with the existing KNNR model

    Synthèse de modèles régionaux d'estimation de crue utilisée en France et au Québec

    Get PDF
    De nombreuses méthodes régionales ont été développées pour améliorer l'estimation de la distribution des débits de crues en des sites où l'on dispose de peu d'information ou même d'aucune information. Cet article présente une synthèse de modèles hydrologiques utilisés en France et au Québec (Canada), à l'occasion d'un séminaire relatif aux " méthodes d'estimation régionale en hydrologie " tenu à Lyon en mai 1997. Les modèles français sont fortement liés à une technique d'extrapolation de la distribution des crues, la méthode du Gradex, qui repose sur l'exploitation probabiliste conjointe des séries hydrométriques et pluviométriques. Ceci explique les deux principaux volets d'études régionales pratiquées en France : les travaux liés à la régionalisation des pluies et ceux liés à la régionalisation des débits. Les modèles québecois comprennent généralement deux étapes : la définition et la détermination de régions hydrologiquement homogènes, puis l'estimation régionale, par le transfert à l'intérieur d'une même région de l'information des sites jaugés à un site non-jaugé ou partiellement jaugé pour lequel on ne dispose pas d'information suffisante. Après avoir donné un aperçu des méthodes pratiquées dans les deux pays, une discussion dégage les caractéristiques principales et les complémentarités des différentes approches et met en évidence l'intérêt de développer une collaboration plus étroite pour mieux tenir compte des particularités et des complémentarités des méthodes développées de part et d'autre. Une des pistes évoquées consiste à combiner l'information régionale pluviométrique (approche française) et hydrométrique (approche québécoise).Design flood estimates at ungauged sites or at gauged sites with short records can be obtained through regionalization techniques. Various methods have been employed in different parts of the world for the regional analysis of extreme hydrological events. These regionalization approaches make different assumptions and hypotheses concerning the hydrological phenomena being modeled, rely on various types of continuous and non-continuous data, and often fall under completely different theories. A research seminar dealing with " regional estimation methods in hydrology " took place in Lyon during the month of May 1997, and brought together various researchers and practitioners mainly from France and the Province of Quebec (Canada). The present paper is based on the conferences and discussions that took place during this seminar and aims to review, classify, comparatively evaluate, and potentially propose improvements to the most prominent regionalization techniques utilized in France and Quebec. The specific objectives of this paper are :· to review the main regional hydrologic models that have been proposed and commonly used during the last three decades ;· to classify the literature into different groups according to the origin of the method, its specific objective, and the technique it adopts ; · to present a comprehensive evaluation of the characteristics of the methods, and to point out the hypotheses, data requirements, strengths and weaknesses of each particular one ; and · to investigate and identify potential improvements to the reviewed methods, by combining and extending the various approaches and integrating their particular strengths.Regionalization approaches adopted in France include the Gradex method which represents a simplified rainfall-runoff model which provides estimates of flood magnitudes of given probabilities and is based on rainfall data which often cover longer periods and are more reliable than flow data (Guillot and Duband, 1967 ; CFGB, 1994). It is based on the hypotheses that beyond a given rainfall threshold (known as the pivot point), all water is transformed into runoff, and that a rainfall event of a given duration generates runoff for the same length of time. These hypotheses are equivalent to assuming that, beyond the pivot point, the rainfall-runoff relationship is linear and that the precipitation and runoff probability curves are parallel on a Gumbel plot.In Quebec (and generally in North America), regional flood frequency analysis involves usually two steps : delineation of homogeneous regions, and regional estimation. In the first step, the focus is on identifying and regrouping sites which seem sufficiently homogeneous or sufficiently similar to the target ungauged site to provide a basis for information transfer. The second step of the analysis consists in inferring flood information (such as quantiles) at the target site using data from the stations identified in the first step of the analysis. Two types of " homogeneous " regions can be proposed : fixed set regions (geographically contiguous or non-contiguous) and neighborhood type of regions. The second type includes the methods of canonical correlation analysis and of the regions of influence. Regional estimation can be accomplished using one of two main approaches : index flood or quantile regression methods.The results of this work indicate that the philosophies of regionalization and the methods utilized in France and Quebec are complementary to each other and are based on different needs and outlooks. While the approaches followed in France are characterized by strong conceptual and geographic aspects with an emphasis on the utilization of information related to other phenomena (such as precipitations), the approaches adopted in Quebec rely on the strength of their statistical and stochastic components and usually condense the spatial and temporal information to a realistic functional form. This dissimilarity in the approaches being followed on either side may be originated by the distinct topographic and climatic characteristics of each region (France and Quebec) and by the differences in basin sizes and hydrometeorologic network densities. The conclusions of the seminar point to the large potential of improvements in regional estimation methods, which may result from an enhanced exchange between scientists from both sides : indeed, there is much to gain from learning about the dissimilarities between the various approaches, comparing their performances, and devising new methods that combine their individual strengths. Hence, the Gradex method for example could benefit from an increased utilization of regional flood information, while flood regionalization methods utilized in Quebec could gain much from the formalization of the use of rainfall information and from the integration of an improved modeling of physical hydrologic phenomena. This should result in the enhancement of the efficiency of regional estimation methods and their ability to handle various practical conditions.It is hoped that this research will contribute towards closing the gap between French and Quebec literature, and more generally between the European and the North American hydrological schools of thought, by narrowing the large literature that is available, by providing the necessary cross-evaluation of regional flood analysis models, and by providing comprehensive propositions for improved approaches for regional hydrologic modeling

    La régionalisation des précipitations : une revue bibliographique des développements récents

    Get PDF
    L'estimation de l'intensité de précipitations extrêmes est un sujet de recherche en pleine expansion. Nous présentons ici une synthèse des travaux de recherche sur l'analyse régionale des précipitations. Les principales étapes de l'analyse régionale revues sont les méthodes d'établissement de régions homogènes, la sélection de fonctions de distributions régionales et l'ajustement des paramètres de ces fonctions.De nombreux travaux sur l'analyse régionale des précipitations s'inspirent de l'approche développée en régionalisation des crues. Les méthodes de types indice de crues ont été utilisées par plusieurs auteurs. Les régions homogènes établies peuvent être contiguës ou non-contiguës. L'analyse multivariée a été utilisée pour déterminer plusieurs régions homogènes au Canada. L'adéquation des sites à l'intérieur d'une région homogène a souvent été validée par une application des L-moments, bien que d'autres tests d'homogénéité aient aussi été utilisés.La loi générale des valeurs extrêmes (GEV) est celle qui a le plus souvent été utilisée dans l'analyse régionale des précipitations. D'autres travaux ont porté sur la loi des valeurs extrêmes à deux composantes (TCEV), de même que sur des applications des séries partielles.Peu de travaux ont porté sur les relations intensité durée dans un contexte régional, ni sur les variations saisonnières des paramètres régionaux. Finalement, les recherches ont débuté sur l'application des concepts d'invariance d'échelle et de loi d'échelle. Ces travaux sont jugés prometteurs.Research on the estimation of extreme precipitation events is currently expanding. This field of research is of great importance in hydraulic engineering not only for the design of dams and dikes, but also for municipal engineering designs. In many cases, local data are scarce. In this context, regionalization methods are very useful tools. This paper summarizes the most recent work on the regionalization of precipitation. Steps normally included in any regionalization work are the delineation of homogenous regions, selection a regional probability distribution function and fitting the parameters.Methods to determine homogenous regions are first reviewed. A great deal of work on precipitation was inspired by methods developed for regional flow analysis, especially the index flood approach. Homogenous regions can be contiguous, but in many cases they are not. The region of influence approach, commonly used in hydrological studies, has not been often applied to precipitation data. Homogenous regions can be established using multivariate statistical approaches such as Principal Component Analysis or Factorial Analysis. These approaches have been used in a number of regions in Canada. Sites within a homogenous region may be tested for their appropriateness by calculating local statistics such as the coefficient of variation, coefficient of skewness and kurtosis, and by comparing these statistics to the regional statistics. Another common approach is the use of L-moments. L-moments are linear combinations of ordered statistics and hence are not as sensitive to outliers as conventional moments. Other homogeneity tests have also been used. They include a Chi-squared test on all regional quantiles associated with a given non-exceedance probability, and a Smirnoff test used to validate the inclusion of a station in the homogenous region.Secondly, we review the distributions and fitting methods used in regionalization of precipitation. The most popular distribution function used is the General Extreme Value (GEV) distribution. This distribution has been recommended for precipitation frequency analysis in the United Kingdom. For regional analysis, the GEV is preferred to the Gumbel distribution, which is often used for site-specific frequency analysis of precipitation extremes. L-moments are also often used to calculate the parameters of the GEV distribution. Some applications of the Two-Component Extreme Value (TCEV) distribution also exist. The TCEV has mostly been used to alleviate the concerns over some of the theoretical and practical restrictions of the GEV.Applications of the Partial Duration Series or Peak-Over-Threshold (POT) approach are also described. In the POT approach, events with a magnitude exceeding a certain threshold are considered in the analysis. The occurrence of such exceedances is modelled as a Poisson process. One of the drawbacks of this method is that it is sometimes necessary to select a relatively high threshold in order to comply with the assumption that observations are independent and identically distributed (i.i.d.). The use of a re-parameterised Generalised Pareto distribution has also been suggested by some researchers.Research on depth-duration relations on a regional scale is also discussed. Empirical approaches used in Canada and elsewhere are described. In most cases, the method consists of establishing a non-linear relationship between a quantile associated with a given duration and its return period to a reference quantile, such as a 1-hour rainfall with a 10-year return period. Depth duration relationships cannot be applied uniformly across Canada for events with durations exceeding two hours. Seasonal variability studies in regionalization are relatively scarce, but are required because of the obvious seasonality of precipitation. In many cases, seasonal regimes may lead to different regionalization approaches for the wet and the dry season. Some research has focused on the use of periodic functions to model regional parameters. Another approach consists of converting the occurrence data of a given event in an angular measurement and developing seasonal indices based on this angular measurement.Other promising avenues of research include the scaling approach. The debate over the possibility of scale invariance for precipitation is ongoing. Simple scaling was studied on a number of precipitation data, but the fact that intermittence is common in precipitation regimes and the presence of numerous zero values in the series does not readily lead to proper application of this approach. Recent research has shown that multiple scaling is likely a more promising avenue

    Regional flood frequency analysis based on peaks-over-threshold approach : a case study for South-Eastern Australia

    Get PDF
    Study region: Southeast Australia Study focus: Regional flood frequency analysis (RFFA) is a widely adopted statistical method to estimate design floods in ungauged catchments. Annual maximum flood (AMF) model is the most popular method in developing RFFA techniques. However, the AMF-based approaches are criticised for its limitations in the range of very frequent to frequent flood estimation. As an alternative, the peaks-over-threshold (POT) based approach has shown theoretical advantages in this flood range. POT based RFFA is currently underemployed internationally due to its complexity in modelling. This study develops POT-based RFFA techniques for south-eastern Australia using data from 151 catchments. A comparison is made between ordinary least squares (OLS) and weighted least squares (WLS) methods in developing POT-based RFFA techniques. New hydrological insights for the region: The OLS based method is found to perform better than the WLS. The median relative error values of the developed prediction equations range 31–38%. The new POT-based RFFA technique overcomes the limitations of the current Australian Rainfall and Runoff, which does not have any RFFA technique for very frequent floods. It is expected that these new POT-based RFFA technique will be used in practice in south-east Australia

    Peaks-over-threshold based regional flood frequency analysis using regularised linear models

    Get PDF
    Regional flood frequency analysis (RFFA) is widely used to estimate design floods in ungauged catchments. Most of the RFFA techniques are based on the annual maximum (AM) flood model; however, research has shown that the peaks-over-threshold (POT) model has greater flexibility than the AM model. There is a lack of studies on POT-based RFFA techniques. This paper presents the development of POT-based RFFA techniques, using regularised linear models (least absolute shrinkage and selection operator, ridge regression and elastic net regression). The results of these regularised linear models are compared with multiple linear regression. Data from 145 stream gauging stations of south-east Australia are used in this study. A leave-one-out cross-validation is adopted to compare these regression models. It has been found that the regularised linear models provide quite accurate flood quantile estimates, with a median relative error in the range of 37 to 47%, which outperform the AM-based RFFA techniques currently recommended in the Australian Rainfall and Runoff guideline. The developed RFFA technique can be used to estimate flood quantiles in ungauged catchments in the study region
    • …
    corecore