645 research outputs found

    Cluster-based Kriging approximation algorithms for complexity reduction

    Get PDF
    Kriging or Gaussian Process Regression is applied in many fields as a non-linear regression model as well as a surrogate model in the field of evolutionary computation. However, the computational and space complexity of Kriging, that is cubic and quadratic in the number of data points respectively, becomes a major bottleneck with more and more data available nowadays. In this paper, we propose a general methodology for the complexity reduction, called cluster Kriging, where the whole data set is partitioned into smaller clusters and multiple Kriging models are built on top of them. In addition, four Kriging approximation algorithms are proposed as candidate algorithms within the new framework. Each of these algorithms can be applied to much larger data sets while maintaining the advantages and power of Kriging. The proposed algorithms are explained in detail and compared empirically against a broad set of existing state-of-the-art Kriging approximation methods on a well-defined testing framework. According to the empirical study, the proposed algorithms consistently outperform the existing algorithms. Moreover, some practical suggestions are provided for using the proposed algorithms.Algorithms and the Foundations of Software technolog

    Understanding and Comparing Scalable Gaussian Process Regression for Big Data

    Full text link
    As a non-parametric Bayesian model which produces informative predictive distribution, Gaussian process (GP) has been widely used in various fields, like regression, classification and optimization. The cubic complexity of standard GP however leads to poor scalability, which poses challenges in the era of big data. Hence, various scalable GPs have been developed in the literature in order to improve the scalability while retaining desirable prediction accuracy. This paper devotes to investigating the methodological characteristics and performance of representative global and local scalable GPs including sparse approximations and local aggregations from four main perspectives: scalability, capability, controllability and robustness. The numerical experiments on two toy examples and five real-world datasets with up to 250K points offer the following findings. In terms of scalability, most of the scalable GPs own a time complexity that is linear to the training size. In terms of capability, the sparse approximations capture the long-term spatial correlations, the local aggregations capture the local patterns but suffer from over-fitting in some scenarios. In terms of controllability, we could improve the performance of sparse approximations by simply increasing the inducing size. But this is not the case for local aggregations. In terms of robustness, local aggregations are robust to various initializations of hyperparameters due to the local attention mechanism. Finally, we highlight that the proper hybrid of global and local scalable GPs may be a promising way to improve both the model capability and scalability for big data.Comment: 25 pages, 15 figures, preprint submitted to KB

    Architectural Quality and the Housing Market: Values of the Late Twentieth Century Built Heritage

    Get PDF
    The assessment of the ‘quality’ of built heritage is a complex transdisciplinary issue, which both public administrations and real estate developers need to carefully consider when making any interventions. Recent international climate regulations underline that currently around 75% of buildings in the EU are not energy efficient. In Italy, those inefficient buildings are more than 50 years old and, if subjected to retrofit interventions, risk being totally transformed and losing their historical value in favor of a more contemporary use. This work aimed to study the residential heritage of the second half of the 20th century in the real estate market and to understand if, how, and in what measure the building and architectonical qualities are recognized and monetized by buyers. The city of Turin was chosen as a study area, and residential building qualities were analyzed using two quality indicators to perform a GWR on market POIs. The results highlighted that housing historical qualities are not homogeneously recognized by the real estate market, in favor of green ones. This work can help both public and private bodies to identify which ‘invisible’ quality residential buildings are immediately exploitable for enhancement strategies, with more respectful retrofitting interventions and a modern protection policy

    Stochastic and deterministic algorithms for continuous black-box optimization

    Get PDF
    Continuous optimization is never easy: the exact solution is always a luxury demand and the theory of it is not always analytical and elegant. Continuous optimization, in practice, is essentially about the efficiency: how to obtain the solution with same quality using as minimal resources (e.g., CPU time or memory usage) as possible? In this thesis, the number of function evaluations is considered as the most important resource to save. To achieve this goal, various efforts have been implemented and applied successfully. One research stream focuses on the so-called stochastic variation (mutation) operator, which conducts an (local) exploration of the search space. The efficiency of those operator has been investigated closely, which shows a good stochastic variation should be able to generate a good coverage of the local neighbourhood around the current search solution. This thesis contributes on this issue by formulating a novel stochastic variation that yields good space coverage. Algorithms and the Foundations of Software technolog

    Optimisation of Mobile Communication Networks - OMCO NET

    Get PDF
    The mini conference “Optimisation of Mobile Communication Networks” focuses on advanced methods for search and optimisation applied to wireless communication networks. It is sponsored by Research & Enterprise Fund Southampton Solent University. The conference strives to widen knowledge on advanced search methods capable of optimisation of wireless communications networks. The aim is to provide a forum for exchange of recent knowledge, new ideas and trends in this progressive and challenging area. The conference will popularise new successful approaches on resolving hard tasks such as minimisation of transmit power, cooperative and optimal routing

    Motion Segmentation Aided Super Resolution Image Reconstruction

    Get PDF
    This dissertation addresses Super Resolution (SR) Image Reconstruction focusing on motion segmentation. The main thrust is Information Complexity guided Gaussian Mixture Models (GMMs) for Statistical Background Modeling. In the process of developing our framework we also focus on two other topics; motion trajectories estimation toward global and local scene change detections and image reconstruction to have high resolution (HR) representations of the moving regions. Such a framework is used for dynamic scene understanding and recognition of individuals and threats with the help of the image sequences recorded with either stationary or non-stationary camera systems. We introduce a new technique called Information Complexity guided Statistical Background Modeling. Thus, we successfully employ GMMs, which are optimal with respect to information complexity criteria. Moving objects are segmented out through background subtraction which utilizes the computed background model. This technique produces superior results to competing background modeling strategies. The state-of-the-art SR Image Reconstruction studies combine the information from a set of unremarkably different low resolution (LR) images of static scene to construct an HR representation. The crucial challenge not handled in these studies is accumulating the corresponding information from highly displaced moving objects. In this aspect, a framework of SR Image Reconstruction of the moving objects with such high level of displacements is developed. Our assumption is that LR images are different from each other due to local motion of the objects and the global motion of the scene imposed by non-stationary imaging system. Contrary to traditional SR approaches, we employed several steps. These steps are; the suppression of the global motion, motion segmentation accompanied by background subtraction to extract moving objects, suppression of the local motion of the segmented out regions, and super-resolving accumulated information coming from moving objects rather than the whole scene. This results in a reliable offline SR Image Reconstruction tool which handles several types of dynamic scene changes, compensates the impacts of camera systems, and provides data redundancy through removing the background. The framework proved to be superior to the state-of-the-art algorithms which put no significant effort toward dynamic scene representation of non-stationary camera systems

    ZipWeave: Towards Efficient and Reliable Measurement based Mobile Coverage Maps

    Get PDF

    Practicable methodologies for delivering comprehensive spatial soils information

    Get PDF
    This thesis is concerned with practicable methodologies for delivering comprehensive spatial soil information to end-users. There is a need for relevant spatial soil information to complement objective decision-making for addressing current problems associated with soil degradation; for modelling, monitoring and measurement of particular soil services; and for the general management of soil resources. These are real-world situations, which operate at spatial scales ranging from field to global scales. As such, comprehensive spatial soil information is tailored to meet the spatial scale specifications of the end user, and is of a nature that fully characterises the whole-soil profile with associated prediction uncertainties, and where possible, both the predictions and uncertainties have been independently validated. ‘Practicable’ is an idealistic pursuit but nonetheless necessary because of a need to equip land-holders, private-sector and non-governmental stakeholders and, governmental departments including soil mapping agencies with the necessary tools to ensure wide application of the methodologies to match the demand for relevant spatial soil information. Practicable methodologies are general and computationally efficient; can be applied to a wide range of soil attributes; can handle variable qualities of data; and are effective when working with very large datasets. In this thesis, delivering comprehensive spatial soil information relies on coupling legacy soil information (principally site observations made in the field) with Digital Soil Mapping (DSM) which comprises quantitative, state-of-the-art technologies for soil mapping. After the General Introduction, a review of the literature is given in Chapter 1 which describes the research context of the thesis. The review describes soil mapping first from a historical perspective and rudimentary efforts of mapping soils and then tracks the succession of advances that have been made towards the realisation of populated, digital spatial soil information databases where measures of prediction certainties are also expressed. From the findings of the review, in order to deliver comprehensive spatial soil information to end-users, new research was required to investigate: 1) a general method for digital soil mapping the whole-profile (effectively pseudo-3D) distribution of soil properties; 2) a general method for quantifying the total prediction uncertainties of the digital soil maps that describe the whole-profile distribution of soil properties; 3) a method for validating the whole-profile predictions of soil properties and the quantifications of their uncertainties; 4) a systematic framework for scale manipulations or upscaling and downscaling techniques for digital soil mapping as a means of generating soil information products tailored to the needs of soil information users. Chapters 2 to 6 set about investigating how we might go about doing these with a succession of practicable methodologies. Chapter 2 addressed the need for whole-profile mapping of soil property distribution. Equal-area spline depth functions coupled with DSM facilitated continuous mapping the lateral and vertical distribution of soil properties. The spline function is a useful tool for deriving the continuous variation of soil properties from soil profile and core observations and is also suitable to use for a number of different soil properties. Generally, mapping the continuous depth function of soil properties reveals that the accuracy of the models is highest at the soil surface but progressively decreases with increasing soil depth. Chapter 3 complements the investigations made in Chapter 2 where an empirical method of quantifying prediction uncertainties from DSM was devised. This method was applied for quantifying the uncertainties of whole-profile digital soil maps. Prediction uncertainty with the devised empirical method is expressed as a prediction interval of the underlying model errors. The method is practicable in the sense that it accounts for all sources of uncertainty and is computationally efficient. Furthermore the method is amenable in situations where complex spatial soil prediction functions such as regression kriging approaches are used. Proper evaluation of digital soil maps requires testing the predictions and the quantification of the prediction uncertainties. Chapter 4 devised two new criteria in which to properly evaluate digital soil maps when additional soil samples collected by probability sampling are used for validation. The first criterion addresses the accuracy of the predictions in the presence of uncertainties and is the spatial average of the statistical expectation of the Mean Square Error of a simulated random value (MSES). The second criterion addresses the quality of the uncertainties which is estimated as the total proportion of the study area where the (1-α)-prediction interval (PI) covers the true value (APCP). Ideally these criteria will be coupled with conventional measures of map quality so that objective decisions can be made about the reliability and subsequent suitability of a map for a given purpose. It was revealed in Chapter 4, that the quantifications of uncertainty are susceptible to bias as a result of using legacy soil data to construct spatial soil prediction functions. As a consequence, in addition to an increasing uncertainty with soil depth, there is increasing misspecification of the prediction uncertainties. Chapter 2, 3, and 4 thus represent a framework for delivering whole-soil profile predictions of soil properties and their uncertainties, where both have been assessed or validated across mapping domains at a range of spatial scales for addressing field, farm, regional, catchment, national, continental or global soil-related problems. The direction of Chapters 5 and 6 however addresses issues specifically related to tailoring spatial soil information to the scale specifications of the end-user through the use of scale manipulations on existing digital soil maps. What is proposed in Chapter 5 is a scaling framework that takes into account the scaling triplet of digital soil maps—extent, resolution, and support—and recommends pedometric methodologies for scale manipulation based on the scale entities of the source and destination maps. Upscaling and downscaling are descriptors for moving up to coarser or down to finer scales respectively but may be too general for DSM. Subsequently Fine-gridding and coarse-gridding are operations where the grid spacing changes but support remains unchanged. Deconvolution and convolution are operations where the support always changes, which may or may not involve changing the grid spacing. While disseveration and conflation operations occur when the support and grid size are equal and both are then changed equally and simultaneously. There is an increasing richness of data sources describing the physical distribution of the Earth’s resources with improved qualities and resolutions. To take advantage of this, Chapter 6 devises a novel procedure for downscaling, involving disseveration. The method attempts to maintain the mass balance of the fine scaled predictions with the available coarse scaled information, through an iterative algorithm which attempts to reconstruct the variation of a property at a prescribed fine scale through an empirical function using environmental or covariate information. One of the advantages associated with the devised method is that soil property uncertainties at the coarse scale can be incorporated into the downscaling algorithm. Finally Chapter 7 presents a synthesis of the investigations made in Chapters 2 to 6 and summarises the pertinent findings. Directly from the investigations carried out during this project there are opportunities for further work; both in terms of addressing shortcomings that were highlighted but not investigated in the thesis, and more generally for advancing digital soil mapping to an operational status and beyond

    An Overview on Application of Machine Learning Techniques in Optical Networks

    Get PDF
    Today's telecommunication networks have become sources of enormous amounts of widely heterogeneous data. This information can be retrieved from network traffic traces, network alarms, signal quality indicators, users' behavioral data, etc. Advanced mathematical tools are required to extract meaningful information from these data and take decisions pertaining to the proper functioning of the networks from the network-generated data. Among these mathematical tools, Machine Learning (ML) is regarded as one of the most promising methodological approaches to perform network-data analysis and enable automated network self-configuration and fault management. The adoption of ML techniques in the field of optical communication networks is motivated by the unprecedented growth of network complexity faced by optical networks in the last few years. Such complexity increase is due to the introduction of a huge number of adjustable and interdependent system parameters (e.g., routing configurations, modulation format, symbol rate, coding schemes, etc.) that are enabled by the usage of coherent transmission/reception technologies, advanced digital signal processing and compensation of nonlinear effects in optical fiber propagation. In this paper we provide an overview of the application of ML to optical communications and networking. We classify and survey relevant literature dealing with the topic, and we also provide an introductory tutorial on ML for researchers and practitioners interested in this field. Although a good number of research papers have recently appeared, the application of ML to optical networks is still in its infancy: to stimulate further work in this area, we conclude the paper proposing new possible research directions
    • 

    corecore