139 research outputs found

    GRAIMATTER Green Paper:Recommendations for disclosure control of trained Machine Learning (ML) models from Trusted Research Environments (TREs)

    Get PDF
    TREs are widely, and increasingly used to support statistical analysis of sensitive data across a range of sectors (e.g., health, police, tax and education) as they enable secure and transparent research whilst protecting data confidentiality.There is an increasing desire from academia and industry to train AI models in TREs. The field of AI is developing quickly with applications including spotting human errors, streamlining processes, task automation and decision support. These complex AI models require more information to describe and reproduce, increasing the possibility that sensitive personal data can be inferred from such descriptions. TREs do not have mature processes and controls against these risks. This is a complex topic, and it is unreasonable to expect all TREs to be aware of all risks or that TRE researchers have addressed these risks in AI-specific training.GRAIMATTER has developed a draft set of usable recommendations for TREs to guard against the additional risks when disclosing trained AI models from TREs. The development of these recommendations has been funded by the GRAIMATTER UKRI DARE UK sprint research project. This version of our recommendations was published at the end of the project in September 2022. During the course of the project, we have identified many areas for future investigations to expand and test these recommendations in practice. Therefore, we expect that this document will evolve over time. The GRAIMATTER DARE UK sprint project has also developed a minimal viable product (MVP) as a suite of attack simulations that can be applied by TREs and can be accessed here (https://github.com/AI-SDC/AI-SDC).If you would like to provide feedback or would like to learn more, please contact Smarti Reel ([email protected]) and Emily Jefferson ([email protected]).The summary of our recommendations for a general public audience can be found at DOI: 10.5281/zenodo.708951

    An extensive study on iterative solver resilience : characterization, detection and prediction

    Get PDF
    Soft errors caused by transient bit flips have the potential to significantly impactan applicalion's behavior. This has motivated the design of an array of techniques to detect, isolate, and correct soft errors using microarchitectural, architectural, compilation­based, or application-level techniques to minimize their impact on the executing application. The first step toward the design of good error detection/correction techniques involves an understanding of an application's vulnerability to soft errors. This work focuses on silent data e orruption's effects on iterative solvers and efforts to mitigate those effects. In this thesis, we first present the first comprehensive characterizalion of !he impact of soft errors on !he convergen ce characteris tics of six iterative methods using application-level fault injection. We analyze the impact of soft errors In terms of the type of error (single-vs multi-bit), the distribution and location of bits affected, the data structure and statement impacted, and varialion with time. We create a public access database with more than 1.5 million fault injection results. We then analyze the performance of soft error detection mechanisms and present the comparalive results. Molivated by our observations, we evaluate a machine-learning based detector that takes as features that are the runtime features observed by the individual detectors to arrive al their conclusions. Our evalualion demonstrates improved results over individual detectors. We then propase amachine learning based method to predict a program's error behavior to make fault injection studies more efficient. We demonstrate this method on asse ssing the performance of soft error detectors. We show that our method maintains 84% accuracy on average with up to 53% less cost. We also show, once a model is trained further fault injection tests would cost 10% of the expected full fault injection runs.“Soft errors” causados por cambios de estado transitorios en bits, tienen el potencial de impactar significativamente el comportamiento de una aplicación. Esto, ha motivado el diseño de una variedad de técnicas para detectar, aislar y corregir soft errors aplicadas a micro-arquitecturas, arquitecturas, tiempo de compilación y a nivel de aplicación para minimizar su impacto en la ejecución de una aplicación. El primer paso para diseñar una buna técnica de detección/corrección de errores, implica el conocimiento de las vulnerabilidades de la aplicación ante posibles soft errors. Este trabajo se centra en los efectos de la corrupción silenciosa de datos en soluciones iterativas, así como en los esfuerzos para mitigar esos efectos. En esta tesis, primeramente, presentamos la primera caracterización extensiva del impacto de soft errors sobre las características convergentes de seis métodos iterativos usando inyección de fallos a nivel de aplicación. Analizamos el impacto de los soft errors en términos del tipo de error (único vs múltiples-bits), de la distribución y posición de los bits afectados, las estructuras de datos, instrucciones afectadas y de las variaciones en el tiempo. Creamos una base de datos pública con más de 1.5 millones de resultados de inyección de fallos. Después, analizamos el desempeño de mecanismos de detección de soft errors actuales y presentamos los resultados de su comparación. Motivados por las observaciones de los resultados presentados, evaluamos un detector de soft errors basado en técnicas de machine learning que toma como entrada las características observadas en el tiempo de ejecución individual de los detectores anteriores al llegar a su conclusión. La evaluación de los resultados obtenidos muestra una mejora por sobre los detectores individualmente. Basados en estos resultados propusimos un método basado en machine learning para predecir el comportamiento de los errores en un programa con el fin de hacer el estudio de inyección de errores mas eficiente. Presentamos este método para evaluar el rendimiento de los detectores de soft errors. Demostramos que nuestro método mantiene una precisión del 84% en promedio con hasta un 53% de mejora en el tiempo de ejecución. También mostramos que una vez que un modelo ha sido entrenado, las pruebas de inyección de errores siguientes costarían 10% del tiempo esperado de ejecución.Postprint (published version

    Delineation of groundwater potential zones by means of ensemble tree supervised classification methods in the Eastern Lake Chad basin

    Get PDF
    This paper presents a machine learning method to map groundwater potential in crystalline domains. First, a spatially-distributed set of explanatory variables for groundwater occurrence is compiled into a geographic information system. Twenty machine learning classifiers are subsequently trained on a sample of 488 boreholes and excavated wells for a region of eastern Chad. This process includes collinearity, cross-validation, feature elimination and parameter fitting routines. Random forest and extra trees classifiers outperformed other algorithms (test score > 0.80, balanced score > 0.80, AUC > 0.87). Fracture density, slope, SAR coherence (interferometric correlation), topographic wetness index, basement depth, distance to channels and slope aspect proved the most relevant explanatory variables. Three major conclusions stem from this work: (1) using a large number of supervised classification algorithms is advisable in groundwater potential studies; (2) the choice of performance metrics constrains the relevance of explanatory variables; and (3) seasonal variations from satellite images contribute to successful groundwater potential mapping

    Closing the Loop: the Capacities and Constraints of Information and Communication Technologies for Development (ICT4D)

    Get PDF
    As a mechanism for collecting and sharing information, information and communications technologies (ICT) hold immense potential for individuals and institutions in low- and middle-income countries. Currently the distribution and adoption of ICTs--particularly mobile devices--has far outpaced the provision of other household services like clean water, sanitation, hygiene, or energy services. At the same time, the development and deployment of Internet of Things (IoT) devices including cellular- and satellite-connected sensors is facilitating more rapid feedback from remote regions where basic services are most limited. When used in conjunction with economic development or public health interventions, these devices and the feedback they provide can inform operation and maintenance activities for field staff and improve the monitoring and evaluation of outcomes for project stakeholders. This dissertation includes three chapters written as journal articles. While each chapter is framed around the work and research efforts being undertaken by the Sustainable Water, Energy, and Environmental Technologies Lab (SweetLab) at Portland State University, the common thread that weaves all three investigations together is the theme of ICT-enabled programmatic feedback. The first chapter introduces the three theoretical lenses that inform these investigations and the ways that ICTs and the data they provide can (1) serve as more appropriate proxies for measuring access to services, (2) reduce information asymmetries between various stakeholders including communities, governments, implementers, and funders, and (3) enable more robust methodologies for measuring outcomes and impacts of interventions within complex adaptive systems. The second chapter presents a critical review of the methodologies and technologies being used to track progress on sanitation and hygiene development goals. Chapter three describes how simple sensors and weight measurements can be combined with complex machine learning algorithms to facilitate more reliable and cost-effective latrine servicing in informal settlements. Chapter four presents the results from an investigation exploring how near-time feedback from sensors installed on motorized boreholes can improve water service delivery and drought resilience in arid regions of Northern Kenya. Finally, chapter five provides a summary of the three manuscripts and discusses the significance of this research for future investigations

    A comparison of the CAR and DAGAR spatial random effects models with an application to diabetics rate estimation in Belgium

    Get PDF
    When hierarchically modelling an epidemiological phenomenon on a finite collection of sites in space, one must always take a latent spatial effect into account in order to capture the correlation structure that links the phenomenon to the territory. In this work, we compare two autoregressive spatial models that can be used for this purpose: the classical CAR model and the more recent DAGAR model. Differently from the former, the latter has a desirable property: its ρ parameter can be naturally interpreted as the average neighbor pair correlation and, in addition, this parameter can be directly estimated when the effect is modelled using a DAGAR rather than a CAR structure. As an application, we model the diabetics rate in Belgium in 2014 and show the adequacy of these models in predicting the response variable when no covariates are available

    A Statistical Approach to the Alignment of fMRI Data

    Get PDF
    Multi-subject functional Magnetic Resonance Image studies are critical. The anatomical and functional structure varies across subjects, so the image alignment is necessary. We define a probabilistic model to describe functional alignment. Imposing a prior distribution, as the matrix Fisher Von Mises distribution, of the orthogonal transformation parameter, the anatomical information is embedded in the estimation of the parameters, i.e., penalizing the combination of spatially distant voxels. Real applications show an improvement in the classification and interpretability of the results compared to various functional alignment methods

    Innovation in manufacturing through digital technologies and applications: Thoughts and Reflections on Industry 4.0

    Get PDF
    The rapid pace of developments in digital technologies offers many opportunities to increase the efficiency, flexibility and sophistication of manufacturing processes; including the potential for easier customisation, lower volumes and rapid changeover of products within the same manufacturing cell or line. A number of initiatives on this theme have been proposed around the world to support national industries under names such as Industry 4.0 (Industrie 4.0 in Germany, Made-in-China in China and Made Smarter in the UK). This book presents an overview of the state of art and upcoming developments in digital technologies pertaining to manufacturing. The starting point is an introduction on Industry 4.0 and its potential for enhancing the manufacturing process. Later on moving to the design of smart (that is digitally driven) business processes which are going to rely on sensing of all relevant parameters, gathering, storing and processing the data from these sensors, using computing power and intelligence at the most appropriate points in the digital workflow including application of edge computing and parallel processing. A key component of this workflow is the application of Artificial Intelligence and particularly techniques in Machine Learning to derive actionable information from this data; be it real-time automated responses such as actuating transducers or informing human operators to follow specified standard operating procedures or providing management data for operational and strategic planning. Further consideration also needs to be given to the properties and behaviours of particular machines that are controlled and materials that are transformed during the manufacturing process and this is sometimes referred to as Operational Technology (OT) as opposed to IT. The digital capture of these properties and behaviours can then be used to define so-called Cyber Physical Systems. Given the power of these digital technologies it is of paramount importance that they operate safely and are not vulnerable to malicious interference. Industry 4.0 brings unprecedented cybersecurity challenges to manufacturing and the overall industrial sector and the case is made here that new codes of practice are needed for the combined Information Technology and Operational Technology worlds, but with a framework that should be native to Industry 4.0. Current computing technologies are also able to go in other directions than supporting the digital ‘sense to action’ process described above. One of these is to use digital technologies to enhance the ability of the human operators who are still essential within the manufacturing process. One such technology, that has recently become accessible for widespread adoption, is Augmented Reality, providing operators with real-time additional information in situ with the machines that they interact with in their workspace in a hands-free mode. Finally, two linked chapters discuss the specific application of digital technologies to High Pressure Die Casting (HDPC) of Magnesium components. Optimizing the HPDC process is a key task for increasing productivity and reducing defective parts and the first chapter provides an overview of the HPDC process with attention to the most common defects and their sources. It does this by first looking at real-time process control mechanisms, understanding the various process variables and assessing their impact on the end product quality. This understanding drives the choice of sensing methods and the associated smart digital workflow to allow real-time control and mitigation of variation in the identified variables. Also, data from this workflow can be captured and used for the design of optimised dies and associated processes

    Proceedings of the National Conference on Water, Food Security and Climate Change in Sri Lanka, BMICH, Colombo, Sri Lanka, 9-11 June 2009. Vol. 2. Water quality, environment and climate change

    Get PDF
    This is the second volume of the proceedings of the National Conference on ‘Water for Food and Environment’, which was held from June 9 –11, 2009 at the Bandaranaike Memorial International Conference Hall (BMICH). The volumes 1 and 3 have been produced as separate documents of this report series. In response to a call for abstracts, 81 abstracts were received from government institutes dealing with water resources and agriculture development, universities, other freelance researchers and researchers from the International Water Management Institute (IWMI). Forty seven of the eighty-one abstracts that were submitted were accepted for compiling full papers. In the past couple of years the sharp increase in food prices worldwide has raised serious concerns about food security, especially in developing countries. To effectively address these concerns a holistic approach is required that encompasses improved agricultural water productivity, adaptation to climate change, targeted and appropriate institutional and financial measures, and a consideration of environmental issues. The main purpose of the conference was to share experiences in these areas and to find opportunities to improve farmers’ incomes and food production, and to promote environmentally sustainable practices in Sri Lanka in the face of growing water scarcity and the challenges of climate change.Wetlands, / Environmental flows / Assessment / Water allocation / Case studies / Rivers / Farming systems / Models / Groundwater / Aquifers / Domestic water / Rainfed farming / Climate change / Wastewater irrigation / Rice
    corecore