1,162 research outputs found

    The privatization of the Russian coal industry: policies and processes in the transformation of a major industry

    Get PDF
    This paper provides an overview of the privatization of the Russian coal industry. It reviews the salient aspects of the Government's privatization policy as it evolved over the years, and looks at the reasons for the successes and the pitfalls encountered along the way. Specific procedures and methods of sale are described in detail. A profile of the new owners of the industry is given, with a look at the implications for competition in the industry and at first performance indicators. As the World Bank has been closely involved in the support of the Government's coal sector restructuring program through provision of financing and policy advice, throughout the paper aspects of World Bank advice are considered.Municipal Financial Management,Banks&Banking Reform,Non Bank Financial Institutions,Environmental Economics&Policies,Water and Industry,Water and Industry,Non Bank Financial Institutions,Mining&Extractive Industry (Non-Energy),Banks&Banking Reform,Municipal Financial Management

    Mining complex trees for hidden fruit : a graph–based computational solution to detect latent criminal networks : a thesis presented in partial fulfilment of the requirements for the degree of Doctor of Philosophy in Information Technology at Massey University, Albany, New Zealand.

    Get PDF
    The detection of crime is a complex and difficult endeavour. Public and private organisations – focusing on law enforcement, intelligence, and compliance – commonly apply the rational isolated actor approach premised on observability and materiality. This is manifested largely as conducting entity-level risk management sourcing ‘leads’ from reactive covert human intelligence sources and/or proactive sources by applying simple rules-based models. Focusing on discrete observable and material actors simply ignores that criminal activity exists within a complex system deriving its fundamental structural fabric from the complex interactions between actors - with those most unobservable likely to be both criminally proficient and influential. The graph-based computational solution developed to detect latent criminal networks is a response to the inadequacy of the rational isolated actor approach that ignores the connectedness and complexity of criminality. The core computational solution, written in the R language, consists of novel entity resolution, link discovery, and knowledge discovery technology. Entity resolution enables the fusion of multiple datasets with high accuracy (mean F-measure of 0.986 versus competitors 0.872), generating a graph-based expressive view of the problem. Link discovery is comprised of link prediction and link inference, enabling the high-performance detection (accuracy of ~0.8 versus relevant published models ~0.45) of unobserved relationships such as identity fraud. Knowledge discovery uses the fused graph generated and applies the “GraphExtract” algorithm to create a set of subgraphs representing latent functional criminal groups, and a mesoscopic graph representing how this set of criminal groups are interconnected. Latent knowledge is generated from a range of metrics including the “Super-broker” metric and attitude prediction. The computational solution has been evaluated on a range of datasets that mimic an applied setting, demonstrating a scalable (tested on ~18 million node graphs) and performant (~33 hours runtime on a non-distributed platform) solution that successfully detects relevant latent functional criminal groups in around 90% of cases sampled and enables the contextual understanding of the broader criminal system through the mesoscopic graph and associated metadata. The augmented data assets generated provide a multi-perspective systems view of criminal activity that enable advanced informed decision making across the microscopic mesoscopic macroscopic spectrum

    GHG inventories and drivers & barriers to climate action : an analysis of Brazilian cities

    Get PDF
    Cities are considered core of the global climate change mitigation and strategic low-carbon development and city-level study is a trend for climate change responses studies. The literature review identified three gaps that guided this research: (I) There is lack of consistent and comparable greenhouse gas (GHG) emissions data at the city level; (II) it is necessary to analyze enabling factors that lead to effective urban climate governance and (III) there is a geographical bias of empirical climate governance studies focused on cities from Global North and developed countries realities. To address these gaps, considering the reality of Global South, this thesis had the general objectives of to analyze (I) the quality and gaps of GHG inventories and (II) the main drivers and barriers to climate agenda strengthening in Brazilian cities. Four manuscripts were developed to reach this objective. The first analyzed the differences among the main existing GHG accounting methodologies for cities and identified gaps in carbon inventories of twenty-four Brazilian cities. The second paper compared GHG emissions results of forty-seven Brazilian cities, applying different GHG accounting methodologies. It highlighted characteristics, similarities, and differences of these methodologies, showing how they can impact GHG results. The third and fourth articles discussed about the climate agenda advance in the city of Recife in Brazil. The third paper is a city profile which discusses how geographical characteristics and the historical urbanization process of the city have contributed to the climate risks and vulnerabilities. It evidences factors that can decisively assist cities to strength the climate agenda, mainly in developing or less developed countries. The fourth paper is a case study which discusses the climate actions adopted in Recife and it examines the main drivers and barriers to their effective implementation, comparing to examples from literature. The main important findings are: (I) There are two main types of GHG reporting gaps: incompleteness and lack of transparency which hinder the accuracy, assessment of results and comparability between cities; (II) to analyze GHG reports and to compare results, it is essential to identify methodology, base year, emission sources included, global warming potential, and calculation methods, information which are not transparent in several reports; (III) the drivers to climate action identified include having committed leadership, being part of a multinational network of cities and multilevel governance which supports existing theory, as well as identified climate risks, much in contrast to other European cases; (IV) It is fundamental to institutionalize the climate agenda in the local government to avoid political interferences, which was considered a primary barrier. The thesis provides insights for academics and policymakers on how develop broader, completer, and more transparent GHG inventories and it evidences precautions that should be taken when analyzing a city GHG report. It also evidences factors that can decisively assist cities to strength the climate agenda, mainly in developing or less developed countries, providing insights to academics and policymakers on low carbon strategies for cities. Moreover, it suggests steps that can assist cities to adopt climate actions, particularly in developing or less developed countries

    Deterministic and Probabilistic Risk Management Approaches in Construction Projects: A Systematic Literature Review and Comparative Analysis

    Get PDF
    Risks and uncertainties are inevitable in construction projects and can drastically change the expected outcome, negatively impacting the project’s success. However, risk management (RM) is still conducted in a manual, largely ineffective, and experience-based fashion, hindering automation and knowledge transfer in projects. The construction industry is benefitting from the recent Industry 4.0 revolution and the advancements in data science branches, such as artificial intelligence (AI), for the digitalization and optimization of processes. Data-driven methods, e.g., AI and machine learning algorithms, Bayesian inference, and fuzzy logic, are being widely explored as possible solutions to RM domain shortcomings. These methods use deterministic or probabilistic risk reasoning approaches, the first of which proposes a fixed predicted value, and the latter embraces the notion of uncertainty, causal dependencies, and inferences between variables affecting projects’ risk in the predicted value. This research used a systematic literature review method with the objective of investigating and comparatively analyzing the main deterministic and probabilistic methods applied to construction RM in respect of scope, primary applications, advantages, disadvantages, limitations, and proven accuracy. The findings established recommendations for optimum AI-based frameworks for different management levels—enterprise, project, and operational—for large or small data sets

    Spatial Data Quality in the IoT Era:Management and Exploitation

    Get PDF
    Within the rapidly expanding Internet of Things (IoT), growing amounts of spatially referenced data are being generated. Due to the dynamic, decentralized, and heterogeneous nature of the IoT, spatial IoT data (SID) quality has attracted considerable attention in academia and industry. How to invent and use technologies for managing spatial data quality and exploiting low-quality spatial data are key challenges in the IoT. In this tutorial, we highlight the SID consumption requirements in applications and offer an overview of spatial data quality in the IoT setting. In addition, we review pertinent technologies for quality management and low-quality data exploitation, and we identify trends and future directions for quality-aware SID management and utilization. The tutorial aims to not only help researchers and practitioners to better comprehend SID quality challenges and solutions, but also offer insights that may enable innovative research and applications

    Software engineering for AI-based systems: A survey

    Get PDF
    AI-based systems are software systems with functionalities enabled by at least one AI component (e.g., for image-, speech-recognition, and autonomous driving). AI-based systems are becoming pervasive in society due to advances in AI. However, there is limited synthesized knowledge on Software Engineering (SE) approaches for building, operating, and maintaining AI-based systems. To collect and analyze state-of-the-art knowledge about SE for AI-based systems, we conducted a systematic mapping study. We considered 248 studies published between January 2010 and March 2020. SE for AI-based systems is an emerging research area, where more than 2/3 of the studies have been published since 2018. The most studied properties of AI-based systems are dependability and safety. We identified multiple SE approaches for AI-based systems, which we classified according to the SWEBOK areas. Studies related to software testing and software quality are very prevalent, while areas like software maintenance seem neglected. Data-related issues are the most recurrent challenges. Our results are valuable for: researchers, to quickly understand the state-of-the-art and learn which topics need more research; practitioners, to learn about the approaches and challenges that SE entails for AI-based systems; and, educators, to bridge the gap among SE and AI in their curricula.This work has been partially funded by the “Beatriz Galindo” Spanish Program BEAGAL18/00064 and by the DOGO4ML Spanish research project (ref. PID2020-117191RB-I00)Peer ReviewedPostprint (author's final draft

    The Problem of Data Extraction in Social Media: A Theoretical Framework

    Get PDF
    In today's rapidly evolving digital landscape, the pervasive growth of social media platforms has resulted in an era of unprecedented data generation. These platforms are responsible for generating vast volumes of data on a daily basis, forming intricate webs of patterns and connections that harbor invaluable insights crucial for informed decision-making. Recognizing the significance of exploring social media data, researchers have increasingly turned their attention towards leveraging this data to address a wide array of social research issues. Unlike conventional data collection methods such as questionnaires, interviews, or focus groups, social media data presents unique challenges and opportunities, demanding specialized techniques for its extraction and analysis. However, the absence of a standardized and systematic approach to collect and preprocess social media data remains a gap in the field. This gap not only compromises the quality and credibility of subsequent data analysis but also hinders the realization of the full potential inherent in social media data. This paper aims to bridge this gap by presenting a comprehensive framework designed for the systematic extraction and processing of social media data. The proposed framework offers a clear, step-by-step methodology for the extraction and processing of social media data for analysis. In an era where social media data serves as a pivotal resource for understanding human behavior, sentiment, and societal dynamics, this framework offers a foundational toolset for researchers and practitioners seeking to harness the wealth of insights concealed within the vast expanse of social media data

    Predictive Modelling Approach to Data-Driven Computational Preventive Medicine

    Get PDF
    This thesis contributes novel predictive modelling approaches to data-driven computational preventive medicine and offers an alternative framework to statistical analysis in preventive medicine research. In the early parts of this research, this thesis presents research by proposing a synergy of machine learning methods for detecting patterns and developing inexpensive predictive models from healthcare data to classify the potential occurrence of adverse health events. In particular, the data-driven methodology is founded upon a heuristic-systematic assessment of several machine-learning methods, data preprocessing techniques, models’ training estimation and optimisation, and performance evaluation, yielding a novel computational data-driven framework, Octopus. Midway through this research, this thesis advances research in preventive medicine and data mining by proposing several new extensions in data preparation and preprocessing. It offers new recommendations for data quality assessment checks, a novel multimethod imputation (MMI) process for missing data mitigation, a novel imbalanced resampling approach, and minority pattern reconstruction (MPR) led by information theory. This thesis also extends the area of model performance evaluation with a novel classification performance ranking metric called XDistance. In particular, the experimental results show that building predictive models with the methods guided by our new framework (Octopus) yields domain experts' approval of the new reliable models’ performance. Also, performing the data quality checks and applying the MMI process led healthcare practitioners to outweigh predictive reliability over interpretability. The application of MPR and its hybrid resampling strategies led to better performances in line with experts' success criteria than the traditional imbalanced data resampling techniques. Finally, the use of the XDistance performance ranking metric was found to be more effective in ranking several classifiers' performances while offering an indication of class bias, unlike existing performance metrics The overall contributions of this thesis can be summarised as follow. First, several data mining techniques were thoroughly assessed to formulate the new Octopus framework to produce new reliable classifiers. In addition, we offer a further understanding of the impact of newly engineered features, the physical activity index (PAI) and biological effective dose (BED). Second, the newly developed methods within the new framework. Finally, the newly accepted developed predictive models help detect adverse health events, namely, visceral fat-associated diseases and advanced breast cancer radiotherapy toxicity side effects. These contributions could be used to guide future theories, experiments and healthcare interventions in preventive medicine and data mining
    corecore