103 research outputs found

    FedLE: Federated Learning Client Selection with Lifespan Extension for Edge IoT Networks

    Full text link
    Federated learning (FL) is a distributed and privacy-preserving learning framework for predictive modeling with massive data generated at the edge by Internet of Things (IoT) devices. One major challenge preventing the wide adoption of FL in IoT is the pervasive power supply constraints of IoT devices due to the intensive energy consumption of battery-powered clients for local training and model updates. Low battery levels of clients eventually lead to their early dropouts from edge networks, loss of training data jeopardizing the performance of FL, and their availability to perform other designated tasks. In this paper, we propose FedLE, an energy-efficient client selection framework that enables lifespan extension of edge IoT networks. In FedLE, the clients first run for a minimum epoch to generate their local model update. The models are partially uploaded to the server for calculating similarities between each pair of clients. Clustering is performed against these client pairs to identify those with similar model distributions. In each round, low-powered clients have a lower probability of being selected, delaying the draining of their batteries. Empirical studies show that FedLE outperforms baselines on benchmark datasets and lasts more training rounds than FedAvg with battery power constraints.Comment: 6 pages, 6 figures, accepted to 2023 IEEE International Conference on Communications (ICC 2023

    Machine Learning Methods and Synthetic Data Generation to Predict Large Wildfires

    Get PDF
    Wildfires are becoming more frequent in different parts of the globe, and the ability to predict when and where they will occur is a complex process. Identifying wildfire events with high probability of becoming a large wildfire is an important task for supporting initial attack planning. Different methods, including those that are physics-based, statistical, and based on machine learning (ML) are used in wildfire analysis. Among the whole, those based on machine learning are relatively novel. In addition, because the number of wildfires is much greater than the number of large wildfires, the dataset to be used in a ML model is imbalanced, resulting in overfitting or underfitting the results. In this manuscript, we propose to generate synthetic data from variables of interest together with ML models for the prediction of large wildfires. Specifically, five synthetic data generation methods have been evaluated, and their results are analyzed with four ML methods. The results yield an improvement in the prediction power when synthetic data are used, offering a new method to be taken into account in Decision Support Systems (DSS) when managing wildfires

    SME default prediction: A systematic methodology-focused review

    Get PDF
    This study reviews the methodologies used in the literature to predict failure in small and medium-sized enterprises (SMEs). We identified 145 SMEs’ default prediction studies from 1972 to early 2023. We summarized the methods used in each study. The focus points are estimation methods, sample re-balancing methods, variable selection techniques, validation methods, and variables included in the literature. More than 1,200 factors used in failure prediction models have been identified, along with 54 unique feature selection techniques and 80 unique estimation methods. Over one-third of the studies do not use any feature selection method, and more than one-quarter use only in-sample validation. Our main recommendation for researchers is to use feature selection and validate results using hold-out samples or cross-validation. As an avenue for further research, we suggest in-depth empirical comparisons of estimation methods, feature selection techniques, and sample re-balancing methods based on some large and commonly used datasets.publishedVersio

    Corporate Credit Rating: A Survey

    Full text link
    Corporate credit rating (CCR) plays a very important role in the process of contemporary economic and social development. How to use credit rating methods for enterprises has always been a problem worthy of discussion. Through reading and studying the relevant literature at home and abroad, this paper makes a systematic survey of CCR. This paper combs the context of the development of CCR methods from the three levels: statistical models, machine learning models and neural network models, summarizes the common databases of CCR, and deeply compares the advantages and disadvantages of the models. Finally, this paper summarizes the problems existing in the current research and prospects the future of CCR. Compared with the existing review of CCR, this paper expounds and analyzes the progress of neural network model in this field in recent years.Comment: 11 page

    Predictive Analysis of Healthcare-Associated Blood Stream Infections in the Neonatal Intensive Care Unit Using Artificial Intelligence: A Single Center Study

    Get PDF
    Background: Neonatal infections represent one of the six main types of healthcare-associated infections and have resulted in increasing mortality rates in recent years due to preterm births or problems arising from childbirth. Although advances in obstetrics and technologies have minimized the number of deaths related to birth, different challenges have emerged in identifying the main factors affecting mortality and morbidity. Dataset characterization: We investigated healthcare-associated infections in a cohort of 1203 patients at the level III Neonatal Intensive Care Unit (ICU) of the “Federico II” University Hospital in Naples from 2016 to 2020 (60 months). Methods: The present paper used statistical analyses and logistic regression to identify an association between healthcare-associated blood stream infection (HABSIs) and the available risk factors in neonates and prevent their spread. We designed a supervised approach to predict whether a patient suffered from HABSI using seven different artificial intelligence models. Results: We analyzed a cohort of 1203 patients and found that birthweight and central line catheterization days were the most important predictors of suffering from HABSI. Conclusions: Our statistical analyses showed that birthweight and central line catheterization days were significant predictors of suffering from HABSI. Patients suffering from HABSI had lower gestational age and birthweight, which led to longer hospitalization and umbilical and central line catheterization days than non-HABSI neonates. The predictive analysis achieved the highest Area Under Curve (AUC), accuracy and F1-macro score in the prediction of HABSIs using Logistic Regression (LR) and Multi-layer Perceptron (MLP) models, which better resolved the imbalanced dataset (65 infected and 1038 healthy)

    Decision support system for the production of miscanthus and willow briquettes

    Get PDF
    The biomass is regarded as a part of renewable energy sources (RES), which can satisfy energy demands. Biomass obtained from plantations is characterized by low bulk density, which increases transport and storage costs. Briquetting is a technology that relies on pressing biomass with the aim of obtaining a denser product (briquettes). In the production of solid biofuels, the technological as well as material variables significantly influence the densification process, and as a result influence the end quality of briquette. This process progresses differently for different materials. Therefore, the optimal selection of process’ parameters is very difficult. It is necessary to use a decision support tool—decision support system (DSS). The purpose of the work was to develop a decision support system that would indicate the optimal parameters for conducting the process of producing Miscanthus and willow briquettes (pre-comminution, milling and briquetting), briquette parameters (durability and specific density) and total energy consumption based on process simulation. Artificial neural networks (ANNs) were used to describe the relationship between individual parameters of the briquette production process. DSS has the form of a web application and is opened from a web browser (it is possible to open it on various types of devices). The modular design allows the modification and expansion the application in the future

    Metaheuristic design of feedforward neural networks: a review of two decades of research

    Get PDF
    Over the past two decades, the feedforward neural network (FNN) optimization has been a key interest among the researchers and practitioners of multiple disciplines. The FNN optimization is often viewed from the various perspectives: the optimization of weights, network architecture, activation nodes, learning parameters, learning environment, etc. Researchers adopted such different viewpoints mainly to improve the FNN's generalization ability. The gradient-descent algorithm such as backpropagation has been widely applied to optimize the FNNs. Its success is evident from the FNN's application to numerous real-world problems. However, due to the limitations of the gradient-based optimization methods, the metaheuristic algorithms including the evolutionary algorithms, swarm intelligence, etc., are still being widely explored by the researchers aiming to obtain generalized FNN for a given problem. This article attempts to summarize a broad spectrum of FNN optimization methodologies including conventional and metaheuristic approaches. This article also tries to connect various research directions emerged out of the FNN optimization practices, such as evolving neural network (NN), cooperative coevolution NN, complex-valued NN, deep learning, extreme learning machine, quantum NN, etc. Additionally, it provides interesting research challenges for future research to cope-up with the present information processing era

    An Evolutionary Neural Network Approach for Slopes Stability Assessment

    Get PDF
    A current big challenge for developed or developing countries is how to keep large-scale transportation infrastructure networks operational under all conditions. Network extensions and budgetary constraints for maintenance purposes are among the main factors that make transportation network management a non-trivial task. On the other hand, the high number of parameters affecting the stability condition of engineered slopes makes their assessment even more complex and difficult to accomplish. Aiming to help achieve the more efficient management of such an important element of modern society, a first attempt at the development of a classification system for rock and soil cuttings, as well as embankments based on visual features, was made in this paper using soft computing algorithms. The achieved results, although interesting, nevertheless have some important limitations to their successful use as auxiliary tools for transportation network management tasks. Accordingly, we carried out new experiments through the combination of modern optimization and soft computing algorithms. Thus, one of the main challenges to overcome is related to the selection of the best set of input features for a feedforward neural network for earthwork hazard category (EHC) identification. We applied a genetic algorithm (GA) for this purpose. Another challenging task is related to the asymmetric distribution of the data (since typically good conditions are much more common than bad ones). To address this question, three training sampling approaches were explored: no resampling, the synthetic minority oversampling technique (SMOTE), and oversampling. Some relevant observations were taken from the optimization process, namely, the identification of which variables are more frequently selected for EHC identification. After finding the most efficient models, a detailed sensitivity analysis was applied over the selected models, allowing us to measure the relative importance of each attribute in EHC identification

    Identifying Critical Neurons in ANN Architectures using Mixed Integer Programming

    Full text link
    We introduce a mixed integer program (MIP) for assigning importance scores to each neuron in deep neural network architectures which is guided by the impact of their simultaneous pruning on the main learning task of the network. By carefully devising the objective function of the MIP, we drive the solver to minimize the number of critical neurons (i.e., with high importance score) that need to be kept for maintaining the overall accuracy of the trained neural network. Further, the proposed formulation generalizes the recently considered lottery ticket optimization by identifying multiple "lucky" sub-networks resulting in optimized architecture that not only performs well on a single dataset, but also generalizes across multiple ones upon retraining of network weights. Finally, we present a scalable implementation of our method by decoupling the importance scores across layers using auxiliary networks. We demonstrate the ability of our formulation to prune neural networks with marginal loss in accuracy and generalizability on popular datasets and architectures.Comment: 16 pages, 3 figures, 5 tables, under revie
    corecore