96 research outputs found

    Towards accurate prediction for high-dimensional and highly-variable cloud workloads with deep learning

    Get PDF
    This is the author accepted manuscript. The final version is available from IEEE via the DOI in this recordResource provisioning for cloud computing necessitates the adaptive and accurate prediction of cloud workloads. However, the existing methods cannot effectively predict the high-dimensional and highly-variable cloud workloads. This results in resource wasting and inability to satisfy service level agreements (SLAs). Since recurrent neural network (RNN) is naturally suitable for sequential data analysis, it has been recently used to tackle the problem of workload prediction. However, RNN often performs poorly on learning longterm memory dependencies, and thus cannot make the accurate prediction of workloads. To address these important challenges, we propose a deep Learning based Prediction Algorithm for cloud Workloads (L-PAW). First, a top-sparse auto-encoder (TSA) is designed to effectively extract the essential representations of workloads from the original high-dimensional workload data. Next, we integrate TSA and gated recurrent unit (GRU) block into RNN to achieve the adaptive and accurate prediction for highly-variable workloads. Using realworld workload traces from Google and Alibaba cloud data centers and the DUX-based cluster, extensive experiments are conducted to demonstrate the effectiveness and adaptability of the L-PAW for different types of workloads with various prediction lengths. Moreover, the performance results show that the L-PAW achieves superior prediction accuracy compared to the classic RNN-based and other workload prediction methods for high-dimensional and highly-variable real-world cloud workloads

    Reconfiguration of optical-NFV network architectures based on cloud resource allocation and QoS degradation cost-aware prediction techniques

    Get PDF
    The high time required for the deployment of cloud resources in Network Function Virtualization network architectures has led to the proposal and investigation of algorithms for predicting trafc or the necessary processing and memory resources. However, it is well known that whatever approach is taken, a prediction error is inevitable. Two types of prediction errors can occur that have a different impact on the increase in network operational costs. In case the predicted values are higher than the real ones, the resource allocation algorithms will allocate more resources than necessary with the consequent introduction of an over-provisioning cost. Conversely, when the predicted values are lower than the real values, the allocation of fewer resources will lead to a degradation of QoS and the introduction of an under-provisioning cost. When over-provisioning and under-provisioning costs are different, most of the prediction algorithms proposed in the literature are not adequate because they are based on minimizing the mean square error or symmetric cost functions. For this reason we propose and investigate a forecasting methodology in which it is introduced an asymmetric cost function capable of weighing the costs of over-provisioning and under-provisioning differently. We have applied the proposed forecasting methodology for resource allocation in a Network Function Virtualization architectures where the Network Function Virtualization Infrastructure Point-of-Presences are interconnected by an elastic optical network.We have veried a cost savings of 40% compared to solutions that provide a minimization of the mean square error

    An Efficient Machine Learning Software Architecture for Internet of Things

    Get PDF
    Internet of Things (IoT) software is becoming a critical infrastructure for many domains. In IoT, sensors monitor their environment and transfer readings to cloud, where Machine Learning (ML) provides insights to decision-makers. In the healthcare domain, the IoT software designers have to consider privacy, real-time performance and cost in addition to ML accuracy. We propose an architecture that decomposes the ML lifecycle into components for deployment on a two-tier cloud, edge-core. It enables IoT time-series data to be consumed by ML models on edge-core infrastructure, with pipeline elements deployed on any tier, dynamically. The architecture feasibility and ML accuracy are validated with three brain-computer interfaces (BCI) based use-cases. The contributions are two-fold: first, we propose a novel ML-IoT pipeline software architecture that encompasses essential components from data ingestion to runtime use of ML models; second, we assess the software on cognitive applications and achieve promising results in comparison to literature

    클라우드 컴퓨팅 환경기반에서 수치 모델링과 머신러닝을 통한 지구과학 자료생성에 관한 연구

    Get PDF
    학위논문(박사) -- 서울대학교대학원 : 자연과학대학 지구환경과학부, 2022. 8. 조양기.To investigate changes and phenomena on Earth, many scientists use high-resolution-model results based on numerical models or develop and utilize machine learning-based prediction models with observed data. As information technology advances, there is a need for a practical methodology for generating local and global high-resolution numerical modeling and machine learning-based earth science data. This study recommends data generation and processing using high-resolution numerical models of earth science and machine learning-based prediction models in a cloud environment. To verify the reproducibility and portability of high-resolution numerical ocean model implementation on cloud computing, I simulated and analyzed the performance of a numerical ocean model at various resolutions in the model domain, including the Northwest Pacific Ocean, the East Sea, and the Yellow Sea. With the containerization method, it was possible to respond to changes in various infrastructure environments and achieve computational reproducibility effectively. The data augmentation of subsurface temperature data was performed using generative models to prepare large datasets for model training to predict the vertical temperature distribution in the ocean. To train the prediction model, data augmentation was performed using a generative model for observed data that is relatively insufficient compared to satellite dataset. In addition to observation data, HYCOM datasets were used for performance comparison, and the data distribution of augmented data was similar to the input data distribution. The ensemble method, which combines stand-alone predictive models, improved the performance of the predictive model compared to that of the model based on the existing observed data. Large amounts of computational resources were required for data synthesis, and the synthesis was performed in a cloud-based graphics processing unit environment. High-resolution numerical ocean model simulation, predictive model development, and the data generation method can improve predictive capabilities in the field of ocean science. The numerical modeling and generative models based on cloud computing used in this study can be broadly applied to various fields of earth science.지구의 변화와 현상을 연구하기 위해 많은 과학자들은 수치 모델을 기반으로 한 고해상도 모델 결과를 사용하거나 관측된 데이터로 머신러닝 기반 예측 모델을 개발하고 활용한다. 정보기술이 발전함에 따라 지역 및 전 지구적인 고해상도 수치 모델링과 머신러닝 기반 지구과학 데이터 생성을 위한 실용적인 방법론이 필요하다. 본 연구는 지구과학의 고해상도 수치 모델과 머신러닝 기반 예측 모델을 기반으로 한 데이터 생성 및 처리가 클라우드 환경에서 효과적으로 구현될 수 있음을 제안한다. 클라우드 컴퓨팅에서 고해상도 수치 해양 모델 구현의 재현성과 이식성을 검증하기 위해 북서태평양, 동해, 황해 등 모델 영역의 다양한 해상도에서 수치 해양 모델의 성능을 시뮬레이션하고 분석하였다. 컨테이너화 방식을 통해 다양한 인프라 환경 변화에 대응하고 계산 재현성을 효과적으로 확보할 수 있었다. 머신러닝 기반 데이터 생성의 적용을 검증하기 위해 생성 모델을 이용한 표층 이하 온도 데이터의 데이터 증강을 실행하여 해양의 수직 온도 분포를 예측하는 모델 훈련을 위한 대용량 데이터 세트를 준비했다. 예측모델 훈련을 위해 위성 데이터에 비해 상대적으로 부족한 관측 데이터에 대해서 생성 모델을 사용하여 데이터 증강을 수행하였다. 모델의 예측성능 비교에는 관측 데이터 외에도 HYCOM 데이터 세트를 사용하였으며, 증강 데이터의 데이터 분포는 입력 데이터 분포와 유사함을 확인하였다. 독립형 예측 모델을 결합한 앙상블 방식은 기존 관측 데이터를 기반으로 하는 예측 모델의 성능에 비해 향상되었다. 데이터합성을 위해 많은 양의 계산 자원이 필요했으며, 데이터 합성은 클라우드 기반 GPU 환경에서 수행되었다. 고해상도 수치 해양 모델 시뮬레이션, 예측 모델 개발, 데이터 생성 방법은 해양 과학 분야에서 예측 능력을 향상시킬 수 있다. 본 연구에서 사용된 클라우드 컴퓨팅 기반의 수치 모델링 및 생성 모델은 지구 과학의 다양한 분야에 광범위하게 적용될 수 있다.1. General Introduction 1 2. Performance of numerical ocean modeling on cloud computing 6 2.1. Introduction 6 2.2. Cloud Computing 9 2.2.1. Cloud computing overview 9 2.2.2. Commercial cloud computing services 12 2.3. Numerical model for performance analysis of commercial clouds 15 2.3.1. High Performance Linpack Benchmark 15 2.3.2. Benchmark Sustainable Memory Bandwidth and Memory Latency 16 2.3.3. Numerical Ocean Model 16 2.3.4. Deployment of Numerical Ocean Model and Benchmark Packages on Cloud Clusters 19 2.4. Simulation results 21 2.4.1. Benchmark simulation 21 2.4.2. Ocean model simulation 24 2.5. Analysis of ROMS performance on commercial clouds 26 2.5.1. Performance of ROMS according to H/W resources 26 2.5.2. Performance of ROMS according to grid size 34 2.6. Summary 41 3. Reproducibility of numerical ocean model on the cloud computing 44 3.1. Introduction 44 3.2. Containerization of numerical ocean model 47 3.2.1. Container virtualization 47 3.2.2. Container-based architecture for HPC 49 3.2.3. Container-based architecture for hybrid cloud 53 3.3. Materials and Methods 55 3.3.1. Comparison of traditional and container based HPC cluster workflows 55 3.3.2. Model domain and datasets for numerical simulation 57 3.3.3. Building the container image and registration in the repository 59 3.3.4. Configuring a numeric model execution cluster 64 3.4. Results and Discussion 74 3.4.1. Reproducibility 74 3.4.2. Portability and Performance 76 3.5. Conclusions 81 4. Generative models for the prediction of ocean temperature profile 84 4.1. Introduction 84 4.2. Materials and Methods 87 4.2.1. Model domain and datasets for predicting the subsurface temperature 87 4.2.2. Model architecture for predicting the subsurface temperature 90 4.2.3. Neural network generative models 91 4.2.4. Prediction Models 97 4.2.5. Accuracy 103 4.3. Results and Discussion 104 4.3.1. Data Generation 104 4.3.2. Ensemble Prediction 109 4.3.3. Limitations of this study and future works 111 4.4. Conclusion 111 5. Summary and conclusion 114 6. References 118 7. Abstract (in Korean) 140박

    On pattern recognition of brain connectivity in resting-state functional MRI

    Get PDF
    Dissertação de mestrado integrado em Biomedical Engineering (specialization on Medical Informatics)The human urge and pursuit for information have led to the development of increasingly complex technologies, and new means to study and understand the most advanced and intricate biological system: the human brain. Large-scale neuronal communication within the brain, and how it relates to human behaviour can be inferred by delving into the brain network, and searching for patterns in connectivity. Functional connectivity is a steady characteristic of the brain, and it has been proved to be very useful for examining how mental disorders affect connections within the brain. The detection of abnormal behaviour in brain networks is performed by experts, such as physicians, who limit the process with human subjectivity, and unwittingly introduce errors in the interpretation. The continuous search for alternatives to obtain faster and robuster results have put Machine Learning and Deep Learning in the leading position of computer vision, as they enable the extraction of meaningful patterns, some beyond human perception. The aim of this dissertation is to design and develop an experiment setup to analyse functional connectivity at a voxel level, in order to find functional patterns. For the purpose, a pipeline was outlined to include steps from data download to data analysis, resulting in four methods: Data Download, Data Preprocessing, Dimensionality Reduction, and Analysis. The proposed experiment setup was modeled using as materials resting state fMRI data from two sources: Life and Health Sciences Research Institute (Portugal), and Human Connectome Project (USA). To evaluate its performance, a case study was performed using the In-House data for concerning a smaller number of subjects to study. The pipeline was successful at delivering results, although limitations concerning the memory of the machine used restricted some aspects of this experiment setup’s testing. With appropriate resources, this experiment setup may support the process of analysing and extracting patterns from any resting state functional connectivity data, and aid in the detection of mental disorders.O desejo e a busca intensos do ser humano por informação levaram ao desenvolvimento de tecnologias cada vez mais complexas e novos meios para estudar e entender o sistema biológico mais avançado e intrincado: o cérebro humano. A comunicação neuronal em larga escala no cérebro, e como ela se relaciona com o comportamento humano, pode ser inferida investigando a rede neuronal cerebral e procurando por padrões de conectividade. A conectividade funcional é uma característica constante do cérebro e provou ser muito útil para examinar como os distúrbios mentais afetam as conexões cerebrais. A deteção de anormalidades em imagens de ressonância magnética é realizada por especialistas, como médicos, que limitam o processo com a subjetividade humana e, inadvertidamente, introduzem erros na interpretação. A busca contínua de alternativas para obter resultados mais rápidos e robustos colocou as técnicas de machine learning e deep learning na posição de liderança de visão computacional, pois permitem a extração de padrões significativos e alguns deles para além da percepção humana. O objetivo desta dissertação é projetar e desenvolver uma configuração experimental para analisar a conectividade funcional ao nível do voxel, a fim de encontrar padrões funcionais. Nesse sentido, foi delineado um pipeline para incluir etapas a começar no download de dados até à análise desses mesmos dados, resultando assim em quatro métodos: Download de Dados, Pré-processamento de Dados, Redução de Dimensionalidade e Análise. A configuração experimental proposta foi modelada usando dados de ressonância magnética funcional de resting-state de duas fontes: Instituto de Ciências da Vida e Saúde (Portugal) e Human Connectome Project (EUA). Para avaliar o seu desempenho, foi realizado um estudo de caso usando os dados internos por considerar um número menor de participantes a serem estudados. O pipeline foi bem-sucedido em fornecer resultados, embora limitações relacionadas com a memória da máquina usada tenham restringido alguns aspetos do teste desta configuração experimental. Com recursos apropriados, esta configuração experimental poderá servir de suporte para o processo de análise e extração de padrões de qualquer conjunto de dados de conectividade funcional em resting-state e auxiliar na deteção de transtornos mentais

    A CNN-LSTM for predicting mortality in the ICU

    Get PDF
    An accurate predicted mortality is crucial to healthcare as it provides an empirical risk estimate for prognostic decision making, patient stratification and hospital benchmarking. Current prediction methods in practice are severity of disease scoring systems that usually involve a fixed set of admission attributes and summarized physiological data. These systems are prone to bias and require substantial manual effort which necessitates an updated approach which can account for most shortcomings. Clinical observation notes allow for recording highly subjective data on the patient that can possibly facilitate higher discrimination. Moreover, deep learning models can automatically extract and select features without human input.This thesis investigates the potential of a combination of a deep learning model and notes for predicting mortality with a higher accuracy. A custom architecture, called CNN-LSTM, is conceptualized for mapping multiple notes compiled in a hospital stay to a mortality outcome. It employs both convolutional and recurrent layers with the former capturing semantic relationships in individual notes independently and the latter capturing temporal relationships between concurrent notes in a hospital stay. This approach is compared to three severity of disease scoring systems with a case study on the MIMIC-III dataset. Experiments are set up to assess the CNN-LSTM for predicting mortality using only the notes from the first 24, 12 and 48 hours of a patient stay. The model is trained using K-fold cross-validation with k=5 and the mortality probability calculated by the three severity scores on the held-out set is used as the baseline. It is found that the CNN-LSTM outperforms the baseline on all experiments which serves as a proof-of-concept of how notes and deep learning can better outcome prediction
    corecore