7,309 research outputs found

    Quality-aware predictive modelling & inferential analytics at the network edge

    Get PDF
    The Internet of Things has grown by an enormous amount of devices over the later years. With the upcoming idea of the Internet of Everything the growth will be even faster. These embedded devices are connected to a central server, e.g. the Cloud. A major task is to send the generated data for further analysis and modelling to this central collection point. The devices’ network and deployed system are constrained due to energy, bandwidth, connectivity, latency, and privacy. To overcome these constraints, Edge Computing has been introduced to enable devices performing computation near the source. With the increase of embedded devices and the Internet of Things, the continuous data transmission between devices and Central Locations reached an infeasible point in which efficient communication and computational offloading are required. Edge Computing enables devices to compute lightweight algorithms locally to reduce the raw-data transmission of the network. The quality of predictive analytics tasks is of high importance as user satisfaction and decision making depend on the outcome. Therefore, this thesis investigates the ability to perform predictive analytics and model inference in Edge Devices with communication-efficient, latency-efficient, and privacy-efficient procedures by focusing on quality-aware results. The first part of the thesis focuses on reducing data transmission between the device and the central location. Two possible energy-efficient methodologies to control the data forwarding are introduced: prediction-based and time-optimised. Both data forwarding strategies aim to maintain the Central Location’s quality of analytics by introducing reconstruction policies. The second part provides a mechanism to enable edge-centric analytics towards latency-efficient network optimisation. One aspect shows the importance of locally generated analytical models in Edge Devices embracing each device’s data subspace. Furthermore, two possible ensemble-pruning methods are introduced that allow the aggregation of individual models at the Central Location towards accurate query predictions. The conclusion chapter presents the importance of privacy-efficient local learning and analytics in Edge Devices. With the aid of Federated Learning, it is possible to train analytical models for privacy-preserving data locally. Furthermore, for continuous changing environments, the parallel deployment of personalisation and generalisation for quality aware predictions is highlighted and demonstrated through experimental evaluation

    Scalable aggregation predictive analytics: a query-driven machine learning approach

    Get PDF
    We introduce a predictive modeling solution that provides high quality predictive analytics over aggregation queries in Big Data environments. Our predictive methodology is generally applicable in environments in which large-scale data owners may or may not restrict access to their data and allow only aggregation operators like COUNT to be executed over their data. In this context, our methodology is based on historical queries and their answers to accurately predict ad-hoc queries’ answers. We focus on the widely used set-cardinality, i.e., COUNT, aggregation query, as COUNT is a fundamental operator for both internal data system optimizations and for aggregation-oriented data exploration and predictive analytics. We contribute a novel, query-driven Machine Learning (ML) model whose goals are to: (i) learn the query-answer space from past issued queries, (ii) associate the query space with local linear regression & associative function estimators, (iii) define query similarity, and (iv) predict the cardinality of the answer set of unseen incoming queries, referred to the Set Cardinality Prediction (SCP) problem. Our ML model incorporates incremental ML algorithms for ensuring high quality prediction results. The significance of contribution lies in that it (i) is the only query-driven solution applicable over general Big Data environments, which include restricted-access data, (ii) offers incremental learning adjusted for arriving ad-hoc queries, which is well suited for query-driven data exploration, and (iii) offers a performance (in terms of scalability, SCP accuracy, processing time, and memory requirements) that is superior to data-centric approaches. We provide a comprehensive performance evaluation of our model evaluating its sensitivity, scalability and efficiency for quality predictive analytics. In addition, we report on the development and incorporation of our ML model in Spark showing its superior performance compared to the Spark’s COUNT method

    Predictive intelligence to the edge through approximate collaborative context reasoning

    Get PDF
    We focus on Internet of Things (IoT) environments where a network of sensing and computing devices are responsible to locally process contextual data, reason and collaboratively infer the appearance of a specific phenomenon (event). Pushing processing and knowledge inference to the edge of the IoT network allows the complexity of the event reasoning process to be distributed into many manageable pieces and to be physically located at the source of the contextual information. This enables a huge amount of rich data streams to be processed in real time that would be prohibitively complex and costly to deliver on a traditional centralized Cloud system. We propose a lightweight, energy-efficient, distributed, adaptive, multiple-context perspective event reasoning model under uncertainty on each IoT device (sensor/actuator). Each device senses and processes context data and infers events based on different local context perspectives: (i) expert knowledge on event representation, (ii) outliers inference, and (iii) deviation from locally predicted context. Such novel approximate reasoning paradigm is achieved through a contextualized, collaborative belief-driven clustering process, where clusters of devices are formed according to their belief on the presence of events. Our distributed and federated intelligence model efficiently identifies any localized abnormality on the contextual data in light of event reasoning through aggregating local degrees of belief, updates, and adjusts its knowledge to contextual data outliers and novelty detection. We provide comprehensive experimental and comparison assessment of our model over real contextual data with other localized and centralized event detection models and show the benefits stemmed from its adoption by achieving up to three orders of magnitude less energy consumption and high quality of inference

    Data and Predictive Analytics Use for Logistics and Supply Chain Management

    Get PDF
    Purpose The purpose of this paper is to explore the social process of Big Data and predictive analytics (BDPA) use for logistics and supply chain management (LSCM), focusing on interactions among technology, human behavior and organizational context that occur at the technology’s post-adoption phases in retail supply chain (RSC) organizations. Design/methodology/approach The authors follow a grounded theory approach for theory building based on interviews with senior managers of 15 organizations positioned across multiple echelons in the RSC. Findings Findings reveal how user involvement shapes BDPA to fit organizational structures and how changes made to the technology retroactively affect its design and institutional properties. Findings also reveal previously unreported aspects of BDPA use for LSCM. These include the presence of temporal and spatial discontinuities in the technology use across RSC organizations. Practical implications This study unveils that it is impossible to design a BDPA technology ready for immediate use. The emergent process framework shows that institutional and social factors require BDPA use specific to the organization, as the technology comes to reflect the properties of the organization and the wider social environment for which its designers originally intended. BDPA is, thus, not easily transferrable among collaborating RSC organizations and requires managerial attention to the institutional context within which its usage takes place. Originality/value The literature describes why organizations will use BDPA but fails to provide adequate insight into how BDPA use occurs. The authors address the “how” and bring a social perspective into a technology-centric area

    Knowledge-centric Analytics Queries Allocation in Edge Computing Environments

    Get PDF
    The Internet of Things involves a huge number of devices that collect data and deliver them to the Cloud. The processing of data at the Cloud is characterized by increased latency in providing responses to analytics queries defined by analysts or applications. Hence, Edge Computing (EC) comes into the scene to provide data processing close to the source. The collected data can be stored in edge devices and queries can be executed there to reduce latency. In this paper, we envision a case where entities located in the Cloud undertake the responsibility of receiving analytics queries and decide on the most appropriate edge nodes for queries execution. The decision is based on statistical signatures of the datasets of nodes and the statistical matching between statistics and analytics queries. Edge nodes regularly update their statistical signatures to support such decision process. Our performance evaluation shows the advantages and the shortcomings of our proposed schema in edge computing environments
    • …
    corecore