276 research outputs found

    Contribuitions and developments on nonintrusive load monitoring

    Get PDF
    Energy efficiency is a key subject in our present world agenda, not only because of greenhouse gas emissions, which contribute to global warming, but also because of possible supply interruptions. In Brazil, energy wastage in the residential market is estimated to be around 15%. Previous studies have indicated that the most savings were achieved with specific appliance, electricity consumption feedback, which caused behavioral changes and encouraged consumers to pursue energy conservation. Nonintrusive Load Monitoring (NILM) is a relatively new term. It aims to disaggregate global consumption at an appliance level, using only a single point of measurement. Various methods have been suggested to infer when appliances are turned on and off, using the analysis of current and voltage aggregated waveforms. Within this context, we aim to provide a methodology for NILM to determine which sets of electrical features and feature extraction rates, obtained from aggregated household data, are essential to preserve equivalent levels of accuracy; thus reducing the amount of data that needs to be transferred to, and stored on, cloud servers. As an addendum to this thesis, a Brazilian appliance dataset, sampled from real appliances, was developed for future NILM developments and research. Beyond that, a low-cost NILM smart meter was developed to encourage consumers to change their habits to more sustainable methods.Eficiência energética é um assunto essencial na agenda mundial. No Brasil, o desperdício de energia no setor residencial é estimado em 15%. Estudos indicaram que maiores ganhos em eficiência são conseguidos quando o usuário recebe as informações de consumo detalhadas por cada aparelho, provocando mudanças comportamentais e incentivando os consumidores na conservação de energia. Monitoramento não intrusivo de cargas (NILM da sigla em inglês) é um termo relativamente novo. A sua finalidade é inferir o consumo de um ambiente até observar os consumos individualizados de cada equipamento utilizando-se de apenas um único ponto de medição. Métodos sofisticados têm sido propostos para inferir quando os aparelhos são ligados e desligados em um ambiente. Dentro deste contexto, este trabalho apresenta uma metodologia para a definição de um conjunto mínimo de características elétricas e sua taxa de extração que reduz a quantidade de dados a serem transmitidos e armazenados em servidores de processamento de dados, preservando níveis equivalentes de acurácia. São utilizadas diferentes técnicas de aprendizado de máquina visando à caracterização e solução do problema. Como adendo ao trabalho, apresenta-se um banco de dados de eletrodomésticos brasileiros, com amostras de equipamentos nacionais para desenvolvimentos futuros em NILM, além de um medidor inteligente de baixo custo para desagregação de cargas, visando tornar o consumo de energia mais sustentável

    Taming Data Quality in AI-Enabled Industrial Internet of Things

    Get PDF
    We address the problem of taming data quality in artificial intelligence (AI)-enabled Industrial Internet of Things systems by devising machine learning pipelines as part of a decentralized edge-to-cloud architecture. We present the design and deployment of our approach from an AI engineering perspective using two industrial case studies.acceptedVersio

    SAR: Learning Cross-Language API Mappings with Little Knowledge

    Get PDF
    To save effort, developers often translate programs from one programming language to another, instead of implementing it from scratch. Translating application program interfaces (APIs) used in one language to functionally equivalent ones available in another language is an important aspect of program translation. Existing approaches facilitate the translation by automatically identifying the API mappings across programming languages. However, these approaches still require a large number of parallel corpora, ranging from pairs of APIs or code fragments that are functionally equivalent, to similar code comments. To minimize the need for parallel corpora, this paper aims at an automated approach that can map APIs across languages with much less a priori knowledge than other approaches. Our approach is based on a realization of the notion of domain adaption, combined with code embedding, to better align two vector spaces. Taking as input large sets of programs, our approach first generates numeric vector representations of the programs (including the APIs used in each language), and it adapts generative adversarial networks (GAN) to align the vectors in different spaces of two languages. For better alignment, we initialize the GAN with parameters derived from API mapping seeds that can be identified accurately with a simple automatic signature-based matching heuristic. Then the cross-language API mappings can be identified via nearest-neighbors queries in the aligned vector spaces. We have implemented the approach (SAR, named after three main technical components, Seeding, Adversarial training, and Refinement) in a prototype for mapping APIs across Java and C# programs. Our evaluation on about 2 million Java files and 1 million C# files shows that the approach can achieve 48% and 78% mapping accuracy in its top-1 and top-10 API mapping results respectively, with only 174 automatically identified seeds, which is more accurate than other approaches using the same or much more mapping seeds

    Towards A Measure Of General Machine Intelligence

    Full text link
    To build general-purpose artificial intelligence systems that can deal with unknown variables across unknown domains, we need benchmarks that measure how well these systems perform on tasks they have never seen before. A prerequisite for this is a measure of a task's generalization difficulty, or how dissimilar it is from the system's prior knowledge and experience. If the skill of an intelligence system in a particular domain is defined as it's ability to consistently generate a set of instructions (or programs) to solve tasks in that domain, current benchmarks do not quantitatively measure the efficiency of acquiring new skills, making it possible to brute-force skill acquisition by training with unlimited amounts of data and compute power. With this in mind, we first propose a common language of instruction, a programming language that allows the expression of programs in the form of directed acyclic graphs across a wide variety of real-world domains and computing platforms. Using programs generated in this language, we demonstrate a match-based method to both score performance and calculate the generalization difficulty of any given set of tasks. We use these to define a numeric benchmark called the generalization index, or the g-index, to measure and compare the skill-acquisition efficiency of any intelligence system on a set of real-world tasks. Finally, we evaluate the suitability of some well-known models as general intelligence systems by calculating their g-index scores.Comment: 31 pages, 15 Figures, 3 Tables; Sample Data and g-index Reference Code at https://github.com/mayahq/g-index-benchmark; g-index toy environment at https://github.com/mayahq/flatland; version 2 added a section about the toy environment; version 3 compressed images to reduce file size; version 4 updated description of flatland toy environmen

    One Deep Music Representation to Rule Them All? : A comparative analysis of different representation learning strategies

    Full text link
    Inspired by the success of deploying deep learning in the fields of Computer Vision and Natural Language Processing, this learning paradigm has also found its way into the field of Music Information Retrieval. In order to benefit from deep learning in an effective, but also efficient manner, deep transfer learning has become a common approach. In this approach, it is possible to reuse the output of a pre-trained neural network as the basis for a new learning task. The underlying hypothesis is that if the initial and new learning tasks show commonalities and are applied to the same type of input data (e.g. music audio), the generated deep representation of the data is also informative for the new task. Since, however, most of the networks used to generate deep representations are trained using a single initial learning source, their representation is unlikely to be informative for all possible future tasks. In this paper, we present the results of our investigation of what are the most important factors to generate deep representations for the data and learning tasks in the music domain. We conducted this investigation via an extensive empirical study that involves multiple learning sources, as well as multiple deep learning architectures with varying levels of information sharing between sources, in order to learn music representations. We then validate these representations considering multiple target datasets for evaluation. The results of our experiments yield several insights on how to approach the design of methods for learning widely deployable deep data representations in the music domain.Comment: This work has been accepted to "Neural Computing and Applications: Special Issue on Deep Learning for Music and Audio

    ECG classification and prognostic approach towards personalized healthcare

    Get PDF
    A very important aspect of personalized healthcare is to continuously monitor an individual’s health using wearable biomedical devices and essentially to analyse and if possible to predict potential health hazards that may prove fatal if not treated in time. The prediction aspect embedded in the system helps in avoiding delays in providing timely medical treatment, even before an individual reaches a critical condition. Despite of the availability of modern wearable health monitoring devices, the real-time analyses and prediction component seems to be missing with these devices. The research work illustrated in this paper, at an outset, focussed on constantly monitoring an individual's ECG readings using a wearable 3-lead ECG kit and more importantly focussed on performing real-time analyses to detect arrhythmia to be able to identify and predict heart risk. Also, current research shows extensive use of heart rate variability (HRV) analysis and machine learning for arrhythmia classification, which however depends on the morphology of the ECG waveforms and the sensitivity of the ECG equipment. Since a wearable 3-lead ECG kit was used, the accuracy of classification had to be dealt with at the machine learning phase, so a unique feature extraction method was developed to increase the accuracy of classification. As a case study a very widely used Arrhythmia database (MIT-BIH, Physionet) was used to develop learning, classification and prediction models. Neuralnet fitting models on the extracted features showed mean-squared error of as low as 0.0085 and regression value as high as 0.99. Current experiments show 99.4% accuracy using k-NN Classification models and show values of Cross-Entropy Error of 7.6 and misclassification error value of 1.2 on test data using scaled conjugate gradient pattern matching algorithms. Software components were developed for wearable devices that took ECG readings from a 3-Lead ECG data acquisition kit in real time, de-noised, filtered and relayed the sample readings to the tele health analytical server. The analytical server performed the classification and prediction tasks based on the trained classification models and could raise appropriate alarms if ECG abnormalities of V (Premature Ventricular Contraction: PVC), A (Atrial Premature Beat: APB), L (Left bundle branch block beat), R (Right bundle branch block beat) type annotations in MITDB were detected. The instruments were networked using IoT (Internet of Things) devices and abnormal ECG events related to arrhythmia, from analytical server could be logged using an FHIR web service implementation, according to a SNOMED coding system and could be accessed in Electronic Health Record by the concerned medic to take appropriate and timely decisions. The system focused on ‘preventive care rather than remedial cure’ which has become a major focus of all the health care and cure institutions across the globe
    • …
    corecore