11 research outputs found

    Independent subspace analysis can cope with the 'curse of dimensionality'

    Get PDF
    We search for hidden independent components, in particular we consider the independent subspace analysis (ISA) task. Earlier ISA procedures assume that the dimensions of the components are known. Here we show a method that enables the non-combinatorial estimation of the components. We make use of a decomposition principle called the ISA separation theorem. According to this separation theorem the ISA task can be reduced to the independent component analysis (ICA) task that assumes one-dimensional components and then to a grouping procedure that collects the respective non-independent elements into independent groups. We show that non-combinatorial grouping is feasible by means of the non-linear f-correlation matrices between the estimated components

    Forecasting with Machine Learning

    Get PDF
    For years, people have been forecasting weather patterns, economic and political events, sports outcomes, and more. In this paper we discussed the ways of using machine learning in forecasting, machine learning is a branch of computer science where algorithms learn from data. The fundamental problem for machine learning and time series is the same: to predict new outcomes based on previously known results. Using the suitable technique of machine learning depend on how much data you have, how noisy the data is, and what kind of new features can be derived from the data. But these techniques can improve accuracy and don’t have to be difficult to implement

    New challenges in wireless and free space optical communications

    Get PDF
    AbstractThis manuscript presents a survey on new challenges in wireless communication systems and discusses recent approaches to address some recently raised problems by the wireless community. At first a historical background is briefly introduced. Challenges based on modern and real life applications are then described. Up to date research fields to solve limitations of existing systems and emerging new technologies are discussed. Theoretical and experimental results based on several research projects or studies are briefly provided. Essential, basic and many self references are cited. Future researcher axes are briefly introduced

    The chronotron: a neuron that learns to fire temporally-precise spike patterns

    Get PDF
    In many cases, neurons process information carried by the precise timing of spikes. Here we show how neurons can learn to generate specific temporally-precise output spikes in response to input spike patterns, thus processing and memorizing information that is fully temporally coded, both as input and as output. We introduce two new supervised learning rules for spiking neurons with temporal coding of information (chronotrons), one that is analytically-derived and highly efficient, and one that has a high degree of biological plausibility. We show how chronotrons can learn to classify their inputs and we study their memory capacity

    Influencing factors in energy use of housing blocks: a new methodology, based on clustering and energy simulations, for decision making in energy refurbishment projects

    Get PDF
    In recent years, big efforts have been dedicated to identify which are the factors with highest influence in the energy consumption of residential buildings. These factors include aspects such as weather dependence, user behaviour, socio-economic situation, type of the energy installations and typology of buildings. The high number of factors increases the complexity of analysis and leads to a lack of confidence in the results of the energy simulation analysis. This fact grows when we move one step up and perform global analysis of blocks of buildings. The aim of this study is to report a new methodology for the assessment of the energy performance of large groups of buildings when considering the real use of energy. We combine two clustering methods, Generative Topographic Mapping and k-means, to obtain reference dwellings that can be considered as representative of the different energy patterns and energy systems of the neighbourhood. Then, simulation of energy demand and indoor temperature against the monitored comfort conditions in a short period is performed to obtain end use load disaggregation. This methodology was applied in a district at Terrassa City (Spain), and six reference dwellings were selected. Results showed that the method was able to identify the main patterns and provide occupants with feasible recommendations so that they can make required decisions at neighbourhood level. Moreover, given that the proposed method is based on the comparison with similar buildings, it could motivate building occupants to implement community improvement actions, as well as to modify their behaviour

    Influencing factors in energy use of housing blocks: a new methodology, based on clustering and energy simulations, for decision making in energy refurbishment projects

    Get PDF
    In recent years, big efforts have been dedicated to identify which are the factors with highest influence in the energy consumption of residential buildings. These factors include aspects such as weather dependence, user behaviour, socio-economic situation, type of the energy installations and typology of buildings. The high number of factors increases the complexity of analysis and leads to a lack of confidence in the results of the energy simulation analysis. This fact grows when we move one step up and perform global analysis of blocks of buildings. The aim of this study is to report a new methodology for the assessment of the energy performance of large groups of buildings when considering the real use of energy. We combine two clustering methods, Generative Topographic Mapping and k-means, to obtain reference dwellings that can be considered as representative of the different energy patterns and energy systems of the neighbourhood. Then, simulation of energy demand and indoor temperature against the monitored comfort conditions in a short period is performed to obtain end use load disaggregation. This methodology was applied in a district at Terrassa City (Spain), and six reference dwellings were selected. Results showed that the method was able to identify the main patterns and provide occupants with feasible recommendations so that they can make required decisions at neighbourhood level. Moreover, given that the proposed method is based on the comparison with similar buildings, it could motivate building occupants to implement community improvement actions, as well as to modify their behaviour.Peer ReviewedPostprint (author's final draft

    Learning the Structure of High-Dimensional Manifolds with Self-Organizing Maps for Accurate Information Extraction

    Get PDF
    This paper was submitted by the author prior to final official version. For official version please see http://hdl.handle.net/1911/70515This work aims to improve the capability of accurate information extraction from high-dimensional data, with a specific neural learning paradigm, the Self-Organizing Map (SOM). The SOM is an unsupervised learning algorithm that can faithfully sense the manifold structure and support supervised learning of relevant information from the data. Yet open problems regarding SOM learning exist. We focus on the following two issues. 1. Evaluation of topology preservation. Topology preservation is essential for SOMs in faithful representation of manifold structure. However, in reality, topology violations are not unusual, especially when the data have complicated structure. Measures capable of accurately quantifying and informatively expressing topology violations are lacking. One contribution of this work is a new measure, the Weighted Differential Topographic Function (WDTF), which differentiates an existing measure, the Topographic Function (TF), and incorporates detailed data distribution as an importance weighting of violations to distinguish severe violations from insignificant ones. Another contribution is an interactive visual tool, TopoView, which facilitates the visual inspection of violations on the SOM lattice. We show the effectiveness of the combined use of the WDTF and TopoView through a simple two-dimensional data set and two hyperspectral images. 2. Learning multiple latent variables from high-dimensional data. We use an existing two-layer SOM-hybrid supervised architecture, which captures the manifold structure in its SOM hidden layer, and then, uses its output layer to perform the supervised learning of latent variables. In the customary way, the output layer only uses the strongest output of the SOM neurons. This severely limits the learning capability. We allow multiple, k, strongest responses of the SOM neurons for the supervised learning. Moreover, the fact that different latent variables can be best learned with different values of k motivates a new neural architecture, the Conjoined Twins, which extends the existing architecture with additional copies of the output layer, for preferential use of different values of k in the learning of different latent variables. We also automate the customization of k for different variables with the statistics derived from the SOM. The Conjoined Twins shows its effectiveness in the inference of two physical parameters from Near-Infrared spectra of planetary ices

    Cartogram representations of self-organizing virtual geographies

    Get PDF
    Model interpretability is a problem for multivariate data in general and, very specifically, for dimensionality reduction techniques as applied to data visualization. The problem is even bigger for nonlinear dimensionality reduction (NLDR) methods, to which interpretability limitations are consubstantial. Data visualization is a key process for knowledge extraction from data that helps us to gain insights into the observed data structure through graphical representations and metaphors. NLDR techniques provide flexible visual insight, but the locally varying representation distor- tion they generate makes interpretation far from intuitive. For some NLDR models, indirect quantitative measures of this mapping distortion can be calculated explicitly and used as part of an interpretative post-processing of the results. In this Master Thesis, we apply a cartogram method, inspired on techniques of geographic representation, to the purpose of data visualization using NLDR models. In particular, we show how this method allows reintroducing the distortion, measured in the visual maps of several self-organizing clustering methods. The main capabilities and limitations of the cartogram visualization of multivariate data using standard and hierarchical self-organizing models were investigated in some detail with artificial data as well as with real information stemming from a neuro-oncology problem that involves the discrimination of human brain tumor types, a problem for which knowledge dis- covery techniques in general, and data visualization in particular should be useful tools

    A soft computing decision support framework for e-learning

    Get PDF
    Tesi per compendi de publicacions.Supported by technological development and its impact on everyday activities, e-Learning and b-Learning (Blended Learning) have experienced rapid growth mainly in higher education and training. Its inherent ability to break both physical and cultural distances, to disseminate knowledge and decrease the costs of the teaching-learning process allows it to reach anywhere and anyone. The educational community is divided as to its role in the future. It is believed that by 2019 half of the world's higher education courses will be delivered through e-Learning. While supporters say that this will be the educational mode of the future, its detractors point out that it is a fashion, that there are huge rates of abandonment and that their massification and potential low quality, will cause its fall, assigning it a major role of accompanying traditional education. There are, however, two interrelated features where there seems to be consensus. On the one hand, the enormous amount of information and evidence that Learning Management Systems (LMS) generate during the e-Learning process and which is the basis of the part of the process that can be automated. In contrast, there is the fundamental role of e-tutors and etrainers who are guarantors of educational quality. These are continually overwhelmed by the need to provide timely and effective feedback to students, manage endless particular situations and casuistics that require decision making and process stored information. In this sense, the tools that e-Learning platforms currently provide to obtain reports and a certain level of follow-up are not sufficient or too adequate. It is in this point of convergence Information-Trainer, where the current developments of the LMS are centered and it is here where the proposed thesis tries to innovate. This research proposes and develops a platform focused on decision support in e-Learning environments. Using soft computing and data mining techniques, it extracts knowledge from the data produced and stored by e-Learning systems, allowing the classification, analysis and generalization of the extracted knowledge. It includes tools to identify models of students' learning behavior and, from them, predict their future performance and enable trainers to provide adequate feedback. Likewise, students can self-assess, avoid those ineffective behavior patterns, and obtain real clues about how to improve their performance in the course, through appropriate routes and strategies based on the behavioral model of successful students. The methodological basis of the mentioned functionalities is the Fuzzy Inductive Reasoning (FIR), which is particularly useful in the modeling of dynamic systems. During the development of the research, the FIR methodology has been improved and empowered by the inclusion of several algorithms. First, an algorithm called CR-FIR, which allows determining the Causal Relevance that have the variables involved in the modeling of learning and assessment of students. In the present thesis, CR-FIR has been tested on a comprehensive set of classical test data, as well as real data sets, belonging to different areas of knowledge. Secondly, the detection of atypical behaviors in virtual campuses was approached using the Generative Topographic Mapping (GTM) methodology, which is a probabilistic alternative to the well-known Self-Organizing Maps. GTM was used simultaneously for clustering, visualization and detection of atypical data. The core of the platform has been the development of an algorithm for extracting linguistic rules in a language understandable to educational experts, which helps them to obtain patterns of student learning behavior. In order to achieve this functionality, the LR-FIR algorithm (Extraction of Linguistic Rules in FIR) was designed and developed as an extension of FIR that allows both to characterize general behavior and to identify interesting patterns. In the case of the application of the platform to several real e-Learning courses, the results obtained demonstrate its feasibility and originality. The teachers' perception about the usability of the tool is very good, and they consider that it could be a valuable resource to mitigate the time requirements of the trainer that the e-Learning courses demand. The identification of student behavior models and prediction processes have been validated as to their usefulness by expert trainers. LR-FIR has been applied and evaluated in a wide set of real problems, not all of them in the educational field, obtaining good results. The structure of the platform makes it possible to assume that its use is potentially valuable in those domains where knowledge management plays a preponderant role, or where decision-making processes are a key element, e.g. ebusiness, e-marketing, customer management, to mention just a few. The Soft Computing tools used and developed in this research: FIR, CR-FIR, LR-FIR and GTM, have been applied successfully in other real domains, such as music, medicine, weather behaviors, etc.Soportado por el desarrollo tecnolĂłgico y su impacto en las diferentes actividades cotidianas, el e-Learning (o aprendizaje electrĂłnico) y el b-Learning (Blended Learning o aprendizaje mixto), han experimentado un crecimiento vertiginoso principalmente en la educaciĂłn superior y la capacitaciĂłn. Su habilidad inherente para romper distancias tanto fĂ­sicas como culturales, para diseminar conocimiento y disminuir los costes del proceso enseñanza aprendizaje le permite llegar a cualquier sitio y a cualquier persona. La comunidad educativa se encuentra dividida en cuanto a su papel en el futuro. Se cree que para el año 2019 la mitad de los cursos de educaciĂłn superior del mundo se impartirĂĄ a travĂ©s del e-Learning. Mientras que los partidarios aseguran que Ă©sta serĂĄ la modalidad educativa del futuro, sus detractores señalan que es una moda, que hay enormes Ă­ndices de abandono y que su masificaciĂłn y potencial baja calidad, provocarĂĄ su caĂ­da, reservĂĄndole un importante papel de acompañamiento a la educaciĂłn tradicional. Hay, sin embargo, dos caracterĂ­sticas interrelacionadas donde parece haber consenso. Por un lado, la enorme generaciĂłn de informaciĂłn y evidencias que los sistemas de gestiĂłn del aprendizaje o LMS (Learning Management System) generan durante el proceso educativo electrĂłnico y que son la base de la parte del proceso que se puede automatizar. En contraste, estĂĄ el papel fundamental de los e-tutores y e-formadores que son los garantes de la calidad educativa. Éstos se ven continuamente desbordados por la necesidad de proporcionar retroalimentaciĂłn oportuna y eficaz a los alumnos, gestionar un sin fin de situaciones particulares y casuĂ­sticas que requieren toma de decisiones y procesar la informaciĂłn almacenada. En este sentido, las herramientas que las plataformas de e-Learning proporcionan actualmente para obtener reportes y cierto nivel de seguimiento no son suficientes ni demasiado adecuadas. Es en este punto de convergencia InformaciĂłn-Formador, donde estĂĄn centrados los actuales desarrollos de los LMS y es aquĂ­ donde la tesis que se propone pretende innovar. La presente investigaciĂłn propone y desarrolla una plataforma enfocada al apoyo en la toma de decisiones en ambientes e-Learning. Utilizando tĂ©cnicas de Soft Computing y de minerĂ­a de datos, extrae conocimiento de los datos producidos y almacenados por los sistemas e-Learning permitiendo clasificar, analizar y generalizar el conocimiento extraĂ­do. Incluye herramientas para identificar modelos del comportamiento de aprendizaje de los estudiantes y, a partir de ellos, predecir su desempeño futuro y permitir a los formadores proporcionar una retroalimentaciĂłn adecuada. AsĂ­ mismo, los estudiantes pueden autoevaluarse, evitar aquellos patrones de comportamiento poco efectivos y obtener pistas reales acerca de cĂłmo mejorar su desempeño en el curso, mediante rutas y estrategias adecuadas a partir del modelo de comportamiento de los estudiantes exitosos. La base metodolĂłgica de las funcionalidades mencionadas es el Razonamiento Inductivo Difuso (FIR, por sus siglas en inglĂ©s), que es particularmente Ăștil en el modelado de sistemas dinĂĄmicos. Durante el desarrollo de la investigaciĂłn, la metodologĂ­a FIR ha sido mejorada y potenciada mediante la inclusiĂłn de varios algoritmos. En primer lugar un algoritmo denominado CR-FIR, que permite determinar la Relevancia Causal que tienen las variables involucradas en el modelado del aprendizaje y la evaluaciĂłn de los estudiantes. En la presente tesis, CR-FIR se ha probado en un conjunto amplio de datos de prueba clĂĄsicos, asĂ­ como conjuntos de datos reales, pertenecientes a diferentes ĂĄreas de conocimiento. En segundo lugar, la detecciĂłn de comportamientos atĂ­picos en campus virtuales se abordĂł mediante el enfoque de Mapeo TopogrĂĄfico Generativo (GTM), que es una alternativa probabilĂ­stica a los bien conocidos Mapas Auto-organizativos. GTM se utilizĂł simultĂĄneamente para agrupamiento, visualizaciĂłn y detecciĂłn de datos atĂ­picos. La parte medular de la plataforma ha sido el desarrollo de un algoritmo de extracciĂłn de reglas lingĂŒĂ­sticas en un lenguaje entendible para los expertos educativos, que les ayude a obtener los patrones del comportamiento de aprendizaje de los estudiantes. Para lograr dicha funcionalidad, se diseñó y desarrollĂł el algoritmo LR-FIR, (extracciĂłn de Reglas LingĂŒĂ­sticas en FIR, por sus siglas en inglĂ©s) como una extensiĂłn de FIR que permite tanto caracterizar el comportamiento general, como identificar patrones interesantes. En el caso de la aplicaciĂłn de la plataforma a varios cursos e-Learning reales, los resultados obtenidos demuestran su factibilidad y originalidad. La percepciĂłn de los profesores acerca de la usabilidad de la herramienta es muy buena, y consideran que podrĂ­a ser un valioso recurso para mitigar los requerimientos de tiempo del formador que los cursos e-Learning exigen. La identificaciĂłn de los modelos de comportamiento de los estudiantes y los procesos de predicciĂłn han sido validados en cuanto a su utilidad por los formadores expertos. LR-FIR se ha aplicado y evaluado en un amplio conjunto de problemas reales, no todos ellos del ĂĄmbito educativo, obteniendo buenos resultados. La estructura de la plataforma permite suponer que su utilizaciĂłn es potencialmente valiosa en aquellos dominios donde la administraciĂłn del conocimiento juegue un papel preponderante, o donde los procesos de toma de decisiones sean una pieza clave, por ejemplo, e-business, e-marketing, administraciĂłn de clientes, por mencionar sĂłlo algunos. Las herramientas de Soft Computing utilizadas y desarrolladas en esta investigaciĂłn: FIR, CR-FIR, LR-FIR y GTM, ha sido aplicadas con Ă©xito en otros dominios reales, como mĂșsica, medicina, comportamientos climĂĄticos, etc.Postprint (published version
    corecore