1,298 research outputs found

    How Quantum Computers Fail: Quantum Codes, Correlations in Physical Systems, and Noise Accumulation

    Full text link
    The feasibility of computationally superior quantum computers is one of the most exciting and clear-cut scientific questions of our time. The question touches on fundamental issues regarding probability, physics, and computability, as well as on exciting problems in experimental physics, engineering, computer science, and mathematics. We propose three related directions towards a negative answer. The first is a conjecture about physical realizations of quantum codes, the second has to do with correlations in stochastic physical systems, and the third proposes a model for quantum evolutions when noise accumulates. The paper is dedicated to the memory of Itamar Pitowsky.Comment: 16 page

    A foundation for machine learning in design

    Get PDF
    This paper presents a formalism for considering the issues of learning in design. A foundation for machine learning in design (MLinD) is defined so as to provide answers to basic questions on learning in design, such as, "What types of knowledge can be learnt?", "How does learning occur?", and "When does learning occur?". Five main elements of MLinD are presented as the input knowledge, knowledge transformers, output knowledge, goals/reasons for learning, and learning triggers. Using this foundation, published systems in MLinD were reviewed. The systematic review presents a basis for validating the presented foundation. The paper concludes that there is considerable work to be carried out in order to fully formalize the foundation of MLinD

    Model-free reconstruction of neuronal network connectivity from calcium imaging signals

    Get PDF
    A systematic assessment of global neural network connectivity through direct electrophysiological assays has remained technically unfeasible even in dissociated neuronal cultures. We introduce an improved algorithmic approach based on Transfer Entropy to reconstruct approximations to network structural connectivities from network activity monitored through calcium fluorescence imaging. Based on information theory, our method requires no prior assumptions on the statistics of neuronal firing and neuronal connections. The performance of our algorithm is benchmarked on surrogate time-series of calcium fluorescence generated by the simulated dynamics of a network with known ground-truth topology. We find that the effective network topology revealed by Transfer Entropy depends qualitatively on the time-dependent dynamic state of the network (e.g., bursting or non-bursting). We thus demonstrate how conditioning with respect to the global mean activity improves the performance of our method. [...] Compared to other reconstruction strategies such as cross-correlation or Granger Causality methods, our method based on improved Transfer Entropy is remarkably more accurate. In particular, it provides a good reconstruction of the network clustering coefficient, allowing to discriminate between weakly or strongly clustered topologies, whereas on the other hand an approach based on cross-correlations would invariantly detect artificially high levels of clustering. Finally, we present the applicability of our method to real recordings of in vitro cortical cultures. We demonstrate that these networks are characterized by an elevated level of clustering compared to a random graph (although not extreme) and by a markedly non-local connectivity.Comment: 54 pages, 8 figures (+9 supplementary figures), 1 table; submitted for publicatio

    Importance sampling the union of rare events with an application to power systems analysis

    Full text link
    We consider importance sampling to estimate the probability μ\mu of a union of JJ rare events HjH_j defined by a random variable x\boldsymbol{x}. The sampler we study has been used in spatial statistics, genomics and combinatorics going back at least to Karp and Luby (1983). It works by sampling one event at random, then sampling x\boldsymbol{x} conditionally on that event happening and it constructs an unbiased estimate of μ\mu by multiplying an inverse moment of the number of occuring events by the union bound. We prove some variance bounds for this sampler. For a sample size of nn, it has a variance no larger than μ(μˉμ)/n\mu(\bar\mu-\mu)/n where μˉ\bar\mu is the union bound. It also has a coefficient of variation no larger than (J+J12)/(4n)\sqrt{(J+J^{-1}-2)/(4n)} regardless of the overlap pattern among the JJ events. Our motivating problem comes from power system reliability, where the phase differences between connected nodes have a joint Gaussian distribution and the JJ rare events arise from unacceptably large phase differences. In the grid reliability problems even some events defined by 57725772 constraints in 326326 dimensions, with probability below 102210^{-22}, are estimated with a coefficient of variation of about 0.00240.0024 with only n=10,000n=10{,}000 sample values

    Domain Generalization in Machine Learning Models for Wireless Communications: Concepts, State-of-the-Art, and Open Issues

    Full text link
    Data-driven machine learning (ML) is promoted as one potential technology to be used in next-generations wireless systems. This led to a large body of research work that applies ML techniques to solve problems in different layers of the wireless transmission link. However, most of these applications rely on supervised learning which assumes that the source (training) and target (test) data are independent and identically distributed (i.i.d). This assumption is often violated in the real world due to domain or distribution shifts between the source and the target data. Thus, it is important to ensure that these algorithms generalize to out-of-distribution (OOD) data. In this context, domain generalization (DG) tackles the OOD-related issues by learning models on different and distinct source domains/datasets with generalization capabilities to unseen new domains without additional finetuning. Motivated by the importance of DG requirements for wireless applications, we present a comprehensive overview of the recent developments in DG and the different sources of domain shift. We also summarize the existing DG methods and review their applications in selected wireless communication problems, and conclude with insights and open questions

    Applications of biased-randomized algorithms and simheuristics in integrated logistics

    Get PDF
    Transportation and logistics (T&L) activities play a vital role in the development of many businesses from different industries. With the increasing number of people living in urban areas, the expansion of on-demand economy and e-commerce activities, the number of services from transportation and delivery has considerably increased. Consequently, several urban problems have been potentialized, such as traffic congestion and pollution. Several related problems can be formulated as a combinatorial optimization problem (COP). Since most of them are NP-Hard, the finding of optimal solutions through exact solution methods is often impractical in a reasonable amount of time. In realistic settings, the increasing need for 'instant' decision-making further refutes their use in real life. Under these circumstances, this thesis aims at: (i) identifying realistic COPs from different industries; (ii) developing different classes of approximate solution approaches to solve the identified T&L problems; (iii) conducting a series of computational experiments to validate and measure the performance of the developed approaches. The novel concept of 'agile optimization' is introduced, which refers to the combination of biased-randomized heuristics with parallel computing to deal with real-time decision-making.Las actividades de transporte y logística (T&L) juegan un papel vital en el desarrollo de muchas empresas de diferentes industrias. Con el creciente número de personas que viven en áreas urbanas, la expansión de la economía a lacarta y las actividades de comercio electrónico, el número de servicios de transporte y entrega ha aumentado considerablemente. En consecuencia, se han potencializado varios problemas urbanos, como la congestión del tráfico y la contaminación. Varios problemas relacionados pueden formularse como un problema de optimización combinatoria (COP). Dado que la mayoría de ellos son NP-Hard, la búsqueda de soluciones óptimas a través de métodos de solución exactos a menudo no es práctico en un período de tiempo razonable. En entornos realistas, la creciente necesidad de una toma de decisiones "instantánea" refuta aún más su uso en la vida real. En estas circunstancias, esta tesis tiene como objetivo: (i) identificar COP realistas de diferentes industrias; (ii) desarrollar diferentes clases de enfoques de solución aproximada para resolver los problemas de T&L identificados; (iii) realizar una serie de experimentos computacionales para validar y medir el desempeño de los enfoques desarrollados. Se introduce el nuevo concepto de optimización ágil, que se refiere a la combinación de heurísticas aleatorias sesgadas con computación paralela para hacer frente a la toma de decisiones en tiempo real.Les activitats de transport i logística (T&L) tenen un paper vital en el desenvolupament de moltes empreses de diferents indústries. Amb l'augment del nombre de persones que viuen a les zones urbanes, l'expansió de l'economia a la carta i les activitats de comerç electrònic, el nombre de serveis del transport i el lliurament ha augmentat considerablement. En conseqüència, s'han potencialitzat diversos problemes urbans, com ara la congestió del trànsit i la contaminació. Es poden formular diversos problemes relacionats com a problema d'optimització combinatòria (COP). Com que la majoria són NP-Hard, la recerca de solucions òptimes mitjançant mètodes de solució exactes sovint no és pràctica en un temps raonable. En entorns realistes, la creixent necessitat de prendre decisions "instantànies" refuta encara més el seu ús a la vida real. En aquestes circumstàncies, aquesta tesi té com a objectiu: (i) identificar COP realistes de diferents indústries; (ii) desenvolupar diferents classes d'aproximacions aproximades a la solució per resoldre els problemes identificats de T&L; (iii) la realització d'una sèrie d'experiments computacionals per validar i mesurar el rendiment dels enfocaments desenvolupats. S'introdueix el nou concepte d'optimització àgil, que fa referència a la combinació d'heurístiques esbiaixades i aleatòries amb informàtica paral·lela per fer front a la presa de decisions en temps real.Tecnologies de la informació i de xarxe
    corecore