65 research outputs found

    Average-Case Complexity

    Full text link
    We survey the average-case complexity of problems in NP. We discuss various notions of good-on-average algorithms, and present completeness results due to Impagliazzo and Levin. Such completeness results establish the fact that if a certain specific (but somewhat artificial) NP problem is easy-on-average with respect to the uniform distribution, then all problems in NP are easy-on-average with respect to all samplable distributions. Applying the theory to natural distributional problems remain an outstanding open question. We review some natural distributional problems whose average-case complexity is of particular interest and that do not yet fit into this theory. A major open question whether the existence of hard-on-average problems in NP can be based on the P\neqNP assumption or on related worst-case assumptions. We review negative results showing that certain proof techniques cannot prove such a result. While the relation between worst-case and average-case complexity for general NP problems remains open, there has been progress in understanding the relation between different ``degrees'' of average-case complexity. We discuss some of these ``hardness amplification'' results

    Digital rights management for electronic documents

    Get PDF
    Ph.DDOCTOR OF PHILOSOPH

    Advanced analytics through FPGA based query processing and deep reinforcement learning

    Get PDF
    Today, vast streams of structured and unstructured data have been incorporated in databases, and analytical processes are applied to discover patterns, correlations, trends and other useful relationships that help to take part in a broad range of decision-making processes. The amount of generated data has grown very large over the years, and conventional database processing methods from previous generations have not been sufficient to provide satisfactory results regarding analytics performance and prediction accuracy metrics. Thus, new methods are needed in a wide array of fields from computer architectures, storage systems, network design to statistics and physics. This thesis proposes two methods to address the current challenges and meet the future demands of advanced analytics. First, we present AxleDB, a Field Programmable Gate Array based query processing system which constitutes the frontend of an advanced analytics system. AxleDB melds highly-efficient accelerators with memory, storage and provides a unified programmable environment. AxleDB is capable of offloading complex Structured Query Language queries from host CPU. The experiments have shown that running a set of TPC-H queries, AxleDB can perform full queries between 1.8x and 34.2x faster and 2.8x to 62.1x more energy efficient compared to MonetDB, and PostgreSQL on a single workstation node. Second, we introduce TauRieL, a novel deep reinforcement learning (DRL) based method for combinatorial problems. The design idea behind combining DRL and combinatorial problems is to apply the prediction capabilities of deep reinforcement learning and to use the universality of combinatorial optimization problems to explore general purpose predictive methods. TauRieL utilizes an actor-critic inspired DRL architecture that adopts ordinary feedforward nets. Furthermore, TauRieL performs online training which unifies training and state space exploration. The experiments show that TauRieL can generate solutions two orders of magnitude faster and performs within 3% of accuracy compared to the state-of-the-art DRL on the Traveling Salesman Problem while searching for the shortest tour. Also, we present that TauRieL can be adapted to the Knapsack combinatorial problem. With a very minimal problem specific modification, TauRieL can outperform a Knapsack specific greedy heuristics.Hoy en día, se han incorporado grandes cantidades de datos estructurados y no estructurados en las bases de datos, y se les aplican procesos analíticos para descubrir patrones, correlaciones, tendencias y otras relaciones útiles que se utilizan mayormente para la toma de decisiones. La cantidad de datos generados ha crecido enormemente a lo largo de los años, y los métodos de procesamiento de bases de datos convencionales utilizados en las generaciones anteriores no son suficientes para proporcionar resultados satisfactorios respecto al rendimiento del análisis y respecto de la precisión de las predicciones. Por lo tanto, se necesitan nuevos métodos en una amplia gama de campos, desde arquitecturas de computadoras, sistemas de almacenamiento, diseño de redes hasta estadísticas y física. Esta tesis propone dos métodos para abordar los desafíos actuales y satisfacer las demandas futuras de análisis avanzado. Primero, presentamos AxleDB, un sistema de procesamiento de consultas basado en FPGAs (Field Programmable Gate Array) que constituye la interfaz de un sistema de análisis avanzado. AxleDB combina aceleradores altamente eficientes con memoria, almacenamiento y proporciona un entorno programable unificado. AxleDB es capaz de descargar consultas complejas de lenguaje de consulta estructurado desde la CPU del host. Los experimentos han demostrado que al ejecutar un conjunto de consultas TPC-H, AxleDB puede realizar consultas completas entre 1.8x y 34.2x más rápido y 2.8x a 62.1x más eficiente energéticamente que MonetDB, y PostgreSQL en un solo nodo de una estación de trabajo. En segundo lugar, presentamos TauRieL, un nuevo método basado en Deep Reinforcement Learning (DRL) para problemas combinatorios. La idea central que está detrás de la combinación de DRL y problemas combinatorios, es aplicar las capacidades de predicción del aprendizaje de refuerzo profundo y el uso de la universalidad de los problemas de optimización combinatoria para explorar métodos predictivos de propósito general. TauRieL utiliza una arquitectura DRL inspirada en el actor-crítico que se adapta a redes feedforward. Además, TauRieL realiza el entrenamieton en línea que unifica el entrenamiento y la exploración espacial de los estados. Los experimentos muestran que TauRieL puede generar soluciones dos órdenes de magnitud más rápido y funciona con un 3% de precisión en comparación con el estado del arte en DRL aplicado al problema del viajante mientras busca el recorrido más corto. Además, presentamos que TauRieL puede adaptarse al problema de la Mochila. Con una modificación específica muy mínima del problema, TauRieL puede superar a una heurística codiciosa de Knapsack Problem.Postprint (published version

    Learning a Scalable Algorithm for Improving Betweenness in the Lightning Network

    Get PDF
    This paper presents a scalable algorithm for solving the Maximum Betweenness Improvement Problem as it occurs in the Bitcoin Lightning Network. In this approach, each node is embedded with a feature vector whereby an Advantage Actor-Critic model identifies key nodes in the network that a joining node should open channels with to maximize its own expected routing opportunities. This model is trained using a custom built environment, lightning-gym, which can randomly generate small scale-free networks or import snapshots of the Lightning Network. After 100 training episodes on networks with 128 nodes, this A2C agent can recommend channels in the Lightning Network that perform competitively with recommendations from centrality based heuristics and in less time. This approach provides a fast, low resource, algorithm for nodes to increase their expected routing opportunities in the Lightning Network

    Ant colony meta-heuristics - Schemes and software framework

    Get PDF
    Master'sMASTER OF SCIENC

    Genetic Algorithms and the Satisfiability of Large-Scale Boolean Expressions.

    Get PDF
    The two new genetic methods overpopulation and bitwise expected value are introduced. In overpopulation a temporary population of size Mn (M 3˘e\u3e 1) is created using genetic operators and the n children with the highest estimated fitness values are selected as the next generation. The rest are discarded. Bitwise expected value (bev) is the fitness estimation function used. Overpopulation and bitwise expected value are applied to the NP-complete problem 3SAT (a special form of Satisfiability in which the boolean expression consists of the conjunction of an arbitrary number of clauses where each clause consists of the disjunction of 3 boolean variables) with excellent empirical results when compared to the performance of the standard genetic algorithm. Overpopulation increases the cost of producing each generation due to the overhead required to maintain the larger temporary population but results in many fewer generations to solution. Using bitwise expected value as a fitness estimator causes the algorithm to take slightly more generations to solution but is much faster to calculate than the fitness function, leading to a decrease in wall-clock time to solution. Theoretical justification for the success of overpopulation is seen as a result of the generalization of the schema growth equation. Bitwise expected value is viewed as an analogy to the Building Block Hypothesis. Empirical evidence of high correlation between bev and the fitness function is presented. We also introduce the target problem concept, in which a difficult problem is transformed into a well-known problem for which a good genetic method of solution is known. As an example of the target problem concept a transformation from the Traveling Salesman Problem to Satisfiability is demonstrated. Overpopulation and bitwise expected value are applied to the resulting boolean expression, with good results. An interesting convergence property is observed
    corecore